Message of the dying crabs


For those who have enjoyed the Discovery Channel show Deadliest Catch, (309 episodes and counting), which follows the lives and fortunes of fishermen who scour the Berring Sea, off Alaskan coast, for snow crabs, there is bad news. The snow crabs, an iconic species, are disappearing.

According to a study published in Science, between 2018 and 2021, some 10 billion snow crabs disappeared; revenues from snow crabs, which averaged about $150 million, fell to $24 million in the 2021-22 crabbing season.

Why? Marine heat waves, caused by global warming.

Snow crabs like it nice and cold; they live in waters that are about 20 C, though they can stay alive even up to 120 C. It was not the heat that killed the crabs, the study says, but the warming waters triggered off a chain of ecological events that resulted in not enough food for the crabs, causing them to starve to death.

“Calculated caloric requirements, reduced spatial distribution, and observed body conditions suggest that starvation played a role in the collapse. The mortality event appears to be one of the largest reported losses of motile marine macrofauna to marine heatwaves globally,” the scientists say.

With the next round of international climate negotiations (COP28) approaching, these findings should serve as a sharp reminder for urgent action.





Source link

Can AI reduce the backlog of cases? Researchers say yes


In India, the judiciary is grappling with an overwhelming backlog of over 50 million pending cases. Some believe that AI has the potential to reduce the number of cases. Researchers from the University of Liverpool used language models to generate legal arguments from case facts. The top method achieved a 63 per cent overlap with benchmark annotations.

AI can summarise, suggest and predict applicable statutes, reducing the time spent on document processing and aiding legal professionals, says Procheta Sen, one of the authors of the paper: “Automated argument generation from legal facts”.

“We used open-source models like GPT-2 and Facebook’s LLaMA for argument generation,” says Sen. LLaMA (Large Language Model Meta AI) is part of a family of large language models (LLMs) released by Facebook’s Meta AI in February 2023.

Large Language Models and the Judiciary

LLMs have found success in various natural language processing (NLP) tasks such as machine translation, summarisation and entity recognition. Starting with the transformer architecture, these models employ pre-trained, fine-tuned and prompt-based approaches to NLP tasks. Pre-trained models such as like BERT and GPT-2 have outperformed baselines in numerous NLP tasks.

Sen, et al’s research paper usedGPT-2 and Flan-T5 models to generate legal arguments from factual information. Under the umbrella of LLMs, these models are fine-tuned using special tokens like ‘[Facts]’ and ‘[Arguments]’ to guide the generation process. Legal documents, known for their length, pose a challenge due to token limits, which could be overcome by using a BERT summariser for content condensation. The dataset had 50 legal documents from the Indian Supreme Court’s corpus, with each sentence labelled with one of seven rhetorical role categories — facts, ruling by lower court, argument, statute, precedent, ratio of decision, ruling by present court. The core idea lies in optimising argument generation through different summaries facilitated by BERT. Evaluation metrics include average word overlap and average semantic similarity.

The researchers used two evaluation metrics that include average word overlap (it measured shared words between generated and actual arguments) and average semantic similarity (similarity between BERT embeddings of generated and actual arguments). They found that, “ with the increase in the number of sentences in the summary, the quality of the generated argument also increased.”

Additionally, it was found that better data quality enhanced also enhances the model’s performance.

Not everything is rosy

But the challenge in understanding the material stems from the poorly structured English sentences in legal case proceedings, says Sen. This lack of refinement hampers the use of existing NLP tools and requires significant human effort for comprehension, she adds.

The limitations also include privacy concerns. When using paid API services as sensitive data might be shared. Similarly, potential biases can exist in larger datasets, but proper fine-tuning and high-quality data can mitigate this issue, according to Sen.

While NLP has developed significantly, Sen feels that the need of the hour is “well-curated data.” Preserving case processing in a structured manner and creating annotated data also consumes lots of time, adds Sen.

While the research did explore a wide area for the judiciary, the data set was very limited. The current work is an initial exploration and more advanced models are planned for the future, says Sen.





Source link

Can AI reduce the backlog of cases? Researchers say yes

Can AI reduce the backlog of cases? Researchers say yes


In India, the judiciary is grappling with an overwhelming backlog of over 50 million pending cases. Some believe that AI has the potential to reduce the number of cases. Researchers from the University of Liverpool used language models to generate legal arguments from case facts. The top method achieved a 63 per cent overlap with benchmark annotations.

AI can summarise, suggest and predict applicable statutes, reducing the time spent on document processing and aiding legal professionals, says Procheta Sen, one of the authors of the paper: “Automated argument generation from legal facts”.

“We used open-source models like GPT-2 and Facebook’s LLaMA for argument generation,” says Sen. LLaMA (Large Language Model Meta AI) is part of a family of large language models (LLMs) released by Facebook’s Meta AI in February 2023.

Large Language Models and the Judiciary

LLMs have found success in various natural language processing (NLP) tasks such as machine translation, summarisation and entity recognition. Starting with the transformer architecture, these models employ pre-trained, fine-tuned and prompt-based approaches to NLP tasks. Pre-trained models such as like BERT and GPT-2 have outperformed baselines in numerous NLP tasks.

Sen, et al’s research paper usedGPT-2 and Flan-T5 models to generate legal arguments from factual information. Under the umbrella of LLMs, these models are fine-tuned using special tokens like ‘[Facts]’ and ‘[Arguments]’ to guide the generation process. Legal documents, known for their length, pose a challenge due to token limits, which could be overcome by using a BERT summariser for content condensation. The dataset had 50 legal documents from the Indian Supreme Court’s corpus, with each sentence labelled with one of seven rhetorical role categories — facts, ruling by lower court, argument, statute, precedent, ratio of decision, ruling by present court. The core idea lies in optimising argument generation through different summaries facilitated by BERT. Evaluation metrics include average word overlap and average semantic similarity.

The researchers used two evaluation metrics that include average word overlap (it measured shared words between generated and actual arguments) and average semantic similarity (similarity between BERT embeddings of generated and actual arguments). They found that, “ with the increase in the number of sentences in the summary, the quality of the generated argument also increased.”

Additionally, it was found that better data quality enhanced also enhances the model’s performance.

Not everything is rosy

But the challenge in understanding the material stems from the poorly structured English sentences in legal case proceedings, says Sen. This lack of refinement hampers the use of existing NLP tools and requires significant human effort for comprehension, she adds.

The limitations also include privacy concerns. When using paid API services as sensitive data might be shared. Similarly, potential biases can exist in larger datasets, but proper fine-tuning and high-quality data can mitigate this issue, according to Sen.

While NLP has developed significantly, Sen feels that the need of the hour is “well-curated data.” Preserving case processing in a structured manner and creating annotated data also consumes lots of time, adds Sen.

While the research did explore a wide area for the judiciary, the data set was very limited. The current work is an initial exploration and more advanced models are planned for the future, says Sen.





Source link

How Large Language Models can revolutionise financial analysis

How Large Language Models can revolutionise financial analysis


Financial analysts and investment professionals will have their lives just that little bit easier as a cutting-edge technology promises to redefine the way they extract critical insights from corporate earnings reports.

A groundbreaking study titled “Towards reducing hallucination in extracting information from financial reports using Large Language Models” by Bhaskarjit Sarmah, Tianjie Zhu, Dhagash Mehta and Stefano Pasquali demonstrates the remarkable potential of Large Language Models (LLMs) to extract information efficiently and accurately from earnings report transcripts.

Precise, reliable

This game-changing approach combines retrieval-augmented generation techniques with metadata integration to extract information from earnings reports. In a comparative analysis of various pre-trained LLMs, the study shows that this innovative method outperforms traditional techniques with unprecedented precision and reliability.

The Q&A section of corporate earnings reports has long been a treasure trove of information for financial analysts and investors. It offers insights and answers to crucial questions about a company’s performance, strategy and financial health. However, the traditional methods of analysing this section, such as detailed reading and note-taking, have been time-consuming and error-prone. Moreover, Optical Character Recognition (OCR) and other automated techniques often struggle to accurately process unstructured transcript text, missing essential linguistic nuances that drive investment decisions.

Enter Large Language Models (LLMs) such as BERT and GPT-3.. These models have the unique ability to understand contextual nuances, enabling them to identify and extract relevant question-answer pairs accurately. LLMs offer a data-driven approach that adapts to the dynamic language patterns found in earnings reports, significantly enhancing both efficiency and precision in information extraction.

However, one persistent challenge with LLMs is the potential for generating responses that deviate from factual accuracy, often referred to as “hallucination.” The study presents an innovative remedy by enhancing LLMs through the integration of retrieval systems. By incorporating external repositories of information, these retrieval-augmented LLMs aim to bolster accuracy and context in generated responses. Nonetheless, challenges remain, particularly when dealing with multiple documents. In such cases, the model might inadvertently extract information from unintended sources, leading to the emergence of hallucinatory responses.

To address these multifaceted challenges comprehensively, the researchers, in addition to integrating retrieval-augmented LLMs, used metadata to mitigate the occurrence of hallucinatory responses. This enhances the reliability and precision of information extracted by the LLMs, while ensuring that responses align more closely with the actual context and requirements of user queries.

Earnings calls of Nifty 50 constituents, a widely recognised and extensive collection of earnings call transcripts, were used as the data source for the study. The dataset encompassing the quarter ending in June 2023, provided a diverse foundation for the research, with transcripts from companies across various sectors being used.

The methodology employed in this study also overcomes the limitations of LLMs, which are trained on data up to a specific cut-off point, lacking access to new information or context that emerges post-training. Retrieval-augmented generation is introduced as a paradigm shift in LLM technology. This approach enhances LLM capabilities by integrating retrieval systems into their architecture, reducing the likelihood of generating false or misleading content.

Superior performance

When documents exceed the context window of LLMs, the study introduces a smart approach called “chunking.” This process involves breaking down documents into smaller, more manageable segments that fit within the context window of the LLM, thereby maintaining accuracy and relevance.

Ground-truth labels are used for a comprehensive examination of earnings reports and a range of randomly selected questions posed during earnings calls. The results indicate that the integration of metadata significantly improves the accuracy and relevance of generated answers. Several evaluation metrics, including BERTScore and Jaro similarity, confirm the superior performance of the proposed approach.

LLMs, when harnessed effectively, have the potential to transform the way financial analysts and investors extract critical insights from earnings reports. The integration of retrieval-augmented generation and metadata not only mitigates hallucinatory responses but also enhances the precision and reliability of the information extraction process. With these advancements in LLM technology, financial professionals can now look forward to a more efficient and accurate analysis process.





Source link

Here comes an AI-based process to shorten drug discovery cycle

Here comes an AI-based process to shorten drug discovery cycle


Drug discovery is a painstaking process. The majority of attempts often go waste after 6-8 years of working on a drug, making it an inordinately expensive exercise. A chip design company, a US-based firm with an R&D centre in Hyderabad, claims that it has found an AI-backed method to cut the process short and save costs.

Venkat Mattela, Founder and Chief Executive Officer of Ceremorphic, said that the company has created a computer-generated cell, which mimics a human cell, making it easy for researchers to understand how it reacts when exposed to a disease or a drug.

Mattela sold his earlier venture Redpine Signals to Silicon Labs for $308 million in 2020

“The aim is to shorten the development cycle and reduce the side effects by choosing the right molecule based on new development methodology,” he told businessline here on Thursday.

“Over 90 per cent of drug candidates fail at the clinical phase-ii stage and the number of new drugs cleared by FDA every year is small. This calls for a new development methodology in the early stages of the design pipeline,” he said.

The firm has started an exclusive Life Sciences division with about 32 people to work on oncology and neurology, with over 70 per cent of them working out of its Hyderabad R&D facility. The team comprised computational biologists, chemists, computer scientists and analog hardware designers.

The company, which taped-out a 5-nanometer chip in 2022, would use a 16-nm chip to work on the drug discovery.

“Accurate AI foundation models, new algorithms and data are critical to producing successful later stage results,” he said.

“Deploying our proprietary analog circuit technology, supercomputing chips and novel AI algorithms, we developed an architecture that can predict the later stage outcomes to increase R&D efficiency,” he said.

He claimed that this innovation would transform the future of drug discovery to accelerate making personalised medicine a reality.

For the subsequent levels in drug discovery, the company plans to forge alliances with relevant life sciences companies.





Source link

Synthesised pyrite for supercapacitors

Synthesised pyrite for supercapacitors


Researchers have synthesised highly crystalline pyrite FeS2 at low temperatures and utilised them for fabricating electrochemical energy storage devices such as batteries and high energy density supercapacitors, says a press release from the Department of Science and Technology (DST).

Transition metal sulfides (TMS) are an important class of inorganic materials and find applications in diverse fields including electrochemical energy storage devices such as batteries and supercapacitors. Solid-state synthetic methods are used to generate metal sulfides from the corresponding metal salts or their equivalent oxides usually by annealing at high temperatures.

However, the experiments were carried out by Savithri Vishwanathan, under the supervision of Dr HSS Ramakrishna Matte at the Centre for Nano and Soft Matter Sciences, Bengaluru, an autonomous institute under DST, demonstrated the low-temperature synthesis of crystalline pyrite FeS2 through a solid-state synthesis route. They have utilised a metastable oxyhydroxide (FeOOH) precursor for this process.

The team reported stabilising this intermediate oxyhydroxide and utilising it as a precursor for sulfidation, in the presence of H2S gas, for the first time in their paper published in the journal Chemical Communications.

Using a metastable precursor helped in lowering the annealing temperature, as FeOOH converted into pyrite FeS2 with fairly good crystallinity at a low temperature. This synthetic route of obtaining sulfides from their corresponding metastable oxyhydroxides can be extended to other transition metals to obtain crystalline materials in an energy intensive way.





Source link

YouTube
Instagram
WhatsApp