First test flight of Gaganyaan is expected in December 2024: ISRO Chairman

First test flight of Gaganyaan is expected in December 2024: ISRO Chairman


The first test flight of Gaganyaan is expected to happen in December 2024, said Indian Space Research Organisation (ISRO) Chairman S Somanath.

The three stages of the Gaganyaan rocket have arrived at the Satish Dhawan Space Centre (SDSC, SHAR). The integration of the crew module is taking place at the Vikram Sarabhai Space Centre, Thiruvananthapuram, he told newspersons after the successful launch of India’s Earth Observation Satellite-08 (EOS-08) into orbit.

All the systems for the Gaganyaan rocket – codenamed G1 – will reach SDSC in November this year and the target for the rocket flight is December, he said.

The Gaganyaan project envisages demonstration of human spaceflight capability by launching a crew of three members to an orbit of 400 km for a three-day mission and bring them back safely to earth, by landing in Indian sea waters.

The project is accomplished through an optimal strategy taking into account inhouse expertise, the experience of Indian industry, intellectual capabilities of Indian academia and research institutions, along with cutting-edge technologies available with international agencies.

The pre-requisites for the Gaganyaan mission include development of critical technologies including the human-rated launch vehicle for carrying crew safely to space, the Life Support System to provide an earth-like environment for the crew in space, crew emergency escape provision and evolving crew management aspects for training, recovery and rehabilitation of crew, according to information on the ISRO website.





Source link

The paradox of powerful AI: When bigger isn’t always better

The paradox of powerful AI: When bigger isn’t always better


Many of us are used to setting alarms with Siri, asking Google for nearby restaurants or telling Alexa to turn up the lights. Even if we don’t use these AI-powered assistants ourselves, we often see others using them. However, when it comes to complex tasks, like drafting an email to the boss, the results can be baffling or even a confusing mess. This stark difference in outcomes raises an important question: Do large language models (LLMs) — the brains behind our virtual assistants and chatbots — really perform as we expect them to?

LLMs have become the cornerstone of modern AI applications. Models like OpenAI’s GPT-4, Google’s Gemini, Meta’s LLAMA, are capable of generating human-like text, translating languages, writing code and even crafting poetry. The excitement around LLMs stems from their ability to handle a diverse range of tasks using a single model. This versatility offers immense potential — imagine a model helping a doctor summarise patient notes while also assisting a software engineer in debugging code.

However, this very diversity also presents a significant challenge — How do we evaluate such a multifaceted tool? Traditional models are typically designed for specific tasks and evaluated against benchmarks tailored to those tasks. But with LLMs, it’s impractical to create benchmarks for every possible application they might be used for. This raises an essential question for researchers and users alike: How can we gauge where an LLM will perform well and where it might stumble?

The LLM dielemma

The crux of the problem lies in understanding human expectations. When deciding where to deploy an LLM, we naturally rely on our interactions with the model. If it performs well on one task, we might assume it will excel at related tasks. This generalisation process — where we infer the capabilities of a model based on limited interactions — is key to understanding and improving the deployment of LLMs.

In a new paper, MIT researchers Keyon Vafa, Ashesh Rambachan and Sendhil Mullainathan took a different approach. In their study — ‘Do large language models perform the way people expect? Measuring the human generalisation function’ — they have explored how humans form beliefs about LLM capabilities and whether these beliefs align with the models’ actual performance.

To start with, the researchers collected a substantial dataset of human generalisations. They surveyed participants, presenting them with examples of how an LLM responded to specific questions. The participants were then asked whether these responses influenced their beliefs about how the model would perform on other, related tasks. This data collection spanned 19,000 examples across 79 tasks, sourced from well-known benchmarks like the MMLU and BIG-Bench.

On analysing the data using sophisticated natural language processing (NLP) techniques, they found that human generalisations are not random; unsurprisingly, they follow consistent, structured patterns that can be predicted using existing NLP methods.

The researchers also evaluated how well different LLMs align with these human generalisations. They tested several models for this, including GPT-4, to see if their performance matched human expectations, and discovered a paradox: larger, more capable models like GPT-4 often performed worse in high-stakes scenarios, precisely because users overestimated their capabilities. In contrast, smaller models sometimes aligned better with human expectations, leading to more reliable deployment in critical applications.

The researchers used a novel approach to evaluate model alignment. Instead of relying on fixed benchmarks, they modelled the human deployment distribution — the set of tasks humans choose based on their beliefs about the model’s capabilities. This method acknowledges that real-world use depends not just on the model’s abilities but also on human perceptions of those abilities.

The findings of this research are both fascinating and cautionary. It highlights that while larger LLMs have impressive capabilities, their misalignment with human generalisations can lead to significant deployment errors.

On the flip side, by understanding and modelling human generalisations, we can better align LLMs with user expectations. This could involve developing better interfaces that help users accurately gauge a model’s strengths and weaknesses or creating more targeted training data that helps models perform consistently across a broader range of tasks.





Source link

Using deep learning to predict stock prices gains maturity

Using deep learning to predict stock prices gains maturity


Alchemists of yore sought to short-cut their way to richness by finding ways to convert metals into gold. Today, it appears that the job of finding a quick pathway to prosperity has been taken up by data scientists. Alchemists, of course, failed to transmute cheap metals into gold, but data scientists may be sniffing success.

Today, it is possible to take data sets of quarterly results of a company, daily high, low and close prices and traded volumes, as well as economic data, news and sentiment analysis, and use the data to train a machine learning model to make predictions. The model essentially sees patterns that we would miss.

Data scientists first clean up the data (filling in any missing values and removing outliers) and make it usable by a model. Then they select a deep learning technique (like Long Short-Term Memory) or Convolutional Neural Networks, for milking the data for insights. LSTM is particularly appropriate for time series data, such as historical stock prices, because it can remember previous information and use it to make predictions. Then the model is trained. The training process involves making the model adjust its parameters to reduce the gap between its own predictions and actual stock prices. Then comes validation of the model, using a separate set of data to verify the model’s accuracy.

In broad terms, AI and deep learning predict stock prices by analysing historical data, identifying patterns and making future price forecasts. This involves collecting and pre-processing data, selecting and training appropriate models, making predictions, and continuously evaluating and refining the models to improve their accuracy and reliability.

Hitting the marks

In the last few years, a lot of work has been done in this area, with each model getting better than the earlier ones. A recent work in this area by three data scientists, Jaydip Sen, Hetvi Waghela and Sneha Rakshit, of the Department of Data Science and Artificial Intelligence at the Praxis Business School, Kolkata, has thrown up a model that boasts of 95.8 per cent accuracy in predicting the next day’s stock prices.

Sen, Waghela and Sneha used Long Short-term Memory (LSTM), which is a special kind of artificial neural network used in deep learning, designed to remember information for long periods of time, for “accurate stock price prediction”. In a recent paper (which is yet to be peer-reviewed), they note that “despite the efficient market hypothesis suggesting that such predictions are impossible, there are propositions in the literature demonstrating that advanced algorithms and predictive models can effectively forecast future stock prices.” They add that in recent times, the use of machine learning and deep learning systems has become popular for market price prediction.

Sen, Waghela and Rakshit believe that an LSTM model could predict stock prices better than other techniques such as convolutional neural networks that has been used by others earlier. “This model automatically retrieves historical stock prices using a Python function, utilising stock ticker names from the NSE within a specified interval determined by a start and end date.”

They took historical prices of 180 stocks across 18 sectors, between January 1, 2005 and April 23, 2024. They trained the data on LSTM model to make predictions. They used three metrics — Huber Loss, Mean Absolute Error and Accuracy Score — which give an idea of how wrong a prediction could be, to assess the performance of their model. For a model to be accurate, it must have low values for Huber loss and MAE, and a high value for the accuracy score.

Golden insight

Sen Waghela and Rakshit observed that their LSTM model yielded the minimum value of Huber loss for the auto sector, the minimum value of MAE for the banking sector, and the maximum value of accuracy score for the PSU banks sector. “Hence, the model is found to be the most accurate for these three sectors,” they note. On the other hand, Huber loss and MAE are the maximum for the media sector and accuracy score is the minimum for the energy sector. Therefore, the performance of the model has been the worst for media and energy sectors on the three metrics.

However, overall, the model’s performance has been highly accurate, since the lowest value of the accuracy score is 0.958315. “In other words, in the worst case, the model correctly predicted the direction of movement of the price (upward or downward) of the stock the next day in 95.83 per cent of cases. Thus, the model can be reliably used in stock trading decisions,” they say.

The use of deep learning techniques for making stock predictions will call for radical changes in regulations, so that the market does not get skewed in favour of those who know how to use techniques at the expense of those who don’t.





Source link

Non-toxic ionic liquids for silk processing

Non-toxic ionic liquids for silk processing


Researchers have found an eco-friendly approach that can eliminate the use of toxic chemicals in silk processing.

Traditionally, toxic chemicals like sodium carbonate, sodium hydroxide, sulfuric acid and lithium bromide have been used to extract silk proteins, fibroin and sericin from various types of raw silk fibres, an important step in the process of making silk from cocoons.

A team at the Institute of Advanced Study in Science and Technology (IASST), Guwahati, has identified Ionic Liquids (ILs) which can be sustainable alternatives to the toxic chemicals currently in use for the silk protein extraction process, according to a press release from the Department of Science and Technology, Government of India. The team, led by Dr Kamatchi Sankaranarayan, has identified four such ILs. Published in Chemistry Select, this research has potential for use in sericin extraction from both mulberry (Bombyx mori) and non-mulberry silks, such as Muga (Antheraea assamensis) and Eri (Philosamia ricini), indigenous to North-east India. Not only does it offer environmentally friendly alternative to traditional chemical methods, it also paves the way for efficient sericin extraction from non-mulberry silks, potentially leading to new applications for these unique fibres.

The researchers explored six different ILs and found some of them were particularly effective in removing sericin without damaging the silk protein structure. The ones showing greatest promise included 1-Butyl-3-methylimidazolium chloride (BMIM.Cl), 1-ethyl-3-methylimidazolium tetrafluoroborate (EMIM.BF4) and Tetraethylammonium bromide (TEAB). TEAB appeared to be highly effective due to its ability to destabilise sericin proteins.

Tackling battery power fade problem with Omics

Researchers at the Lawrence-Berkeley National Lab, California, US, have discovered a way to tackle the ‘power fade’ problem in batteries. The “power fade problem” in batteries refers to the gradual decrease in the battery’s ability to deliver power over time. The degradation affects the performance and efficiency of batteries, particularly in applications requiring high power output — which is why you have to throw away batteries after about five years. The power fade occurs due to factors such as ageing of electrodes and decomposition of electrolytes.

The L-B scientists took to ‘omics’ to study the power fade problem. The ‘Omics technique’ refers to a suite of technologies used to explore the roles, relationships, and actions of the various types of molecules that make up the cells of an organism.

The scientists “wanted to see if we could use a similar approach to examine the chemical signatures of the battery’s components and identify the reactions contributing to power fade and where they were occurring.”

The researchers focused their analysis on lithium metal batteries with high-voltage, high-density layered oxides containing nickel, manganese and cobalt. Contrary to prior research, which has typically thought the power fade problem was from the battery’s anode, the team observed that power fade stems from the cathode side. This was where particles cracked and corroded over time, hindering charge movement and reducing battery efficiency. “It was a non-obvious outcome,” Youngmin Ko, a postdoc researcher, said in a press release “We found that mixing salts in the electrolyte could suppress the reactivity of typically reactive species, which formed a stabilising, corrosion-resistant coating.”





Source link

Forward into past

Forward into past


You hail a taxi. A stylish cab called ‘Electrobat’ comes along. It is an electric vehicle.

A distance ahead, the driver stops, begs your forgiveness, for he needs just a jiffy to swap batteries. Three minutes later you are on your way, in a clean, green, electric vehicle — that has just avoided a huge pile of waste, the disposal of which has been a headache for the authorities.

Did I just take you to the future? Not at all. Quite the opposite — I took you to the past, to the year 1897, to New York’s infamous Manhattan, when electric vehicles, which we so covet these days, was the norm. Three years earlier, two engineers named Henry Morris and Pedro Salom, made an electric car and called it Electrobat. Rather than sell the vehicles, they got into running a taxi service. They called it — Electric Wagon & Carriage Company. The venture drove off with a dozen vehicles, but in just two years, was operating a hundred.

The battery it used was the old familiar lead-acid battery. It weighed a thousand pounds (454 kg, or about half as heavy as a Maruti Alto). With that weight, you would think that the vehicle can’t travel fast, but ‘fast’ is a relative term, conveying a different sense in every different context. In the context of 1899, it was an atrociously high speed of 25 km per hour. Guess what was the ‘waste’ that the vehicles avoided? Horse manure.

Electric Wagon & Carriage Co ran successfully until it lost its way due to a fraud, notes an article in National Geographic. In due course, internal combustion engines came and steamrolled the EVs out of business. But now, the EVs are having their sweet revenge.





Source link

De-identifying patients with AI: Ensuring patient privacy while enhancing research

De-identifying patients with AI: Ensuring patient privacy while enhancing research


In India, the most populous country in the world, efficient management — access, storage and retrieval — of healthcare data is increasingly critical. Imagine having access to health records of millions of patients, a treasure trove of information that could dramatically improve public health policies, advance medical research and enhance patient care. However, this also brings a significant challenge: protecting patient’s privacy.

A recent study “Generation and De-Identification of Indian Clinical Discharge Summaries using LLMs” by Sanjeet Singh et al, from the Indian Institute of Technology, Kanpur, (IIT Kanpur) and technology company Miimansa dives into this pressing issue.

The researchers explored how artificial intelligence (AI) can be harnessed to de-identify patient records, ensuring that sensitive information remains confidential while still being useful for research and policy-making.

Healthcare data is incredibly valuable. It can reveal patterns about the spread of diseases, the effectiveness of treatments, and the needs of different patient groups. In India, over 330 million patient records have already been linked with unique central IDs. This vast amount of data, roughly equivalent to the population of the US, represents an underutilised resource with the potential to revolutionise public health. However, it also poses a risk. If not handled properly, this data can expose individuals to privacy breaches. The consequences can be severe, from personal embarrassment to identity theft and financial loss.

To mitigate these risks, healthcare data must be de-identified, stripping it of any personal information that could reveal the patient’s identity. Natural Language Processing (NLP), a branch of AI that deals with the interaction between computers and human language, offers powerful tools for de-identification. NLP can scan through text, identify personal health information (PHI), and mask it.

However, there’s a catch: AI systems are only as good as the data they are trained on. Most existing systems have been trained on data from Western countries and they might not perform well on Indian data, given the cultural and linguistic differences.

De-identification of personal health information (PHI) is also critical to ensure compliance with privacy regulations such as the Indian Digital Personal Data Protection Act, 2023, (DPDPA) and similar laws like GDPR in Europe and HIPAA in the US.

The study from IIT Kanpur and Miimansa tackled this challenge head-on. Using a dataset of fully de-identified discharge summaries from an Indian hospital (the Sanjay Gandhi Post Graduate Institute of Medical Sciences, Lucknow), the researchers ran existing de-identification models, including commercial solutions. These models were originally trained on non-Indian datasets which primarily included data from US healthcare institutions. The results were telling: Models trained on non-Indian data did not perform well — a clear indication that AI models need to be trained on region-specific data to be effective.

Synthetic solution

To overcome this limitation, the researchers turned to a clever solution: synthetic data. By using large language models (LLMs) like Gemini, Gemma, Mistral and Llama3, they generated synthetic clinical reports that mimicked real patient data but did not correspond to actual patients, avoiding privacy issues. Training AI models on synthetic data dramatically improved their performance on the real Indian data.

This approach also ensures that healthcare data can be used safely for research and policy-making without risking patient privacy. For India, this could mean more accurate health statistics and better public health interventions.

While the results of this study are promising, there is still a long way to go. AI systems need continuous improvement and validation. The researchers plan to establish an active learning workflow that combines AI models with human expertise. This means that while AI will do the heavy lifting, human experts will refine and validate the results, creating a feedback loop that continuously enhances the system’s accuracy and reliability.

In a country as diverse and populous as India, blend of technology and human touch will be crucial in building a robust, resilient and responsive healthcare system.





Source link

YouTube
Instagram
WhatsApp