spot_img
HomeNews & Current EventsResearch Indicates AI Models Can Suffer Cognitive Decline from...

Research Indicates AI Models Can Suffer Cognitive Decline from Exposure to Low-Quality Online Data

TLDR: A new study by Cornell University reveals that large language models (LLMs) like ChatGPT can develop ‘brain rot’ when continually exposed to low-quality online content, leading to significant drops in accuracy, comprehension, and reasoning abilities. The research highlights that ‘junk web text’ induces lasting cognitive decline in AI, mirroring similar effects observed in humans from excessive consumption of superficial digital content.

A groundbreaking study conducted by researchers at Cornell University has unveiled a concerning phenomenon: artificial intelligence models, including prominent large language models (LLMs) such as ChatGPT and Gemini, are susceptible to a form of ‘brain rot’ when trained on or exposed to extensive amounts of low-quality online content. The findings, detailed in a paper titled ‘LLMs Can Get “Brain Rot”’, suggest that ‘junk web text’ induces a lasting cognitive decline in these advanced AI systems, drawing parallels to the negative impact of similar content on human cognition.

The study’s authors exposed an LLM to online gibberish and observed a significant deterioration in its capabilities. Specifically, the accuracy of AI models using this ‘brain rot’ content plummeted from 74.9 percent to 57.2 percent. Furthermore, their ability to accurately understand information with substantial context saw an even more drastic reduction, dropping from 84.4 percent to 52.3 percent. This indicates a profound hit to the cognitive and comprehensive capabilities of LLMs after prolonged exposure to low-quality data.

Researchers identified two primary parameters for defining ‘junk content’ on social media platforms like X (formerly Twitter). One category focused on short, viral posts with high engagement (retweets and likes), while the other included clickbait posts characterized by false claims and attention-grabbing language. These types of posts were then used to assess their impact on AI models such as Llama 3 and Qwen 2.5.

Upon analyzing the affected LLMs, the researchers noted a behavioral symptom termed ‘thought-skipping,’ where models would prematurely abandon reasoning chains. Additionally, models fed on junk data exhibited ‘dark traits,’ showing tendencies towards psychopathy and narcissism. The study also explored potential remedies, attempting to retrain the compromised LLMs with fresh, high-quality content. While some improvement in reasoning accuracy was observed, it did not fully restore the models to their original baseline performance, a phenomenon the researchers described as ‘persistent representational drift.’ This suggests that the damage inflicted by ‘brain rot’ may not be easily reversible.

Also Read:

The implications of this research are significant, especially as generative AI continues its rapid progression and finds applications across diverse fields including computing, education, finance, and medicine. With companies like Google, Microsoft, OpenAI, and Anthropic investing billions into AI technology, ensuring the quality of training data becomes paramount to prevent the degradation of AI models and maintain their reliability and accuracy. The study underscores the critical need for robust quality control mechanisms in the data pipelines that feed these powerful AI systems.

Nikhil Patel
Nikhil Patelhttps://blogs.edgentiq.com
Nikhil Patel is a tech analyst and AI news reporter who brings a practitioner's perspective to every article. With prior experience working at an AI startup, he decodes the business mechanics behind product innovations, funding trends, and partnerships in the GenAI space. Nikhil's insights are sharp, forward-looking, and trusted by insiders and newcomers alike. You can reach him out at: [email protected]

- Advertisement -

spot_img

Gen AI News and Updates

spot_img

- Advertisement -