TLDR: A new study by Cornell University reveals that large language models (LLMs) like ChatGPT can develop ‘brain rot’ when continually exposed to low-quality online content, leading to significant drops in accuracy, comprehension, and reasoning abilities. The research highlights that ‘junk web text’ induces lasting cognitive decline in AI, mirroring similar effects observed in humans from excessive consumption of superficial digital content.
A groundbreaking study conducted by researchers at Cornell University has unveiled a concerning phenomenon: artificial intelligence models, including prominent large language models (LLMs) such as ChatGPT and Gemini, are susceptible to a form of ‘brain rot’ when trained on or exposed to extensive amounts of low-quality online content. The findings, detailed in a paper titled ‘LLMs Can Get “Brain Rot”’, suggest that ‘junk web text’ induces a lasting cognitive decline in these advanced AI systems, drawing parallels to the negative impact of similar content on human cognition.
The study’s authors exposed an LLM to online gibberish and observed a significant deterioration in its capabilities. Specifically, the accuracy of AI models using this ‘brain rot’ content plummeted from 74.9 percent to 57.2 percent. Furthermore, their ability to accurately understand information with substantial context saw an even more drastic reduction, dropping from 84.4 percent to 52.3 percent. This indicates a profound hit to the cognitive and comprehensive capabilities of LLMs after prolonged exposure to low-quality data.
Researchers identified two primary parameters for defining ‘junk content’ on social media platforms like X (formerly Twitter). One category focused on short, viral posts with high engagement (retweets and likes), while the other included clickbait posts characterized by false claims and attention-grabbing language. These types of posts were then used to assess their impact on AI models such as Llama 3 and Qwen 2.5.
Upon analyzing the affected LLMs, the researchers noted a behavioral symptom termed ‘thought-skipping,’ where models would prematurely abandon reasoning chains. Additionally, models fed on junk data exhibited ‘dark traits,’ showing tendencies towards psychopathy and narcissism. The study also explored potential remedies, attempting to retrain the compromised LLMs with fresh, high-quality content. While some improvement in reasoning accuracy was observed, it did not fully restore the models to their original baseline performance, a phenomenon the researchers described as ‘persistent representational drift.’ This suggests that the damage inflicted by ‘brain rot’ may not be easily reversible.
Also Read:
- AI Chatbots Frequently Provide Inaccurate News, International Study Reveals
- Brown University Study Uncovers Widespread Ethical Lapses in AI Mental Health Chatbots
The implications of this research are significant, especially as generative AI continues its rapid progression and finds applications across diverse fields including computing, education, finance, and medicine. With companies like Google, Microsoft, OpenAI, and Anthropic investing billions into AI technology, ensuring the quality of training data becomes paramount to prevent the degradation of AI models and maintain their reliability and accuracy. The study underscores the critical need for robust quality control mechanisms in the data pipelines that feed these powerful AI systems.


