Journalism of Courage
Advertisement

AI models like ChatGPT can develop ‘brain rot’ from online junk content, study reveals

A new study reveals that low quality online content, often referred to as “brain rot” can dull an AI model’s reasoning and cognitive skills, much like it does in humans.

AI "brain rot" is real, and it can hurt an AI model's reasoning and thinking capabilties.AI "brain rot" is real, and it can hurt an AI model's reasoning and thinking capabilties. (Image Source: YouTube/GingersDiary)

Generative AI has progressed quite rapidly, finding use in various fields like computing, education, finance, medical and others. Although AI still suffers from hallucinations, companies like Google, Microsoft, OpenAI and Anthropic are investing billions of dollars into the technology.

Large language models – which are the backbones of AI chatbots like ChatGPT and Gemini are trained on information from all over the internet, a new study by Cornell University suggests even AI suffers from “brain rot” after long exposure to low-quality data.

In a paper titled LLMs Can Get “Brain Rot”, researchers say they exposed an LLM to online gibberish and noted that “junk web text induces lasting cognitive decline.” Studies have previously shown that such content negatively affects human cognitive capabilities like reasoning and focus as well, and now the same can be said for AI models.

Researchers say they used two parameters to identify junk content on the social media platform X. While one test focused on short and viral posts that had a lot of retweets and likes, the second one included clickbait posts with false claims and attention-grabbing words. These posts were then used to study the impact on AI models like Llama 3 and Qwen 2.5.

In the study, it was noted that the accuracy of these AI models using brain rot content fell from 74.9 per cent to 57.2 per cent. As it turns out, these AI models were unable to accurately understand information with a lot of context, with the capability dropping from 84.4 per cent to 52.3 per cent.

After these LLMs were exposed to more junk content, their cognitive and comprehensive capabilities took a drastic hit. Moreover, the low quality also made them less reliable, as they generated more incorrect responses. Upon analysing these LLMs, researchers noted that models trained on junk data engaged in “thought skipping”, a process where LLMs often skip some steps while reasoning. LLMs fed on junk data also had some “dark traits”, which made them lean towards psychopathy and narcissism.

When researchers tried to fix these LLMs trained on junk data by retraining with fresh content, the results were partially better. While reasoning accuracy improved slightly, it still fell short of the original baseline – a phenomenon researchers describe as “persistent representational drift.”

Story continues below this ad

The research paper suggested that all popular LLMs be periodically checked for cognitive decline in a three-step process. The first one includes the routine cognitive evaluation of AI models to detect early signs of reasoning decline, while the second and third ones are aimed at controlling data quality during pre-training and studying how viral or low-quality content reshapes its learning patterns.

From the homepage
Tags:
  • artificial intelligence
Edition
Install the Express App for
a better experience
Featured
Trending Topics
News
Multimedia
Follow Us
Long ReadsFelony, fear, fight for domination: How Sigma & Co 'turned murder into business' in Bihar’s Sitamarhi
X