Model Collapse – when AI feeds on its own content

Share post:

As AI-generated data floods the internet, it risks being ingested by future AI models, leading to a feedback loop that degrades quality. Research shows that AI systems trained on their own output can suffer from “model collapse,” where diversity and accuracy decline over time. For example, when AI models are trained repeatedly on their own content, like handwritten digits or text, they tend to produce more homogeneous and less accurate results, drifting away from the original data they were meant to mimic.

There is an exceptional illustration of this in an article in the NY Times.

This phenomenon poses significant challenges for AI development. As models increasingly consume AI-generated content, the quality of their outputs deteriorates, which could affect everything from medical advice to historical accuracy. Additionally, the lack of diversity in data can lead to biased and limited outputs, further compromising the reliability of AI systems. This trend highlights the importance of using high-quality, diverse human-generated data to train AI models and prevent the negative effects of self-generated data loops.

To mitigate these risks, AI companies are exploring strategies like watermarking AI-generated content, paying for high-quality data, and using synthetic data selectively under human supervision. These measures aim to ensure that AI continues to learn and evolve based on diverse and accurate inputs rather than becoming trapped in a cycle of self-reference and diminishing returns. As the reliance on AI grows, addressing these issues will be crucial for maintaining the effectiveness and safety of AI technologies.

SUBSCRIBE NOW

Related articles

AI and Cyber Security: Practical Insights. Hashtag Trending Weekend Edition (repeat episode)

Unlocking AI: Understanding the Expanding Role of AI in Business and Cybersecurity This is our repeat episode and if...

You.com versus Perplexity.ai. Two AI’s go to head with an twist. A debate between AI’s with an AI judge

This is a bit longer than our average article, but hopefully it's also a little bit of fun....

Is Windows Intelligent Media Search the next “Recall?”

Microsoft is reportedly working on a new AI feature for Windows 11, called "Intelligent Media Search," which can...

Facial recognition company fined for “illegallly creating a database of faces”

Clearview AI, a U.S.-based facial recognition company, has been fined €30.5 million ($33.7 million) by the Netherlands’ data...

Become a member

New, Relevant Tech Stories. Our article selection is done by industry professionals. Our writers summarize them to give you the key takeaways