Skip to main content

artificial intelligence collapse risks

AI training of AI in LLMs may result in model collapse, researchers suggest

AI training of AI in LLM
A study published in Nature warns that using AI-generated datasets to train subsequent machine learning models may lead to model collapse, polluting their outputs. The research indicates that, after a few generations, original content is supplanted by unrelated gibberish, underscoring the necessity of reliable data for AI training.

Generative AI tools, including Large Language Models (LLMs), have gained widespread popularity, primarily being trained on human-generated inputs. However, as these AI models become more prevalent on the internet, there is a risk of computer-generated content being used to train other AI models, or even themselves, in a recursive manner.

Ilia Shumailov and his team have developed mathematical models to illustrate the phenomenon of model collapse in AI systems. Their research shows that AI models may disregard certain outputs, such as infrequent lines of text in training data, leading to self-training on a limited subset of the dataset.

Shumailov and his team examined the responses of AI models to training datasets primarily generated by artificial intelligence. Their findings reveal that using AI-generated data leads to a degradation in learning capabilities over successive generations, culminating in model collapse.

The majority of recursively trained language models analyzed showed a pattern of generating repetitive phrases. As an example, when medieval architecture text was used as the initial input, the ninth generation's output consisted of a list of jackrabbits.

According to the authors, model collapse is an inevitable result of using training datasets produced by earlier generations of AI models. They suggest that successful training with AI-generated data is possible if stringent data filtering measures are implemented.

Simultaneously, firms leveraging human-produced content for AI training could develop models that outperform those of their rivals.

Further detail: In the paper 'AI Models Collapse When Trained on Recursively Generated Data,' Ilia Shumailov et al., Nature, 2024.

Source

Comments

Popular posts from this blog

NASA chile scientists comet 3i atlas nickel mystery

NASA and Chilean Scientists Study 3I/ATLAS, A Comet That Breaks the Rules Interstellar visitors are rare guests in our Solar System , but when they appear they often rewrite the rules of astronomy. Such is the case with 3I/ATLAS , a fast-moving object that has left scientists puzzled with its bizarre behaviour. Recent findings from NASA and Chilean researchers reveal that this comet-like body is expelling an unusual plume of nickel — without the iron that typically accompanies it. The discovery challenges conventional wisdom about how comets form and evolve, sparking both excitement and controversy across the scientific community. A Cosmic Outsider: What Is 3I/ATLAS? The object 3I/ATLAS —the third known interstellar traveler after "Oumuamua (2017) and 2I/Borisov (2019) —was first detected in July 2025 by the ATLAS telescope network , which scans he skies for potentially hazardous objects. Earlier images from Chile's Vera C. Rubin Observatory had unknowingly captured it, but ...

Quantum neural algorithms for creating illusions

Quantum Neural Networks and Optical Illusions: A New Era for AI? Introduction At first glance, optical illusions, quantum mechanics, and neural networks may appear unrelated. However, my recent research in APL Machine Learning Leverages "quantum tunneling" to create a neural network that perceives optical illusions similarly to humans. Neural Network Performance The neural network I developed successfully replicated human perception of the Necker cube and Rubin's vase illusions, surpassing the performance of several larger, conventional neural networks in computer vision tasks. This study may offer new perspectives on the potential for AI systems to approximate human cognitive processes. Why Focus on Optical Illusions? Understanding Visual Perception O ptical illusions mani pulate our visual  perce ption,  presenting scenarios that may or may not align with reality. Investigating these illusions  provides valuable understanding of brain function and dysfunction, inc...

fractal universe cosmic structure mandelbrot

Is the Universe a Fractal? Unraveling the Patterns of Nature The Cosmic Debate: Is the Universe a Fractal? For decades, cosmologists have debated whether the universe's large-scale structure exhibits fractal characteristics — appearing identical across scales. The answer is nuanced: not entirely, but in certain res pects, yes. It's a com plex matter. The Vast Universe and Its Hierarchical Structure Our universe is incredibly vast, com prising a p proximately 2 trillion galaxies. These galaxies are not distributed randomly but are organized into hierarchical structures. Small grou ps ty pically consist of u p to a dozen galaxies. Larger clusters contain thousands, while immense su perclusters extend for millions of light-years, forming intricate cosmic  patterns. Is this where the story comes to an end? Benoit Mandelbrot and the Introduction of Fractals During the mid-20th century, Benoit Mandelbrot introduced fractals to a wider audience . While he did not invent the conce pt —...