Degenerative AI

Degenerative AI & What Happens When AI Trains on AI Data: Artificial Intelligence Trends

AI models need so much data that they are beginning to be trained on AI-generated data. An article shows how this could lead to “degenerative AI”.

According to The New York Times (When A.I.’s Output Is a Threat to A.I. Itself, written by Aatish Bhatia and available here), Sam Altman of OpenAI wrote in February that the company generated about 100 billion words per day — a million novels’ worth of text, every day, an unknown share of which finds its way onto the internet.

A.I.-generated text may show up as a restaurant review, a dating profile or a social media post. And it may show up as a news article, too: NewsGuard, a group that tracks online misinformation, recently identified over a thousand websites that churn out error-prone A.I.-generated news articles.

Advertisement
Syllo

All this A.I.-generated information can make it harder for us to know what’s real. And it also poses a problem for A.I. companies. As they trawl the web for new data to train their next models on — an increasingly challenging task — they’re likely to ingest some of their own A.I.-generated content, creating an unintentional feedback loop in which what was once the output from one A.I. becomes the input for another.

In the long run, this cycle may pose a threat to A.I. itself. Research has shown that when generative A.I. is trained on a lot of its own output, it can get a lot worse. Here are a couple of illustrations of that (based on research by Ilia Shumailov and others):

This exercise illustrates how training of AI to mimic handwritten digits by training the AI on previous AI outputs causes all the digits to converge into a single shape after 30 generations:

Degenerative AI

This exercise shows how training of AI by the researchers by asking it to complete a sentence that started with “To cook a turkey for Thanksgiving, you…,” began to repeat phrases incoherently after just four generations:

Advertisement
Nextpoint
Degenerative AI

Another team of researchers at Rice University studied what would happen when the kinds of A.I. that generate images are repeatedly trained on their own output — a problem that could already be occurring as A.I.-generated images flood the web.

They found that glitches and image artifacts started to build up in the A.I.’s output, eventually producing distorted images with wrinkled patterns and mangled fingers.

And even after addressing that issue, another one emerged – a lack of diversity in AI generated images:

From this (which already seems to be notably lacking some ethnic groups):

To this, after four generations, where the faces all appeared to converge:

This “degenerative AI” not only affects the quality of the output, but it also affects energy and costs as well. Computer scientists at N.Y.U. found that when there is a lot of A.I.-generated content in the training data, it takes more computing power to train A.I. — which translates into more energy and more money.

“Models won’t scale anymore as they should be scaling,” said ​​Julia Kempe, the N.Y.U. professor who led this work.

The leading A.I. models already cost tens to hundreds of millions of dollars to train, and they consume staggering amounts of energy, so this can be a sizable problem.

Potential solutions include for AI companies to pay for the data they are training on to ensure quality levels – which may improve the quality but limit the parameters (not to mention that they are currently using a lot of data for free and litigating over what they state is the “fair use” of it) – and the use of “watermarking” to detect AI content (which can’t always be reliably detected and can be easily subverted).

This is certainly a potential issue which could jeopardize the potential improvement of AI models over time. The age-old expression of “garbage in, garbage out” is more significant than ever when it comes to the advancement of AI models.

So, what do you think? Are you concerned about the potential for “degenerative AI”? Please share any comments you might have or if you’d like to know more about a particular topic.

Image Copyright © Columbia Pictures Some of you will “see what I did there”! 😉

Disclaimer: The views represented herein are exclusively the views of the authors and speakers themselves, and do not necessarily represent the views held by my employer, my partners or my clients. eDiscovery Today is made available solely for educational purposes to provide general information about general eDiscovery principles and not to provide specific legal advice applicable to any particular circumstance. eDiscovery Today should not be used as a substitute for competent legal advice from a lawyer you have retained and who has agreed to represent you.


Discover more from eDiscovery Today by Doug Austin

Subscribe to get the latest posts sent to your email.

5 comments

Leave a Reply