Self-Consuming AI Resources

Advances in generative artificial intelligence (AI) algorithms for text, imagery, and other data types have led to the temptation to use AI-synthesized data to train next-generation models. Repeating this process creates a self-consuming loop whose properties are poorly understood. Unfortunately, recent research has shown that repeated training with synthetic data forms a self-consuming feedback loop that causes the model distribution to drift away from reality, reinforcing biases, amplifying artifacts, and lowering the quality and diversity of next-generation models, a phenomena often referred to as model collapse or model autophagy disorder (MAD).

The goal of this webpage is to collect and help organize the growing literature on AI self-consuming feedback loops. To post new links or correct existing links, please email selfconsumingAI@gmail.com

Foundations

Mitigation & Prevention

Fairness & Bias

Social Implications

Synthetic Data and Downstream Tasks

Selected Media Coverage