Does curation prevent model collapse in iterative deployment?

Determine whether the curation of valid traces during iterative deployment of large language models fully prevents model collapse, or whether curation merely delays collapse without eliminating it.

Background

The paper studies an iterative deployment mechanism where each new generation of a LLM is fine-tuned on curated valid traces from previous deployments. This process is theoretically connected to REINFORCE with an implicit binary reward and empirically shown to improve planning capabilities.

A key safety concern discussed is model collapse: prior work has shown that training on models’ own synthetic outputs can shrink distributions and degrade capabilities. The authors highlight that while curation might delay collapse by filtering valid outputs, it remains explicitly unknown whether curation fully prevents collapse in iteratively deployed systems.

References

A final property to note is model collapse \citep{shumailov-et-al-nature2024}. Data curation might delay model collapse by filtering for valid traces, it is still unknown whether this fully prevents collapse.

Iterative Deployment Improves Planning Skills in LLMs  (2512.24940 - Corrêa et al., 31 Dec 2025) in Subsection: Implications to AI Safety