Papers
Topics
Authors
Recent
Search
2000 character limit reached

Theoretical Perspectives on Data Quality and Synergistic Effects in Pre- and Post-Training Reasoning Models

Published 1 Mar 2026 in cs.LG, cs.AI, and stat.ML | (2603.01293v1)

Abstract: LLMs are pretrained on massive datasets and later instruction-tuned via supervised fine-tuning (SFT) or reinforcement learning (RL). Best practices emphasize large, diverse pretraining data, whereas post-training operates differently: SFT relies on smaller, high-quality datasets, while RL benefits more from scale, with larger amounts of feedback often outweighing label quality. Yet it remains unclear why pretraining and RL require large datasets, why SFT excels on smaller ones, and what defines high-quality SFT data. In this work, we theoretically analyze transformers trained on an in-context weight prediction task for linear regression. Our analysis reveals several key findings: $(i)$ balanced pretraining data can induce latent capabilities later activated during post-training, and $(ii)$ SFT learns best from a small set of examples challenging for the pretrained model, while excessively large SFT datasets may dilute informative pretraining signals. In contrast, RL is most effective on large-scale data that is not overly difficult for the pretrained model. We validate these theoretical insights with experiments on large nonlinear transformer architectures.

Summary

  • The paper demonstrates that small, carefully curated SFT datasets aligned with pretraining gaps yield optimal performance, while excessive scaling causes interference.
  • It models pretraining as population-level in-context learning with linear self-attention and derives precise error bounds and convergence rates for SFT and RL paradigms.
  • Empirical validations using LSA and GPT-2 confirm the theoretical predictions, emphasizing balanced data curation and the need for diversified outcome supervision.

Theoretical Analysis of Data Quality and Synergistic Effects in LLM Pre- and Post-Training

Motivation and Conceptual Framework

The paper investigates the theoretical underpinnings of data quality, scaling, and interaction dynamics between pretraining and post-training for LLMs, specifically focusing on reasoning ability and in-context learning. While empirical best practices dictate massive, diverse pretraining and careful selection of post-training data, there has been limited rigorous understanding of why these protocols yield certain behaviors. The authors bridge this gap by studying transformer models on an in-context weight prediction task for linear regression, enabling tractable yet expressive analysis of transformer learning dynamics.

Key facets delineated in the work include:

  • Latent capabilities induced during pretraining and subsequently activated by post-training.
  • Contrasts between process supervision (SFT) and outcome supervision (RL/RLHF) from an optimization and generalization perspective.
  • Precise characterization of "hard" and "informative" examples and the role of interference in post-training.

Formal Pretraining and Post-Training Modeling

Pretraining is modeled as population-level in-context learning such that the transformer acquires parameters encoding the statistics of the input distribution. The theoretical analysis leverages a linear self-attention (LSA) abstraction to elucidate the role of empirical covariances and optimal weight initializations.

Post-training is decomposed into two major paradigms:

  • Supervised Fine-Tuning (SFT): The model receives ground-truth chain-of-thought intermediate supervision, which minimizes KL divergence and encourages coverage over plausible reasoning trajectories.
  • Outcome Supervision (OS / RL): The model is optimized only for final reasoning outcomes, with the RL loss landscape typically exhibiting sharp curvature and sensitivity to distributional alignment.

The authors present formal loss functions and recurrence relations for both SFT and RL, analyzing convergence, sparsity preservation, and adaptation under covariate shift in input statistics. Figure 1

Figure 1

Figure 1

Figure 1: Optimal SFT data selection minimizes post-test error in the absence of distributional interference (r=0r=0).

Analytical Results: SFT vs. RL/RLHF Optimization Landscapes

Theoretical results reveal several novel and empirically validated insights:

  • SFT learns most efficiently from a small collection of examples that the pretrained model finds difficult. Excessive SFT data volume introduces detrimental interference, eroding the latent structure acquired during pretraining.
  • The optimal SFT allocation requires data concentrated in directions poorly covered by the pretraining covariance, aligning with adaptation shifts.
  • RL benefits from scale: stability and effective adaptation are achieved only when outcome supervision datasets are large and not overly difficult relative to the pretrained model's capabilities. The RL loss landscape is sharply curved, demanding careful learning rates and substantial data diversity.

Theoretical error bounds, convergence rates, and population regime asymptotics substantiate these claims. The paper analytically demonstrates double descent phenomena and catastrophic error increases when SFT dataset size approaches model dimensionality in the presence of distributional interference. Figure 2

Figure 2

Figure 2

Figure 2: Post-training error escalates with interference; increasing SFT prompts (BB) past optimal causes pronounced degradation.

Data Selection, Scaling, and Quality Trade-offs

Optimal data selection for SFT is derived analytically by minimizing post-test error through covariance alignment. The results concretize empirical intuition: high-quality SFT data should be hard and diverse, targeting latent gaps left by pretraining. Scaling up SFT data—either in number of prompts (BB) or prompt length (nn)—increases the risk of interference, leading to performance degradation.

Population and mean-field analyses, leveraging random matrix theory, provide closed-form and asymptotic formulas for test error as a function of data parameters, covariance structure, and adaptation shift. Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3

Figure 3: Supervised fine-tuning performance: increasing BB or nn initially decreases error, but excessive scaling leads to reversal due to interference.

Empirical Validation with LSA and GPT-2

Experimental results on linear self-attention and nonlinear GPT-2 architectures confirm theoretical predictions:

  • For SFT, optimal performance is obtained at small, carefully selected datasets aligned with adaptation directions. Scaling SFT data volume beyond this point increases test error, corroborating theoretical double descent and interference phenomena.
  • RL/OS training always benefits from larger volume and diversity of outcome-supervised data, with stability and robustness emerging only at large scale. Figure 4

Figure 4

Figure 4

Figure 4

Figure 4

Figure 4

Figure 4: Supervised fine-tuning in GPT-2: test loss minima are attained at modest SFT dataset sizes; further scaling induces error increases.

Figure 5

Figure 5

Figure 5: Test error as a function of prompt length (n=1000n=1000): optimal SFT scaling minimizes post-test error, but overparameterization yields U-shaped trends.

Practical and Theoretical Implications

The findings imply critical guidance for LLM development:

  • Pretraining should emphasize spectral balance and broad distributional coverage, as latent generalization is strictly bottlenecked by pretraining data diversity.
  • SFT must be curated to cover gaps and challenging cases, avoiding over-saturation that diminishes prior structure.
  • RL post-training is effective only when preceded by balanced pretraining and executed with massive, diverse feedback data to mitigate instability from sharp loss landscape curvature.

The synergy between pretraining and post-training is shown to be fundamental: SFT activates latent reasoning abilities via challenging examples, while RL serves as scale-intensive refinement, particularly when the pretrained model already possesses partial alignment in the target adaptation directions.

Future developments will likely focus on principled data allocation, leveraging theoretical error expansions and mean-field approaches for optimal pretrain/post-train synergy, advancing robustness in reasoning models.

Conclusion

This paper provides rigorous theoretical and empirical insight into the interaction dynamics between pretraining and post-training data selection in reasoning-centric LLMs. The results formalize key empirical phenomena—such as the necessity of balanced pretraining, the optimality of small, high-quality SFT datasets, and the scale dependence of RL—guided by precise error bounds and optimization landscape analysis. These findings underpin principled data curation and training paradigms for both efficiency and robustness in advanced LLMs (2603.01293).

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 26 likes about this paper.