Papers
Topics
Authors
Recent
Search
2000 character limit reached

Goldilocks RL: Tuning Task Difficulty to Escape Sparse Rewards for Reasoning

Published 16 Feb 2026 in cs.LG and cs.AI | (2602.14868v1)

Abstract: Reinforcement learning has emerged as a powerful paradigm for unlocking reasoning capabilities in LLMs. However, relying on sparse rewards makes this process highly sample-inefficient, as models must navigate vast search spaces with minimal feedback. While classic curriculum learning aims to mitigate this by ordering data based on complexity, the right ordering for a specific model is often unclear. To address this, we propose Goldilocks, a novel teacher-driven data sampling strategy that aims to predict each question's difficulty for the student model. The teacher model selects questions of appropriate difficulty for the student model, i.e., questions that are neither too easy nor too hard (Goldilocks principle), while training the student with GRPO. By leveraging the student's performance on seen samples, the teacher continuously adapts to the student's evolving abilities. On OpenMathReasoning dataset, Goldilocks data sampling improves the performance of models trained with standard GRPO under the same compute budget.

Summary

  • The paper introduces Goldilocks RL, a dynamic teacher-student framework that selects tasks of optimal difficulty to maximize the learning signal in reasoning models.
  • It demonstrates significant improvements in sample efficiency and validation accuracy across various model scales compared to standard GRPO approaches.
  • The framework reduces wasted compute by focusing on prompts near the decision frontier, where reward variance is highest, ensuring effective use of training data.

Goldilocks RL: A Teacher-Driven Curriculum for Efficient Reinforcement Learning on Reasoning Tasks

Motivation and Context

Addressing the computational inefficiency of reward-sparse RL for reasoning in LLMs, the paper "Goldilocks RL: Tuning Task Difficulty to Escape Sparse Rewards for Reasoning" (2602.14868) formalizes and empirically investigates the intersection between curriculum learning and outcome-based RL. The authors identify key deficiencies in conventional curriculum strategies: history-based schedules fail to generalize and are computationally impractical in web-scale RL, while category-based curricula mischaracterize instance-level difficulty. Goldilocks RL introduces a zero-shot utility estimator—a teacher trained online to select tasks of optimal difficulty for the current capabilities of a reasoning model, thereby maximizing the RL learning signal and avoiding wasted compute on trivial or infeasible examples.

Goldilocks Framework: Theory and Implementation

The Goldilocks architecture is a two-agent framework: a dynamically trained teacher, and a student model fine-tuned via RL. The student executes RL updates (GRPO or variants) on data scheduled by the teacher, while the teacher is updated asynchronously to predict the mean reward uncertainty (learning potential) per prompt. Unlike prior CL schemes, the teacher makes selection decisions in a streaming fashion without requiring repeated student exposures or external metadata. Figure 1

Figure 1: Overview of the Goldilocks Framework. The training cycle proceeds as follows: (1) KcandidateK_{\text{candidate}} questions are sampled; (2) the Teacher selects the optimal one; (3) the Student generates rollouts; (4) Student updates via GRPO; (5) Teacher targets are computed; (6) Teacher updates asynchronously.

The framework exploits a theoretical insight: with outcome supervision and group-relative policy optimization, the policy gradient's norm is proportional to pq(1−pq)\sqrt{p_q(1-p_q)}, where pqp_q is the probability the student solves the prompt. Maximum gradient is achieved on prompts near the "decision frontier" (pq≈0.5p_q \approx 0.5), while confidently solved or intractable tasks contribute zero signal and inhibit learning progress. Goldilocks RL formalizes this by explicitly training the teacher to predict and select prompts with high reward variance for the student.

Technical Details

The student is a Transformer LM (e.g., Qwen2.5, Phi-4, Olmo2) trained via GRPO. Rollouts per prompt are grouped, and rewards decomposed into format compliance and final correctness (binary). The teacher, parameterized as a lightweight LM with a head, predicts the empirical standard deviation of group reward as a utility proxy, constraining the range by a scaled sigmoid. A candidate set of KK prompts is periodically drawn; using ϵ\epsilon-greedy, the teacher selects the one with maximal predicted utility (exploitation) or at random (exploration).

Rollout feedback is used to compute empirical success rates and to update a replay buffer, from which the teacher's regression head is periodically updated on the MSE loss targeting observed grad norms.

Empirical Evaluation

Experiments are conducted on OpenMathReasoning (Moshkov et al., 23 Apr 2025) with >3M chain-of-thought math problems, using several model scales and architectures. Baselines are standard GRPO-trained policies under identical global batch size and compute budgets. All key experiments normalize training steps by effective policy device allocation, controlling for total compute.

Tabled results demonstrate consistent improvement in sample efficiency and final validation accuracy across student scales with Goldilocks:

  • Olmo2-1B: Baseline 11.7% → Goldilocks 14.9%
  • Qwen2.5-1.5B: Baseline 30.6% → Goldilocks 33.4%
  • Qwen3-4B: Baseline 58.1% → Goldilocks 59.7%
  • Phi-4-mini-instruct: Baseline 37.1% → Goldilocks 41.0%

Validation accuracy evolves faster and saturates at higher levels for Goldilocks-tuned models, with particularly pronounced efficiency gains at fixed compute. Figure 2

Figure 2: Evolution of validation accuracy over training steps.

Goldilocks students consistently attain higher training accuracy and learning rates versus baseline GRPO, indicating more frequent optimizations on non-trivial, frontier tasks. Figure 3

Figure 3: Average Training Reward (Success Rate). Goldilocks achieves higher training accuracy significantly earlier.

Goldilocks maintains substantially higher reward variance throughout training, suppressing the fraction of zero-variance (all correct/all wrong) batches (thus avoiding zero gradients). Figure 4

Figure 4

Figure 4: Training Reward Std.

Figure 5

Figure 5: Fraction of questions yielding zero reward variance. Goldilocks reduces "wasted" compute on uninformative samples.

The increased gradient norm during optimization is empirically confirmed, supporting the claim that Goldilocks focuses compute on maximally informative samples. Figure 6

Figure 6: Optimization Dynamics. Larger gradient norms under Goldilocks enhance learning signal and prevent stagnation.

Teacher Model Analysis

The online-trained teacher rapidly adapts to the student's shifting capability profile, maintaining low error on unseen candidate difficulty estimation. Figure 7

Figure 7: Teacher Mean Absolute Error (MAE) on unseen samples.

The predicted difficulty distribution adapts as training progresses, with the mean difficulty selected decreasing as regions of the curriculum are mastered. Figure 8

Figure 8: Evolution of Teacher Predictions. Prediction mean and variance track model proficiency on the task distribution.

Ablations and Robustness

Goldilocks RL's gains are robust to advanced loss formulations (e.g., DAPO), and entropy-regularized objectives, confirming that scheduling non-trivial, high-potential samples provides improvements orthogonal to specific RL loss functions.

Practical and Theoretical Implications

The zero-shot, instance-level scheduling mechanism demonstrates strong generalization: the teacher rapidly infers frontier difficulties on previously unseen data, without resource-draining revisitations. This property renders Goldilocks immediately applicable to massive corpora in both finetuning and potentially pretraining, as the teacher-student cycle can filter and prioritize trillion-token streams for high-utility optimization.

Theoretically, the direct connection between the outcome reward variance and the RL gradient evokes future work on explicit signal-maximizing curricula, scalable resource allocation for multi-agent student-teacher systems, and more fine-grained utility modeling as policy complexity grows. The approach is compatible with further scheduling refinements—such as adaptive candidate set composition, model-based exploration in the data pool, or integration with process-based feedback mechanisms.

Conclusion

Goldilocks RL pioneers efficient RL-based reasoning model scaling by combining outcome-variance-driven sample scheduling with a practical online teacher-student system. It achieves consistently higher data efficiency and performance, demonstrating the value of dynamically tuning task difficulty to match model capability—thereby maximizing the learning signal and compute utilization during RL finetuning for challenging reasoning tasks.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We found no open problems mentioned in this paper.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 2 tweets with 34 likes about this paper.