Kaplan & Chinchilla Scaling Laws
- Kaplan and Chinchilla Scaling Laws are empirical and theoretical relationships that predict how test loss in neural language models scales with parameter count, dataset size, and compute cost.
- They provide power-law formulas for optimal resource allocation, guiding the balance between increasing model size and training data to achieve efficient performance improvements.
- Extensions incorporate architecture-specific factors like latency and sparsity, enabling principled LLM design and deployment under practical constraints.
Kaplan and Chinchilla Scaling Laws are a class of empirical and theoretical relationships that characterize how neural LLMs' test loss scales with respect to model parameter count, dataset size, compute cost, and, in modern extensions, architecture-dependent inference characteristics. These scaling laws provide quantitative recipes for allocating compute among model capacity and data, predict training behavior across orders of magnitude in scale, and now underpin principled design of LLM architectures under deployment constraints.
1. Origins and Mathematical Forms of Scaling Laws
The original scaling law formulation by Kaplan et al. (2020) established that trained transformer models exhibit predictable decreases in test cross-entropy loss as a power-law in both parameter size and number of unique training tokens : where , , , , and are empirical constants determined by fitting—, for their original GPT-style models. Compute-optimal model size under the constraint was derived as , , indicating an allocation favoring larger models.
Hoffmann et al. (2022) ("Chinchilla") revised these exponents using an expanded parameter and data regime, and by accounting for total (including embedding) parameter count and compute: Their fitting yields , , , , , and critically, a compute-optimal allocation —implying a balanced "50:50" split between increasing model and data for fixed compute (Pearce et al., 2024, Porian et al., 2024).
Empirical tests and analytic reparametrizations, correcting for embedding layer and small-scale biases, demonstrate that Kaplan's originally steeper exponent is a finite- artifact. Simulating Chinchilla under the original Kaplan conventions replicates the $0.73$ exponent, with both analyses converging to $0.5$ at large scale (Pearce et al., 2024, Porian et al., 2024).
2. Theoretical Explanations and Unification
Contemporary work provides rigorous theoretical underpinnings for these empirical power laws:
- An information-theoretic analysis establishes, in Barron-like single-hidden-layer settings, tight upper bounds whose compute-optimal minimizer fulfills (i.e., a linear scaling) matching Chinchilla's empirical law up to logarithmic corrections (Jeon et al., 2022).
- The "Effective Frontier" framework abstracts task learning as progressive coverage of patterns from a long-tailed (Zipf) distribution. Loss is attributed to the mass of unlearned tail patterns above a cutoff ("Effective Frontier"), which depends on the available resource :
Distinct scaling exponents emerge for , , and training steps . The key result is a Max-Bottleneck principle: when multiple resource constraints apply (e.g., , , ), loss is dictated by the slowest-decaying term
Constrained optimization over , with yields the two limiting regimes: the Kaplan law (compute-limited, ) and the Chinchilla law (data-limited, ), reconciling them as equilibrium solutions of a unified scaling equation (Zou et al., 1 Feb 2026).
3. Practical Fitting, Methodological Refinements, and Calibration
Subsequent analyses identified methodological artifacts impacting exponent estimates and fitting consistency:
- Exclusion of embedding and output layers from measures produced a $0.73$ exponent at small scale; total parameter and compute accounting resolves this to Chinchilla's $0.50$ (Pearce et al., 2024).
- Warmup duration, last-layer computational cost, and optimizer hyperparameter scaling systematically skew power-law estimates. Incorporating per-run, size-adaptive warmup, counting all FLOPs (including head and embedding), and per-model optimizer tuning yields scaling curves with —matching Chinchilla (Porian et al., 2024).
- Out-of-sample accuracy of parametric scaling forms is augmented by robust Huber-loss or L-BFGS fitting. Nonparametric ML regression (e.g., kernel or neural network surface for ) further improves frontier estimation in some empirical settings (Barkeshli et al., 15 Jan 2026).
4. Extensions: Architecture, Latency, and Sparsity
Modern scaling law frameworks explicitly incorporate architecture- and efficiency-aware factors:
- Latency and memory bandwidth: Empirical runtime is much better predicted by the volume of memory copy operations (dominant term in accelerator-bound environments) than by pure FLOP count. The closed-form throughput and loss prediction as a function of transformer hyperparameters allows analytic architecture optimization for fixed wall-clock time budgets, pointing to wide–shallow configurations as preferable for a given parameter count (Inbar et al., 2024).
- Inference cost: Shape-aware scaling laws co-optimize , , and model "aspect ratio" , with loss penalized as
These forms enable Pareto-efficient inference-optimized model selection and demonstrate that for a fixed , wider, shallower models achieve up to faster inference at identical task accuracy, confirming model-shape-dependence in scaling (Bian et al., 30 Jan 2025).
- Conditional scaling over architecture: Parameterizing loss exponents as functions of hidden size and MLP/attention ratio,
yields models (e.g., "Panda" and "Surefire") that outperform baselines both in terms of loss and inference throughput (Bian et al., 21 Oct 2025).
- Sparsity: The "average parameter count" over pre-training, , replaces static in the Chinchilla law, unifying sparse and dense pre-training in a single scaling framework. Empirical results confirm that for matched and , "lossless" compression at up to sparsity yields no training or downstream accuracy deficit, decoupling training-time compute from inference speed (Jin et al., 21 Jan 2025).
5. Origin and Robustness of Scaling Laws
Recent theoretical and synthetic experiments reinforce the universality and origin of scaling laws:
- Scaling exponents' values are not reducible solely to data distribution tails; robust power-law scaling is observed in synthetic settings devoid of Zipf structure (e.g., transformers on random walks), implying an architectural emergence of power-law spectra or optimization effects (Barkeshli et al., 15 Jan 2026).
- Data complexity, as modulated via dataset generative process or architecture parameterization, alters the data exponent , while the parameter exponent typically remains stable (around $0.5$–$0.76$ in 2-layer transformers and task-agnostic settings).
- The necessity of accounting for irreducible offsets is emphasized for fit validity. One-dimensional power-law fits, including such offsets, outperform both fixed functional forms and exponentials in predictive accuracy.
6. Empirical Design Implications and Current Best Practices
Compiled results across multiple studies yield convergent recipes for LLM design under budget:
- Use total (including embeddings) parameter and compute counts in all scaling law reporting (Pearce et al., 2024, Porian et al., 2024).
- For compute-optimal training, allocate compute such that for sufficiently large, ensuring the "Chinchilla frontier" is approached (Pearce et al., 2024).
- For practical deployment, augment Chinchilla or related scaling laws by incorporating shape (aspect ratio) or conditional architectural parameters to optimize for latency-constrained inference, with wider–shallower models yielding substantial efficiency gains for a fixed (Inbar et al., 2024, Bian et al., 30 Jan 2025, Bian et al., 21 Oct 2025).
- In sparse pre-training, the effective parameter controlling train loss is the arithmetic mean ; use the average-parameter version of the scaling law for both dense and sparse regimes (Jin et al., 21 Jan 2025).
- For high-complexity or high-dimensional input, theories predict that more of the budget should be channeled into increasing rather than ; this is supported by both analytic upper bound derivations and empirical sweeps (Jeon et al., 2022).
- Nonparametric frontier fitting and sensitivity analysis over (, ) allow for more robust extrapolation at new scales and in atypical regimes (Barkeshli et al., 15 Jan 2026).
The combined legacy of Kaplan and Chinchilla scaling laws is a rigorously grounded, empirically validated, and now architecture-aware methodology for predicting LLM scaling and guiding efficient allocation of resources in pretraining and deployment. Extensions to account for architectural details, memory, and inference cost have established that contemporary model selection and scaling optimization can be performed entirely with closed-form, hyperparameter-only loss models (Inbar et al., 2024, Bian et al., 30 Jan 2025, Bian et al., 21 Oct 2025).