CHGNet: Universal ML Interatomic Potential
- CHGNet is a charge-informed graph neural network-based force field pretrained on extensive DFT data to approximate the universal potential energy surface for inorganic materials.
- Its architecture employs message-passing with E(3)-equivariant updates that accurately capture energies, forces, stresses, and magnetic moments.
- Fine-tuning with targeted high-energy DFT snapshots mitigates systematic PES softening, enabling robust predictions in active learning and high-throughput materials screening.
The CHGNet Universal Machine Learning Interatomic Potential (uMLIP) is a charge-informed, graph neural network-based force field designed to approximate the universal potential energy surface for inorganic crystalline materials. Pretrained on an extensive dataset of density functional theory (DFT) energies, forces, stresses, and magnetic moments from the Materials Project, CHGNet aims to combine domain transferability across chemistries with near-DFT accuracy in predicting static and dynamic properties relevant to materials design, phase stability, ion transport, and defect physics (Deng et al., 2023, Lian et al., 3 Jul 2025).
1. Model Architecture and Representational Formalism
CHGNet encodes atomic configurations as undirected graphs , where is a set of atomic nodes and is a set of edges connecting neighbors within a cutoff (typically 5–6 Å). Node features are learnable embeddings of atomic number (one-hot or continuous), optionally extended to include atomic volume or site magnetic moment if available. Edge features are formed by radial basis expansions of the interatomic distance ; advanced variants include angular triplet features to capture three-body correlations (Deng et al., 2023, Huang et al., 7 Apr 2025, Pitfield et al., 24 Jul 2025).
The message-passing block comprises layers. For layer 0:
- Message aggregation: 1
- Node update: 2 where the 3 functions are small feed-forward neural networks, which can be E(3)-equivariant (ensuring rotational equivariance in forces).
After 4 passes, each atom’s embedding 5 is mapped by a readout MLP 6 to a per-atom energy, with total energy given by
7
Atomic forces are obtained by gradient backpropagation:
8
The same differentiated model yields virial stress; for models with charge-informed nodes, on-site magnetic moments 9 are predicted from intermediate features via a dedicated linear head, acting as a proxy for local charge or valence state (Deng et al., 2023).
2. Pretraining Dataset and Loss Functions
CHGNet is pretrained on the Materials Project Trajectory Dataset (MPtrj), comprising 0 million DFT labels covering 89–94 elements, drawn from static and relaxation trajectories of 1146,000 distinct compounds. These include total energies, atomic forces, cell stresses, and magnetic moments in spin-polarized runs (Deng et al., 2023, Zhou et al., 2024, Huang et al., 7 Apr 2025).
The multi-task training objective is a weighted sum (usually mean squared error) over:
- Total energy per atom
- Atomic forces
- Virial stress
- Magnetic moments
Weighting is typically 2 (Huang et al., 7 Apr 2025). Optimization employs Adam or RAdam with initial learning rates around 3, batch sizes 4–5, and early stopping on a hold-out set. For computational efficiency, FastCHGNet accelerates the reference implementation via fused CUDA kernels, direct force/stress heads, and batch-parallel basis generation, reducing multi-GPU pretraining time from 8.3 days (A100) to 1.53 h (32×A100) while cutting memory by %%%%2627%%%% (Zhou et al., 2024).
3. Universality, Systematic Softening, and Fine-Tuning Strategies
CHGNet's universal character derives from extensive element coverage, cross-chemistry pretraining, and rotationally equivariant representations. The resulting model is directly transferable to out-of-sample inputs, enabling zero-shot prediction for previously unseen crystals, surfaces, or defected supercells (Deng et al., 2023, Deng et al., 2024, Lian et al., 3 Jul 2025).
However, models pretrained on equilibrium-centric data develop a systematic "softening" of the potential energy surface (PES): a global underestimation of PES curvature and energy barriers, manifesting as slopes 81 in parity plots of ML vs DFT forces, low phonon frequencies, and underestimated surface/defect/migration energies (Deng et al., 2024). This softening is formally a near-uniform rescaling in PES curvature across off-equilibrium configurations.
The bias can be dramatically reduced by fine-tuning the pretrained weights with a modest dataset—sometimes a single high-energy (e.g., transition-state or defect) DFT snapshot suffices. The protocol incorporates additional DFT labels via transfer learning (e.g., 1–100 structures), using the same multi-task loss and optimizer but with all or selected layers unfrozen (Deng et al., 2024, Žguns et al., 10 Sep 2025, Lian et al., 3 Jul 2025). In practice, fine-tuning on targeted high-energy configurations "re-anchors" the PES in the relevant region, yielding dramatic reductions in MAE and restoring parity slopes close to unity.
4. Validation: Performance Benchmarks and Applications
4.1. Bulk and Surface Structure Energetics
In bulk and relaxed structures, CHGNet attains MAE ≈ 30 meV/atom (energy), 77 meV/Å (force), and 0.35 GPa (stress) in held-out tests (Deng et al., 2023, Focassio et al., 2024). On surface calculations, zero-shot surface energy RMSE is 90.51 J/m0 but is systematically underestimated due to training set bias; fine-tuning is recommended for accurate work on surfaces (Focassio et al., 2024, Deng et al., 2024).
4.2. NEB Barrier Prediction and Li-Ion Conductors
Fine-tuned CHGNet reduces migration-barrier MAE from 0.23–0.24 eV (pretrained) to 0.07–0.09 eV and 1 from 0.94 to 0.98 on test and training sets. This enables reliable high-throughput NEB-based screening of candidate Li-ion conductor frameworks, achieving 100–10002 wall-clock speedup over DFT-based NEBs (Lian et al., 3 Jul 2025, Kang et al., 14 Aug 2025). For example, aliovalent-doped variants Li3Mg4Al5PO6 and Li7TiPO8F9 were predicted with fine-tuned CHGNet to have low migration barriers and conductivities of 0.19 mS/cm, 0.024 mS/cm respectively.
4.3. Molecular Dynamics and Transport
Long-time molecular dynamics (MD) using CHGNet yields force and dynamical accuracy at the ab-initio level: e.g., 200–20000 speedup for 101–102-step trajectories in superionic conductors, and accurate Arrhenius behavior for Li/Na/Mg diffusion. When deployed as-is, universal CHGNet systematically underpredicts diffusion coefficients in defective materials (e.g., for H in Mg), but fine-tuning on a few thousand targeted ab-initio frames restores DFT-level transport and activation energies (Angeletti et al., 2024).
4.4. Alloy Mixing and Thermodynamics
In alloy theory, out-of-the-box CHGNet predictions for mixing enthalpy are typically accurate to 340–50 meV/atom RMSE, which does not suffice for thermodynamic topology (correct sign and relative ordering of phases). Augmenting the universal CHGNet with tens of DFT-labeled SQS supercells at key compositions enables fine-tuned models to reach the 410 meV/atom regime needed for reliable phase diagram prediction (Casillas-Trujillo et al., 2024, Zhu et al., 2024).
5. Integration with Active Learning and Global Optimization
CHGNet serves as the backbone for active-learning pipelines combining foundation uMLIP prediction with local correction via a 5-GPR (Gaussian Process Regression) model using Smooth Overlap of Atomic Positions (SOAP) descriptors (Pitfield et al., 24 Jul 2025). The workflow:
- Propose trial structures (by RSS, Basin Hopping, GOFEE, Replica Exchange).
- Relax with current CHGNet+6 model.
- Select promising structures for DFT evaluation.
- Compute 7, add to GPR training set.
- Retrain the GPR correction; iterate.
The 8 model rectifies systematic errors in the universal backbone, accelerates global structure searches, and is particularly effective when the base uMLIP misorders global minima or mispredicts energetics in new chemical environments.
6. Limitations, Generalization Scope, and Future Directions
While CHGNet provides wide across-chemistry coverage and near-ab-initio accuracy after modest fine-tuning, its limitations stem from:
- Pronounced PES softening on surfaces, defects, and high-energy states absent from the training set (Deng et al., 2024, Lian et al., 3 Jul 2025).
- Lack of explicit electrostatics and van der Waals interactions in the baseline model (user-correctable via additive corrections).
- Incomplete transferability to liquid, highly disordered, or strongly correlated (e.g., 9-electron) systems unless those states enter training/fine-tune sets.
- In alloy/CALPHAD applications, formation-energy errors of 30–100 meV/atom can qualitatively upset phase equilibria; only sub-10 meV/atom errors eliminate this risk (Zhu et al., 2024).
Future best practices include:
- Systematic expansion of the pretraining dataset to include high-energy (off-equilibrium) configurations, transition states, defects, and free surfaces (Deng et al., 2024, Žguns et al., 10 Sep 2025).
- Routine use of ultra-efficient fine-tuning (minimum 01–100 targeted DFT labels) before production in new compositional/structural regimes.
- Hybridization with E(3)-equivariant architectures and protocolized active learning to ensure coverage of critical phase-space regions, particularly for multi-phase or multi-valent systems.
7. Summary Table: Key Model Features and Performance Metrics
| Property | Universal (Pretrained) CHGNet | Fine-Tuned CHGNet |
|---|---|---|
| Energy MAE (bulk/test) | 30 meV/atom (Deng et al., 2023) | 2 meV/atom (TS) (Lian et al., 3 Jul 2025) |
| Force MAE | 77 meV/Å (Deng et al., 2023) | 13 meV/Å (TS) (Lian et al., 3 Jul 2025) |
| MD/NEB Speedup vs DFT | 1100–10002 (Lian et al., 3 Jul 2025) | 3100–10004 (Lian et al., 3 Jul 2025) |
| Surface Energy RMSE | 0.51 J/m5 (Focassio et al., 2024) | Reduced after fine-tuning (Deng et al., 2024) |
| Alloy 6 | 41.5 meV/atom (Casillas-Trujillo et al., 2024) | <10 meV/atom (with retraining) |
| Migration Barrier MAE | 0.23–0.24 eV (Lian et al., 3 Jul 2025) | 0.07–0.09 eV (Lian et al., 3 Jul 2025) |
| Transfer to new chemistry | High (across MP elements) | High (if covered in fine-tune) |
Fine-tuning with a minimal set of high-energy DFT structures systematically eliminates the curvature softening endemic to universal pretraining, restores DFT-level accuracy, and enables robust, data-efficient adaptation of CHGNet to challenging regions of the potential energy surface critical for materials discovery (Deng et al., 2024, Žguns et al., 10 Sep 2025, Lian et al., 3 Jul 2025).