Effective scaling to billion-parameter radiology foundation models
Investigate why the Curia-2 g Vision Transformer with 1.3 billion parameters does not significantly outperform the Curia-2 L Vision Transformer with 303 million parameters and determine training strategies or architectural modifications that resolve the challenge of effectively scaling self-supervised multi-modal CT and MRI foundation models to the billion-parameter regime.
References
Yet, the performance of Curia-2 g remains close to that of Curia-2 L, indicating that the challenge of scaling to billion-parameter models in medical imaging is not yet fully resolved.
— Curia-2: Scaling Self-Supervised Learning for Radiology Foundation Models
(2604.01987 - Saporta et al., 2 Apr 2026) in Conclusion