Low-dimensionality of cores for complex multi-step reasoning
Determine whether low-dimensional algorithmic core subspaces persist for complex multi-step reasoning tasks in transformer-based language models, as opposed to only simpler settings such as Markov-chain prediction, modular addition, and subject–verb agreement.
References
First, whether cores remain low-dimensional for complex multi-step reasoning tasks is untested -- though subject--verb agreement cores remain one-dimensional across GPT-2 Small, Medium, and Large despite a 6.6-fold increase in parameters (117M to 774M) and a threefold increase in depth (12 to 36 layers), suggesting core dimensionality may not depend on model scale.
— Transformers converge to invariant algorithmic cores
(2602.22600 - Schiffman, 26 Feb 2026) in Discussion — Limitations and future directions