Establish whether Tabular Language Models learn tabular invariances or rely on non-tabular mechanisms
Determine whether Tabular Language Models that serialize tables into text (including models such as Tabula-8B and related fine-tuned LLMs) actually learn tabular invariances—specifically row-permutation and column-permutation invariance—and generalize across tabular prediction tasks, or whether their observed performance primarily arises from non-tabular mechanisms such as instruction-following abilities and format familiarity.
References
Yet it remains unclear whether TLMs actually learn tabular invariances and generalize over tabular data or succeed through other mechanisms entirely.
— The Illusion of Generalization: Re-examining Tabular Language Model Evaluation
(2602.04031 - Gorla et al., 3 Feb 2026) in Section 1 (Introduction)