Scaling laws for learned membership inference
Determine the scaling laws that govern the performance of learned membership inference attacks against fine-tuned autoregressive language models, quantifying how attack effectiveness varies with training diversity, classifier capacity, and feature complexity.
References
This paper demonstrates proof-of-concept, and the performance ceiling is likely higher; finding the scaling laws that govern learned membership inference remains an open question.
— Learning the Signature of Memorization in Autoregressive Language Models
(2604.03199 - Ilić et al., 3 Apr 2026) in Discussion, Scaling subsection