LinBridge: A Learnable Framework for Interpreting Nonlinear Neural Encoding Models
Abstract: Neural encoding of artificial neural networks (ANNs) links their computational representations to brain responses, offering insights into how the brain processes information. Current studies mostly use linear encoding models for clarity, even though brain responses are often nonlinear. This has sparked interest in developing nonlinear encoding models that are still interpretable. To address this problem, we propose LinBridge, a learnable and flexible framework based on Jacobian analysis for interpreting nonlinear encoding models. LinBridge posits that the nonlinear mapping between ANN representations and neural responses can be factorized into a linear inherent component that approximates the complex nonlinear relationship, and a mapping bias that captures sample-selective nonlinearity. The Jacobian matrix, which reflects output change rates relative to input, enables the analysis of sample-selective mapping in nonlinear models. LinBridge employs a self-supervised learning strategy to extract both the linear inherent component and nonlinear mapping biases from the Jacobian matrices of the test set, allowing it to adapt effectively to various nonlinear encoding models. We validate the LinBridge framework in the scenario of neural visual encoding, using computational visual representations from CLIP-ViT to predict brain activity recorded via functional magnetic resonance imaging (fMRI). Our experimental results demonstrate that: 1) the linear inherent component extracted by LinBridge accurately reflects the complex mappings of nonlinear neural encoding models; 2) the sample-selective mapping bias elucidates the variability of nonlinearity across different levels of the visual processing hierarchy. This study presents a novel tool for interpreting nonlinear neural encoding models and offers fresh evidence about hierarchical nonlinearity distribution in the visual cortex.
- Mostafa Abdou. Connecting neural response measurements & computational models of language: a non-comprehensive guide. arXiv preprint arXiv:2203.05300, 2022.
- A massive 7t fmri dataset to bridge cognitive neuroscience and artificial intelligence. Nature Neuroscience, 25:116 – 126, 2021.
- Neural additive vector autoregression models for causal discovery in time series. In Discovery Science: 24th International Conference, DS 2021, Halifax, NS, Canada, October 11–13, 2021, Proceedings 24, pp. 446–460. Springer, 2021.
- Brains and algorithms partially converge in natural language processing. Communications biology, 5(1):134, 2022.
- Mark S Cohen. Parametric analysis of fmri data using linear systems methods. Neuroimage, 6(2):93–103, 1997.
- Study on representation invariances of cnns and human visual information processing based on data augmentation. Brain Sciences, 10(9):602, 2020.
- Gabornet visual encoding: A lightweight region-based visual encoding model with good expressiveness and biological interpretability. Frontiers in Neuroscience, 15:614182, 2021.
- Seeing it all: Convolutional network layers map the function of the human visual system. NeuroImage, 152:184–194, 2017. doi: https://doi.org/10.1016/j.neuroimage.2016.10.001.
- The jacobian matrix and global univalence of mappings. Mathematische Annalen, 159(2):81–93, 1965.
- A multi-modal parcellation of human cerebral cortex. Nature, 536(7615):171–178, 2016.
- Shared computational principles for language processing in humans and deep language models. Nature neuroscience, 25(3):369–380, 2022.
- Umut Güçlü and Marcel A. J. van Gerven. Deep neural networks reveal a gradient in the complexity of neural representations across the ventral stream. 35(27):10005–10014, 2015. doi: 10.1523/JNEUROSCI.5023-14.2015.
- Functional connectivity in resting-state fmri: is linear correlation sufficient? Neuroimage, 54(3):2218–2225, 2011.
- Modeling task fmri data via deep convolutional autoencoder. IEEE transactions on medical imaging, 37(7):1551–1561, 2017.
- Neuroanatomy, visual cortex. 2018.
- Incorporating context into language encoding models for fmri. Advances in neural information processing systems, 31, 2018.
- Deep supervised, but not unsupervised, models may explain it cortical representation. PLOS Computational Biology, 10(11):1–29, 11 2014. doi: 10.1371/journal.pcbi.1003915.
- Economy statistical recurrent units for inferring nonlinear granger causality. arXiv preprint arXiv:1911.09879, 2019.
- An improved gabornet visual encoding model with nonlinear mapping and data augmentation. In 2022 7th International Conference on Computational Intelligence and Applications (ICCIA), pp. 180–184. IEEE, 2022.
- Dissecting neural computations in the human auditory pathway using deep neural networks for speech. Nature Neuroscience, 26:2213–2225, 2023. doi: 10.1038/s41593-023-01468-4.
- Microsoft coco: Common objects in context. In Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13, pp. 740–755. Springer, 2014.
- Coupling artificial neurons in bert and biological neurons in the human brain. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pp. 8888–8896, 2023.
- Toward a realistic model of speech processing in the brain with self-supervised learning. In S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh (eds.), Advances in Neural Information Processing Systems, volume 35, pp. 33428–33443. Curran Associates, Inc., 2022.
- Encoding and decoding in fmri. NeuroImage, 56(2):400–410, 2011a.
- Encoding and decoding in fmri. Neuroimage, 56(2):400–410, 2011b.
- Representation learning with contrastive predictive coding. arXiv preprint arXiv:1807.03748, 2018.
- Learning transferable visual models from natural language supervision. In International conference on machine learning, pp. 8748–8763. PMLR, 2021.
- Learnable latent embeddings for joint behavioural and neural analysis. Nature, 617(7960):360–368, 2023.
- The neural architecture of language: Integrative modeling converges on predictive processing. Proceedings of the National Academy of Sciences, 118(45):e2105646118, 2021.
- Revealing vision-language integration in the brain with multimodal networks. arXiv preprint arXiv:2406.14481, 2024.
- Granger causality using jacobian in neural networks. Chaos: An Interdisciplinary Journal of Nonlinear Science, 33(2), 2023.
- Brain encoding models based on multimodal transformers can transfer across language and vision. Advances in Neural Information Processing Systems, 36, 2024.
- Neural granger causality. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(8):4267–4279, 2021.
- Many but not all deep neural network audio models capture brain responses and exhibit correspondence between model stages and brain regions. PLOS Biology, 21(12):1–70, 12 2023. doi: 10.1371/journal.pbio.3002366.
- Self-supervised models of audio effectively explain human cortical responses to speech. In Kamalika Chaudhuri, Stefanie Jegelka, Le Song, Csaba Szepesvari, Gang Niu, and Sivan Sabato (eds.), Proceedings of the 39th International Conference on Machine Learning, volume 162 of Proceedings of Machine Learning Research, pp. 21927–21944. PMLR, 17–23 Jul 2022.
- Laurens Van der Maaten and Geoffrey Hinton. Visualizing data using t-sne. Journal of machine learning research, 9(11), 2008.
- Incorporating natural language into vision models improves prediction and understanding of higher visual cortex. bioRxiv, 2022. doi: 10.1101/2022.09.27.508760.
- Better models of human high-level visual cortex emerge from natural language supervision with a large and diverse dataset. Nature Machine Intelligence, 5(12):1415–1426, 2023.
- Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proceedings of the National Academy of Sciences, 111(23):8619–8624, 2014.
- Using goal-driven deep learning models to understand sensory cortex. Nature neuroscience, 19(3):356–365, 2016.
- A visual encoding model based on deep neural networks and transfer learning for brain activity measured by functional magnetic resonance imaging. Journal of neuroscience methods, 325:108318, 2019.
- Zijun Zhang. Improved adam optimizer for deep neural networks. In 2018 IEEE/ACM 26th international symposium on quality of service (IWQoS), pp. 1–2. Ieee, 2018.
- Fine-grained artificial neurons in audio-transformers for disentangling neural auditory encoding. In Findings of the Association for Computational Linguistics: ACL 2023, pp. 7943–7956, 2023.
- Jacobian regularizer-based neural granger causality. arXiv preprint arXiv:2405.08779, 2024.
- Unsupervised neural network models of the ventral visual stream. Proceedings of the National Academy of Sciences, 118(3):e2014196118, 2021. doi: 10.1073/pnas.2014196118.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.