Interpreting Neural Networks through Mahalanobis Distance
Abstract: This paper introduces a theoretical framework that connects neural network linear layers with the Mahalanobis distance, offering a new perspective on neural network interpretability. While previous studies have explored activation functions primarily for performance optimization, our work interprets these functions through statistical distance measures, a less explored area in neural network research. By establishing this connection, we provide a foundation for developing more interpretable neural network models, which is crucial for applications requiring transparency. Although this work is theoretical and does not include empirical data, the proposed distance-based interpretation has the potential to enhance model robustness, improve generalization, and provide more intuitive explanations of neural network decisions.
- Explainable ai frameworks: Navigating the present challenges and unveiling innovative applications. Algorithms, 17(6):227, 2024. doi: 10.3390/a17060227.
- Christopher M. Bishop. Pattern Recognition and Machine Learning. Springer, 2006.
- Weight uncertainty in neural networks. In Proceedings of the 32nd International Conference on Machine Learning (ICML), pages 1613–1622. PMLR, 2015.
- Radial basis functions, multi-variable functional interpolation and adaptive networks. Royal Signals and Radar Establishment Malvern (United Kingdom) RSRE Memorandum, 4148, 1988.
- Fast and accurate deep network learning by exponential linear units (elus). In 4th International Conference on Learning Representations (ICLR), 2016. URL https://arxiv.org/abs/1511.07289.
- The mahalanobis distance. Chemometrics and Intelligent Laboratory Systems, 50(1):1–18, 2000. doi: 10.1016/S0169-7439(99)00047-7.
- Visualizing higher-layer features of a deep network. In Technical Report, University of Montreal, 2009.
- A density-based algorithm for discovering clusters in large spatial databases with noise. In Proceedings of the Second International Conference on Knowledge Discovery and Data Mining (KDD-96), pages 226–231. AAAI Press, 1996.
- Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the 13th International Conference on Artificial Intelligence and Statistics (AISTATS), pages 249–256, 2010.
- Interpreting artificial intelligence models: A systematic review on the application of lime and shap in alzheimer’s disease detection. Brain Informatics, 11(1):1–19, 2024. doi: 10.1186/s40708-023-00195-9.
- Delving deep into rectifiers: Surpassing human-level performance on imagenet classification. In Proceedings of the IEEE International Conference on Computer Vision, pages 1026–1034, 2015.
- Gaussian error linear units (gelus). arXiv preprint arXiv:1606.08415, 2016. URL https://arxiv.org/abs/1606.08415.
- Long short-term memory. Neural Computation, 9(8):1735–1780, 1997.
- Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the 32nd International Conference on Machine Learning (ICML), pages 448–456, 2015. URL https://arxiv.org/abs/1502.03167.
- Adaptive mixtures of local experts. Neural Computation, 3(1):79–87, 1991.
- Ian T. Jolliffe. Principal Component Analysis. Springer, 2 edition, 2002.
- A survey of computational imaging methods for inverse problems. IEEE Signal Processing Magazine, 34(6):85–95, 2017.
- Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In Proceedings of the 35th International Conference on Machine Learning (ICML), pages 2668–2677. PMLR, 2018a.
- Interpretability beyond feature attribution: Quantitative testing with concept activation vectors (tcav). In International Conference on Machine Learning, pages 2668–2677. PMLR, 2018b.
- Imagenet classification with deep convolutional neural networks. In Advances in Neural Information Processing Systems, volume 25, pages 1097–1105, 2012.
- Deep learning for case-based reasoning through prototypes: A neural network that explains its predictions. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 32, 2018.
- Zachary C. Lipton. The mythos of model interpretability. arXiv preprint arXiv:1606.03490, 2016. URL https://arxiv.org/abs/1606.03490.
- ngpt: Normalized transformer with representation learning on the hypersphere, 2024. URL https://arxiv.org/abs/2410.01131.
- A unified approach to interpreting model predictions. Advances in Neural Information Processing Systems, 30:4765–4774, 2017.
- Rectifier nonlinearities improve neural network acoustic models. In Proc. icml, volume 30, page 3, 2013.
- J. MacQueen. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, volume 1, pages 281–297. University of California Press, 1967a.
- J. B. MacQueen. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Volume 1: Statistics, pages 281–297. University of California Press, 1967b.
- Prasanta Chandra Mahalanobis. On the generalized distance in statistics. Proceedings of the National Institute of Sciences of India, 2(1):49–55, 1936.
- Mark Markov. Lime vs shap: A comparative analysis of interpretability tools. MarkovML Blog, 2020. URL https://www.markovml.com/blog/lime-vs-shap-comparative-analysis.
- Perceptrons: An Introduction to Computational Geometry. MIT Press, Cambridge, MA, USA, 1969. ISBN 9780262631832.
- Fionn Murtagh. A survey of recent advances in hierarchical clustering algorithms. The Computer Journal, 26(4):354–359, 1983.
- Rectified linear units improve restricted boltzmann machines. In Proceedings of the 27th International Conference on Machine Learning (ICML), pages 807–814, 2010.
- Radford M. Neal. Bayesian Learning for Neural Networks, volume 118 of Lecture Notes in Statistics. Springer-Verlag, 1996.
- Searching for activation functions. In International Conference on Learning Representations, 2018. URL https://openreview.net/forum?id=Hk4_qw5K-. arXiv preprint arXiv:1710.05941.
- Douglas A. Reynolds. Gaussian mixture models. In Encyclopedia of Biometrics, pages 659–663. Springer, 2009.
- ”why should i trust you?”: Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD international conference on knowledge discovery and data mining, pages 1135–1144. ACM, 2016.
- Frank Rosenblatt. The perceptron: A probabilistic model for information storage and organization in the brain. Psychological Review, 65(6):386–408, 1958.
- Cynthia Rudin. Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence, 1(5):206–215, 2019.
- Deep inside convolutional networks: Visualising image classification models and saliency maps. arXiv preprint arXiv:1312.6034, 2013. URL https://arxiv.org/abs/1312.6034.
- Ulrike Von Luxburg. A tutorial on spectral clustering. Statistics and Computing, 17(4):395–416, 2007.
- Distance metric learning for large margin nearest neighbor classification. In International Conference on Machine Learning (ICML), pages 1317–1324, 2009. URL https://proceedings.mlr.press/v5/weinberger09a.html.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.