2000 character limit reached
Almost Sure Asymptotic Freeness of Neural Network Jacobian with Orthogonal Weights
Published 11 Aug 2019 in math.PR, cs.LG, and stat.ML | (1908.03901v4)
Abstract: A well-conditioned Jacobian spectrum has a vital role in preventing exploding or vanishing gradients and speeding up learning of deep neural networks. Free probability theory helps us to understand and handle the Jacobian spectrum. We rigorously show almost sure asymptotic freeness of layer-wise Jacobians of deep neural networks as the wide limit. In particular, we treat the case that weights are initialized as Haar distributed orthogonal matrices.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.