Papers
Topics
Authors
Recent
Search
2000 character limit reached

Deep Learning: Our Miraculous Year 1990-1991

Published 12 May 2020 in cs.NE | (2005.05744v3)

Abstract: In 2020-2021, we celebrated that many of the basic ideas behind the deep learning revolution were published three decades ago within fewer than 12 months in our "Annus Mirabilis" or "Miraculous Year" 1990-1991 at TU Munich. Back then, few people were interested, but a quarter century later, neural networks based on these ideas were on over 3 billion devices such as smartphones, and used many billions of times per day, consuming a significant fraction of the world's compute.

Citations (5)

Summary

  • The paper introduces the first very deep neural networks with unsupervised pre-training to counteract the vanishing gradient problem.
  • The paper pioneers neural network compression techniques that effectively distill knowledge from a teacher model to a student model.
  • The work lays the foundation for LSTM, adversarial networks, and dynamic weight programming, shaping modern AI architectures.

Overview of "Deep Learning: Our Miraculous Year 1990-1991"

The paper "Deep Learning: Our Miraculous Year 1990-1991" authored by Jürgen Schmidhuber provides a comprehensive account of a pivotal period in the development of deep learning theories and practices—documented by the author's research group at TU Munich. It recounts a concentrated phase of innovation that significantly contributed to machine learning and artificial intelligence, forming the foundation of several contemporary applications. This summary highlights key insights, contributions, and subsequent ramifications that emerged from this "Miraculous Year."

Key Contributions

  1. First Very Deep Learner: The paper elucidates the development of the first very deep neural networks with unsupervised or self-supervised pre-training, overcoming previous limitations in training depth. The use of predictive coding to minimize description length at each layer was particularly instrumental in addressing the vanishing gradient issue.
  2. Neural Network Compression: The introduction of a method to compress one neural network into another, also known retrospectively as "distillation," was foundational in knowledge transfer between models. The technique allowed effectively transferring learned behavior from a teacher network to a student network.
  3. Fundamental Deep Learning Problem: Insights into the vanishing/exploding gradient problem were identified, analyzing the challenges of deep neural network training. This understanding laid groundwork for subsequent solutions, such as Long Short-Term Memory (LSTM) networks.
  4. Development of LSTM: The paper chronicles the inception of LSTM, designed to retain information over extended sequences while managing gradient flow effectively, a significant stride in sequential data processing. LSTM has since become a pivotal architecture in diverse applications, from language processing to speech recognition.
  5. Adversarial Generative Networks: Concepts similar to Generative Adversarial Networks (GANs) were introduced as early as 1990, with adversarial networks generating novel outputs to challenge predictive models, a framework inspiring later advances in model innovation through adversarial interactions.
  6. Fast Weight Programmers: Introducing networks capable of manipulating their weights dynamically, the notion of Fast Weight Programmers anticipated modern Transformer architectures, which leverage dynamic and context-sensitive connections through mechanisms akin to self-attention.

Implications

The work documented in Schmidhuber's paper demonstrates the transition from theoretical constructs to robust implementations in deep learning, informing both the academic discourse and commercial application ecosystems. The detailed treatment of gradient-based learning, optimization strategies, and network architectures provided insights that continue to steer innovations. Specifically, LSTMs and their derivatives are central to technologies in natural language processing, autonomous systems, and beyond. Moreover, the retrospective discussion on unsupervised pre-training versus pure supervised learning reflects ongoing debates in efficient model training strategies.

Future Developments

Looking forward, the foundation laid during this "Miraculous Year" will likely persist in guiding future inquiries into scalable and adaptable neural networks. As computational capabilities evolve, the principles of compressing large networks and dynamically adjusting learning algorithms in real-time will drive further evolution in model efficiency and applicability. The pioneering efforts in hierarchical reinforcement learning and sequential attention mechanisms are poised for broader exploration in designing more sophisticated, human-like AI systems.

In summary, the insights from 1990-1991 delineate a critical juncture in machine learning history, one that continues to resonate throughout the tapestry of modern AI research and development. The methodologies and philosophies espoused during this period have entrenched themselves as cornerstones in the quest for effective and intelligent neural network models.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (1)

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 3 likes about this paper.

HackerNews

  1. Deep Learning: Our Year 1990-1991 (1 point, 0 comments) 
  2. Deep Learning: Our Year 1990-1991 (1 point, 0 comments)