Tuned Compositional Feature Replays for Efficient Stream Learning
Abstract: Our brains extract durable, generalizable knowledge from transient experiences of the world. Artificial neural networks come nowhere close to this ability. When tasked with learning to classify objects by training on non-repeating video frames in temporal order (online stream learning), models that learn well from shuffled datasets catastrophically forget old knowledge upon learning new stimuli. We propose a new continual learning algorithm, Compositional Replay Using Memory Blocks (CRUMB), which mitigates forgetting by replaying feature maps reconstructed by combining generic parts. CRUMB concatenates trainable and re-usable "memory block" vectors to compositionally reconstruct feature map tensors in convolutional neural networks. Storing the indices of memory blocks used to reconstruct new stimuli enables memories of the stimuli to be replayed during later tasks. This reconstruction mechanism also primes the neural network to minimize catastrophic forgetting by biasing it towards attending to information about object shapes more than information about image textures, and stabilizes the network during stream learning by providing a shared feature-level basis for all training examples. These properties allow CRUMB to outperform an otherwise identical algorithm that stores and replays raw images, while occupying only 3.6% as much memory. We stress-tested CRUMB alongside 13 competing methods on 7 challenging datasets. To address the limited number of existing online stream learning datasets, we introduce 2 new benchmarks by adapting existing datasets for stream learning. With only 3.7-4.1% as much memory and 15-43% as much runtime, CRUMB mitigates catastrophic forgetting more effectively than the state-of-the-art. Our code is available at https://github.com/MorganBDT/crumb.git.
- Michael McCloskey and Neal J Cohen “Catastrophic interference in connectionist networks: The sequential learning problem” In Psychology of learning and motivation 24 Elsevier, 1989, pp. 109–165
- Roger Ratcliff “Connectionist models of recognition memory: constraints imposed by learning and forgetting functions.” In Psychological review 97.2 American Psychological Association, 1990, pp. 285
- Robert M French “Catastrophic forgetting in connectionist networks” In Trends in cognitive sciences 3.4 Elsevier, 1999, pp. 128–135
- “Remind your neural network to prevent catastrophic forgetting” In European Conference on Computer Vision, 2020, pp. 466–483 Springer
- “Measuring catastrophic forgetting in neural networks” In Thirty-second AAAI conference on artificial intelligence, 2018
- “Continuous learning in single-incremental-task scenarios” In Neural Networks Elsevier, 2019
- “A continual learning survey: Defying forgetting in classification tasks” In IEEE transactions on pattern analysis and machine intelligence 44.7 IEEE, 2021, pp. 3366–3385
- “How catastrophic can catastrophic forgetting be in linear regression?” In Conference on Learning Theory, 2022, pp. 4028–4079 PMLR
- Hitesh Vaidya, Travis Desell and Alexander G Ororbia “Reducing Catastrophic Forgetting in Self Organizing Maps with Internally-Induced Generative Replay (Student Abstract)” In Proceedings of the AAAI Conference on Artificial Intelligence 36.11, 2022, pp. 13069–13070
- “Overcoming Catastrophic Forgetting beyond Continual Learning: Balanced Training for Neural Machine Translation” In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022, pp. 2023–2036
- Ameya Prabhu, Philip HS Torr and Puneet K Dokania “Gdumb: A simple approach that questions our progress in continual learning” In European conference on computer vision, 2020, pp. 524–540 Springer
- “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
- Cecilia S Lee and Aaron Y Lee “Clinical applications of continual learning machine learning” In The Lancet Digital Health 2.6 Elsevier, 2020, pp. e279–e281
- “The toybox dataset of egocentric visual object transformations” In arXiv preprint arXiv:1806.06034, 2018
- Ali Borji, Saeed Izadi and Laurent Itti “ilab-20m: A large-scale controlled object dataset to investigate deep learning” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2221–2230
- “Learning without forgetting” In IEEE transactions on pattern analysis and machine intelligence 40.12 IEEE, 2017, pp. 2935–2947
- “Efficient lifelong learning with a-gem” In arXiv preprint arXiv:1812.00420, 2018
- “Overcoming catastrophic interference using conceptor-aided backpropagation”, 2018
- Friedemann Zenke, Ben Poole and Surya Ganguli “Continual learning through synaptic intelligence” In Proceedings of the 34th International Conference on Machine Learning-Volume 70, 2017, pp. 3987–3995 JMLR. org
- “Overcoming catastrophic forgetting by incremental moment matching” In Advances in neural information processing systems, 2017, pp. 4652–4662
- “Overcoming catastrophic forgetting in continual learning by exploring eigenvalues of hessian matrix” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2023
- Junfeng Wen, Yanshuai Cao and Ruitong Huang “Few-Shot Self Reminder to Overcome Catastrophic Forgetting” In arXiv preprint arXiv:1812.00543, 2018
- “Online continual learning in image classification: An empirical survey” In Neurocomputing 469 Elsevier, 2022, pp. 28–51
- Zeyang Zhang, Yidong Chen and Changle Zhou “Self-growing binary activation network: A novel deep learning model with dynamic architecture” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2022
- “Pathnet: Evolution channels gradient descent in super neural networks” In arXiv preprint arXiv:1701.08734, 2017
- “Random path selection for continual learning” In Advances in Neural Information Processing Systems 32, 2019
- “Overcoming catastrophic forgetting with hard attention to the task” In International Conference on Machine Learning, 2018, pp. 4548–4557 PMLR
- Tameem Adel, Han Zhao and Richard E Turner “Continual learning with adaptive weights (claw)” In arXiv preprint arXiv:1911.09514, 2019
- “Sparse Distributed Memory is a Continual Learner” In International Conference on Learning Representations, 2023
- “Progress & compress: A scalable framework for continual learning” In arXiv preprint arXiv:1805.06370, 2018
- Siavash Golkar, Michael Kagan and Kyunghyun Cho “Continual learning via neural pruning” In arXiv preprint arXiv:1903.04476, 2019
- “Overcoming long-term catastrophic forgetting through adversarial neural pruning and synaptic consolidation” In IEEE Transactions on Neural Networks and Learning Systems 33.9 IEEE, 2021, pp. 4243–4256
- “Efficient architecture search for continual learning” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2022
- “Large scale incremental learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 374–382
- “ICARL: Incremental classifier and representation learning” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2001–2010
- “Gradient based sample selection for online continual learning” In arXiv preprint arXiv:1903.08671, 2019
- “Variational continual learning” In arXiv preprint arXiv:1710.10628, 2017
- David Lopez-Paz “Gradient episodic memory for continual learning” In Advances in Neural Information Processing Systems, 2017, pp. 6467–6476
- “Rainbow Memory: Continual Learning with a Memory of Diverse Samples” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 8218–8227
- “Continual learning with deep generative replay” In Advances in Neural Information Processing Systems, 2017, pp. 2990–2999
- Anthony Robins “Catastrophic forgetting, rehearsal and pseudorehearsal” In Connection Science 7.2 Taylor & Francis, 1995, pp. 123–146
- “Mnemonics training: Multi-class incremental learning without forgetting” In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, 2020, pp. 12245–12254
- “Pseudo-recursal: Solving the catastrophic forgetting problem in deep neural networks” In arXiv preprint arXiv:1802.03875, 2018
- “Generative feature replay for class-incremental learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020, pp. 226–227
- “Generative feature replay with orthogonal weight modification for continual learning” In 2021 International Joint Conference on Neural Networks (IJCNN), 2021, pp. 1–8 IEEE
- Gido M Ven, Zhe Li and Andreas S Tolias “Class-Incremental Learning with Generative Classifiers” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 3611–3620
- “Latent replay for real-time continual learning” In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020, pp. 10203–10209 IEEE
- “Memory recall: A simple neural network training framework against catastrophic forgetting” In IEEE Transactions on Neural Networks and Learning Systems 33.5 IEEE, 2021, pp. 2010–2022
- Herve Jegou, Matthijs Douze and Cordelia Schmid “Product quantization for nearest neighbor search” In IEEE transactions on pattern analysis and machine intelligence 33.1 IEEE, 2010, pp. 117–128
- “Play it again: reactivation of waking experience and memory” In Trends in neurosciences 33.5 Elsevier, 2010, pp. 220–229
- Penelope A Lewis and Simon J Durrant “Overlapping memory replay during sleep builds cognitive schemata” In Trends in cognitive sciences 15.8 Elsevier, 2011, pp. 343–351
- “Replay of learned neural firing sequences during rest in human motor cortex” In Cell Reports 31.5 Elsevier, 2020, pp. 107581
- James L McClelland, Bruce L McNaughton and Randall C O’Reilly “Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory.” In Psychological review 102.3 American Psychological Association, 1995, pp. 419
- “Lifelong Learning With Cycle Memory Networks” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2023
- “CORe50: a new dataset and benchmark for continuous object recognition” In Conference on Robot Learning, 2017, pp. 17–26 PMLR
- “Object identification from few examples by improving the invariance of a deep convolutional neural network” In 2016 IEEE/RSJ international conference on intelligent robots and systems (IROS), 2016, pp. 4904–4911 IEEE
- “Learning multiple layers of features from tiny images” Citeseer, 2009
- “Imagenet: A large-scale hierarchical image database” In 2009 IEEE conference on computer vision and pattern recognition, 2009, pp. 248–255 Ieee
- “SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and¡ 0.5 MB model size” In arXiv preprint arXiv:1602.07360, 2016
- Yaoyao Liu, Bernt Schiele and Qianru Sun “Adaptive aggregation networks for class-incremental learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 2544–2553
- Matthias De Lange and Tinne Tuytelaars “Continual prototype evolution: Learning online from non-stationary data streams” In arXiv preprint arXiv:2009.00919, 2020
- “Understanding the role of training regimes in continual learning” In arXiv preprint arXiv:2006.06958, 2020
- “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778
- “PyTorch: An imperative style, high-performance deep learning library” In arXiv preprint arXiv:1912.01703, 2019
- “Memory aware synapses: Learning what (not) to forget” In Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 139–154
- “How transferable are features in deep neural networks?” In arXiv preprint arXiv:1411.1792, 2014
- Yutian Chen, Max Welling and Alex Smola “Super-samples from kernel herding” In arXiv preprint arXiv:1203.3472, 2012
- Pang Wei Koh and Percy Liang “Understanding black-box predictions via influence functions” In Proceedings of the 34th International Conference on Machine Learning-Volume 70, 2017, pp. 1885–1894 JMLR. org
- Pratik Prabhanjan Brahma and Adrienne Othon “Subset Replay Based Continual Learning for Scalable Improvement of Autonomous Systems” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2018
- “Prototype-Guided Memory Replay for Continual Learning” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2023
- “ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness.” In International Conference on Learning Representations, 2019
- “On the Robustness, Generalization, and Forgetting of Shape-Texture Debiased Continual Learning” In arXiv preprint arXiv:2211.11174, 2022
- “Mobilenetv2: Inverted residuals and linear bottlenecks” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4510–4520
- “Attention is all you need” In arXiv preprint arXiv:1706.03762, 2017
- “An image is worth 16x16 words: Transformers for image recognition at scale” In International Conference on Learning Representations, 2021
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.