Papers
Topics
Authors
Recent
Search
2000 character limit reached

Tuned Compositional Feature Replays for Efficient Stream Learning

Published 6 Apr 2021 in cs.CV and cs.AI | (2104.02206v8)

Abstract: Our brains extract durable, generalizable knowledge from transient experiences of the world. Artificial neural networks come nowhere close to this ability. When tasked with learning to classify objects by training on non-repeating video frames in temporal order (online stream learning), models that learn well from shuffled datasets catastrophically forget old knowledge upon learning new stimuli. We propose a new continual learning algorithm, Compositional Replay Using Memory Blocks (CRUMB), which mitigates forgetting by replaying feature maps reconstructed by combining generic parts. CRUMB concatenates trainable and re-usable "memory block" vectors to compositionally reconstruct feature map tensors in convolutional neural networks. Storing the indices of memory blocks used to reconstruct new stimuli enables memories of the stimuli to be replayed during later tasks. This reconstruction mechanism also primes the neural network to minimize catastrophic forgetting by biasing it towards attending to information about object shapes more than information about image textures, and stabilizes the network during stream learning by providing a shared feature-level basis for all training examples. These properties allow CRUMB to outperform an otherwise identical algorithm that stores and replays raw images, while occupying only 3.6% as much memory. We stress-tested CRUMB alongside 13 competing methods on 7 challenging datasets. To address the limited number of existing online stream learning datasets, we introduce 2 new benchmarks by adapting existing datasets for stream learning. With only 3.7-4.1% as much memory and 15-43% as much runtime, CRUMB mitigates catastrophic forgetting more effectively than the state-of-the-art. Our code is available at https://github.com/MorganBDT/crumb.git.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (75)
  1. Michael McCloskey and Neal J Cohen “Catastrophic interference in connectionist networks: The sequential learning problem” In Psychology of learning and motivation 24 Elsevier, 1989, pp. 109–165
  2. Roger Ratcliff “Connectionist models of recognition memory: constraints imposed by learning and forgetting functions.” In Psychological review 97.2 American Psychological Association, 1990, pp. 285
  3. Robert M French “Catastrophic forgetting in connectionist networks” In Trends in cognitive sciences 3.4 Elsevier, 1999, pp. 128–135
  4. “Remind your neural network to prevent catastrophic forgetting” In European Conference on Computer Vision, 2020, pp. 466–483 Springer
  5. “Measuring catastrophic forgetting in neural networks” In Thirty-second AAAI conference on artificial intelligence, 2018
  6. “Continuous learning in single-incremental-task scenarios” In Neural Networks Elsevier, 2019
  7. “A continual learning survey: Defying forgetting in classification tasks” In IEEE transactions on pattern analysis and machine intelligence 44.7 IEEE, 2021, pp. 3366–3385
  8. “How catastrophic can catastrophic forgetting be in linear regression?” In Conference on Learning Theory, 2022, pp. 4028–4079 PMLR
  9. Hitesh Vaidya, Travis Desell and Alexander G Ororbia “Reducing Catastrophic Forgetting in Self Organizing Maps with Internally-Induced Generative Replay (Student Abstract)” In Proceedings of the AAAI Conference on Artificial Intelligence 36.11, 2022, pp. 13069–13070
  10. “Overcoming Catastrophic Forgetting beyond Continual Learning: Balanced Training for Neural Machine Translation” In Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers), 2022, pp. 2023–2036
  11. Ameya Prabhu, Philip HS Torr and Puneet K Dokania “Gdumb: A simple approach that questions our progress in continual learning” In European conference on computer vision, 2020, pp. 524–540 Springer
  12. “Overcoming catastrophic forgetting in neural networks” In Proceedings of the national academy of sciences 114.13 National Acad Sciences, 2017, pp. 3521–3526
  13. Cecilia S Lee and Aaron Y Lee “Clinical applications of continual learning machine learning” In The Lancet Digital Health 2.6 Elsevier, 2020, pp. e279–e281
  14. “The toybox dataset of egocentric visual object transformations” In arXiv preprint arXiv:1806.06034, 2018
  15. Ali Borji, Saeed Izadi and Laurent Itti “ilab-20m: A large-scale controlled object dataset to investigate deep learning” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2016, pp. 2221–2230
  16. “Learning without forgetting” In IEEE transactions on pattern analysis and machine intelligence 40.12 IEEE, 2017, pp. 2935–2947
  17. “Efficient lifelong learning with a-gem” In arXiv preprint arXiv:1812.00420, 2018
  18. “Overcoming catastrophic interference using conceptor-aided backpropagation”, 2018
  19. Friedemann Zenke, Ben Poole and Surya Ganguli “Continual learning through synaptic intelligence” In Proceedings of the 34th International Conference on Machine Learning-Volume 70, 2017, pp. 3987–3995 JMLR. org
  20. “Overcoming catastrophic forgetting by incremental moment matching” In Advances in neural information processing systems, 2017, pp. 4652–4662
  21. “Overcoming catastrophic forgetting in continual learning by exploring eigenvalues of hessian matrix” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2023
  22. Junfeng Wen, Yanshuai Cao and Ruitong Huang “Few-Shot Self Reminder to Overcome Catastrophic Forgetting” In arXiv preprint arXiv:1812.00543, 2018
  23. “Online continual learning in image classification: An empirical survey” In Neurocomputing 469 Elsevier, 2022, pp. 28–51
  24. Zeyang Zhang, Yidong Chen and Changle Zhou “Self-growing binary activation network: A novel deep learning model with dynamic architecture” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2022
  25. “Pathnet: Evolution channels gradient descent in super neural networks” In arXiv preprint arXiv:1701.08734, 2017
  26. “Random path selection for continual learning” In Advances in Neural Information Processing Systems 32, 2019
  27. “Overcoming catastrophic forgetting with hard attention to the task” In International Conference on Machine Learning, 2018, pp. 4548–4557 PMLR
  28. Tameem Adel, Han Zhao and Richard E Turner “Continual learning with adaptive weights (claw)” In arXiv preprint arXiv:1911.09514, 2019
  29. “Sparse Distributed Memory is a Continual Learner” In International Conference on Learning Representations, 2023
  30. “Progress & compress: A scalable framework for continual learning” In arXiv preprint arXiv:1805.06370, 2018
  31. Siavash Golkar, Michael Kagan and Kyunghyun Cho “Continual learning via neural pruning” In arXiv preprint arXiv:1903.04476, 2019
  32. “Overcoming long-term catastrophic forgetting through adversarial neural pruning and synaptic consolidation” In IEEE Transactions on Neural Networks and Learning Systems 33.9 IEEE, 2021, pp. 4243–4256
  33. “Efficient architecture search for continual learning” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2022
  34. “Large scale incremental learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 374–382
  35. “ICARL: Incremental classifier and representation learning” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2017, pp. 2001–2010
  36. “Gradient based sample selection for online continual learning” In arXiv preprint arXiv:1903.08671, 2019
  37. “Variational continual learning” In arXiv preprint arXiv:1710.10628, 2017
  38. David Lopez-Paz “Gradient episodic memory for continual learning” In Advances in Neural Information Processing Systems, 2017, pp. 6467–6476
  39. “Rainbow Memory: Continual Learning with a Memory of Diverse Samples” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 8218–8227
  40. “Continual learning with deep generative replay” In Advances in Neural Information Processing Systems, 2017, pp. 2990–2999
  41. Anthony Robins “Catastrophic forgetting, rehearsal and pseudorehearsal” In Connection Science 7.2 Taylor & Francis, 1995, pp. 123–146
  42. “Mnemonics training: Multi-class incremental learning without forgetting” In Proceedings of the IEEE/CVF conference on Computer Vision and Pattern Recognition, 2020, pp. 12245–12254
  43. “Pseudo-recursal: Solving the catastrophic forgetting problem in deep neural networks” In arXiv preprint arXiv:1802.03875, 2018
  44. “Generative feature replay for class-incremental learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, 2020, pp. 226–227
  45. “Generative feature replay with orthogonal weight modification for continual learning” In 2021 International Joint Conference on Neural Networks (IJCNN), 2021, pp. 1–8 IEEE
  46. Gido M Ven, Zhe Li and Andreas S Tolias “Class-Incremental Learning with Generative Classifiers” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 3611–3620
  47. “Latent replay for real-time continual learning” In 2020 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2020, pp. 10203–10209 IEEE
  48. “Memory recall: A simple neural network training framework against catastrophic forgetting” In IEEE Transactions on Neural Networks and Learning Systems 33.5 IEEE, 2021, pp. 2010–2022
  49. Herve Jegou, Matthijs Douze and Cordelia Schmid “Product quantization for nearest neighbor search” In IEEE transactions on pattern analysis and machine intelligence 33.1 IEEE, 2010, pp. 117–128
  50. “Play it again: reactivation of waking experience and memory” In Trends in neurosciences 33.5 Elsevier, 2010, pp. 220–229
  51. Penelope A Lewis and Simon J Durrant “Overlapping memory replay during sleep builds cognitive schemata” In Trends in cognitive sciences 15.8 Elsevier, 2011, pp. 343–351
  52. “Replay of learned neural firing sequences during rest in human motor cortex” In Cell Reports 31.5 Elsevier, 2020, pp. 107581
  53. James L McClelland, Bruce L McNaughton and Randall C O’Reilly “Why there are complementary learning systems in the hippocampus and neocortex: insights from the successes and failures of connectionist models of learning and memory.” In Psychological review 102.3 American Psychological Association, 1995, pp. 419
  54. “Lifelong Learning With Cycle Memory Networks” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2023
  55. “CORe50: a new dataset and benchmark for continuous object recognition” In Conference on Robot Learning, 2017, pp. 17–26 PMLR
  56. “Object identification from few examples by improving the invariance of a deep convolutional neural network” In 2016 IEEE/RSJ international conference on intelligent robots and systems (IROS), 2016, pp. 4904–4911 IEEE
  57. “Learning multiple layers of features from tiny images” Citeseer, 2009
  58. “Imagenet: A large-scale hierarchical image database” In 2009 IEEE conference on computer vision and pattern recognition, 2009, pp. 248–255 Ieee
  59. “SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and¡ 0.5 MB model size” In arXiv preprint arXiv:1602.07360, 2016
  60. Yaoyao Liu, Bernt Schiele and Qianru Sun “Adaptive aggregation networks for class-incremental learning” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 2544–2553
  61. Matthias De Lange and Tinne Tuytelaars “Continual prototype evolution: Learning online from non-stationary data streams” In arXiv preprint arXiv:2009.00919, 2020
  62. “Understanding the role of training regimes in continual learning” In arXiv preprint arXiv:2006.06958, 2020
  63. “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778
  64. “PyTorch: An imperative style, high-performance deep learning library” In arXiv preprint arXiv:1912.01703, 2019
  65. “Memory aware synapses: Learning what (not) to forget” In Proceedings of the European Conference on Computer Vision (ECCV), 2018, pp. 139–154
  66. “How transferable are features in deep neural networks?” In arXiv preprint arXiv:1411.1792, 2014
  67. Yutian Chen, Max Welling and Alex Smola “Super-samples from kernel herding” In arXiv preprint arXiv:1203.3472, 2012
  68. Pang Wei Koh and Percy Liang “Understanding black-box predictions via influence functions” In Proceedings of the 34th International Conference on Machine Learning-Volume 70, 2017, pp. 1885–1894 JMLR. org
  69. Pratik Prabhanjan Brahma and Adrienne Othon “Subset Replay Based Continual Learning for Scalable Improvement of Autonomous Systems” In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, 2018
  70. “Prototype-Guided Memory Replay for Continual Learning” In IEEE Transactions on Neural Networks and Learning Systems IEEE, 2023
  71. “ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness.” In International Conference on Learning Representations, 2019
  72. “On the Robustness, Generalization, and Forgetting of Shape-Texture Debiased Continual Learning” In arXiv preprint arXiv:2211.11174, 2022
  73. “Mobilenetv2: Inverted residuals and linear bottlenecks” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 4510–4520
  74. “Attention is all you need” In arXiv preprint arXiv:1706.03762, 2017
  75. “An image is worth 16x16 words: Transformers for image recognition at scale” In International Conference on Learning Representations, 2021
Citations (2)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.