Papers
Topics
Authors
Recent
Search
2000 character limit reached

Learning a Low-Rank Feature Representation: Achieving Better Trade-Off between Stability and Plasticity in Continual Learning

Published 14 Dec 2023 in cs.LG and cs.CV | (2312.08740v1)

Abstract: In continual learning, networks confront a trade-off between stability and plasticity when trained on a sequence of tasks. To bolster plasticity without sacrificing stability, we propose a novel training algorithm called LRFR. This approach optimizes network parameters in the null space of the past tasks' feature representation matrix to guarantee the stability. Concurrently, we judiciously select only a subset of neurons in each layer of the network while training individual tasks to learn the past tasks' feature representation matrix in low-rank. This increases the null space dimension when designing network parameters for subsequent tasks, thereby enhancing the plasticity. Using CIFAR-100 and TinyImageNet as benchmark datasets for continual learning, the proposed approach consistently outperforms state-of-the-art methods.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (17)
  1. “Understanding the role of training regimes in continual learning,” in Proc. Adv. Neural Inf. Process. Syst, 2020, vol. 33, pp. 7308–7320.
  2. “Reconciling meta-learning and continual learning with online mixtures of tasks,” in Proc. Adv. Neural Inf. Process. Syst, 2019, vol. 32, pp. 9122–9133.
  3. “Packnet: Adding multiple tasks to a single network by iterative pruning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2018, pp. 7765–7773.
  4. “Selective experience replay for lifelong learning,” in Proc. AAAI Conf. Artif. Intell., 2018, pp. 3302–3309.
  5. “Experience replay for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2019, vol. 32, pp. 350–360.
  6. “Overcoming catastrophic forgetting in neural networks,” in Proc. Nat. Acad. Sci., 2017, vol. 114, pp. 3521–3526.
  7. “Gradient episodic memory for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2017, vol. 30, pp. 6467–6476.
  8. “Gradient projection memory for continual learning,” in Proc. Int. Conf. Learn. Represent., 2021.
  9. “Training networks in null space of feature covariance for continual learning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2021, pp. 184–193.
  10. “Efficient lifelong learning with a-GEM,” in Proc. Int. Conf. Learn. Represent., 2019.
  11. “Balancing stability and plasticity through advanced null space in continual learning,” in Proc. Eur. Conf. Comput. Vis., 2022, pp. 219–236.
  12. “Rethinking the value of network pruning,” in Proc. Int. Conf. Learn. Represent., 2019.
  13. “The lottery ticket hypothesis: Finding sparse, trainable neural networks,” in Proc. Int. Conf. Learn. Represent., 2019.
  14. “Learning efficient convolutional networks through network slimming,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 2736–2744.
  15. “Memory aware synapses: Learning what (not) to forget,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 139–154.
  16. “Improved schemes for episodic memory-based lifelong learning,” in Proc. Adv. Neural Inf. Process. Systs., 2020, vol. 33, pp. 1023–1035.
  17. “Continual learning of context-dependent processing in neural networks,” Nat. Mach. Intell., vol. 1, no. 8, pp. 364–372, 2019.

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.