Learning a Low-Rank Feature Representation: Achieving Better Trade-Off between Stability and Plasticity in Continual Learning
Abstract: In continual learning, networks confront a trade-off between stability and plasticity when trained on a sequence of tasks. To bolster plasticity without sacrificing stability, we propose a novel training algorithm called LRFR. This approach optimizes network parameters in the null space of the past tasks' feature representation matrix to guarantee the stability. Concurrently, we judiciously select only a subset of neurons in each layer of the network while training individual tasks to learn the past tasks' feature representation matrix in low-rank. This increases the null space dimension when designing network parameters for subsequent tasks, thereby enhancing the plasticity. Using CIFAR-100 and TinyImageNet as benchmark datasets for continual learning, the proposed approach consistently outperforms state-of-the-art methods.
- “Understanding the role of training regimes in continual learning,” in Proc. Adv. Neural Inf. Process. Syst, 2020, vol. 33, pp. 7308–7320.
- “Reconciling meta-learning and continual learning with online mixtures of tasks,” in Proc. Adv. Neural Inf. Process. Syst, 2019, vol. 32, pp. 9122–9133.
- “Packnet: Adding multiple tasks to a single network by iterative pruning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2018, pp. 7765–7773.
- “Selective experience replay for lifelong learning,” in Proc. AAAI Conf. Artif. Intell., 2018, pp. 3302–3309.
- “Experience replay for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2019, vol. 32, pp. 350–360.
- “Overcoming catastrophic forgetting in neural networks,” in Proc. Nat. Acad. Sci., 2017, vol. 114, pp. 3521–3526.
- “Gradient episodic memory for continual learning,” in Proc. Adv. Neural Inf. Process. Systs., 2017, vol. 30, pp. 6467–6476.
- “Gradient projection memory for continual learning,” in Proc. Int. Conf. Learn. Represent., 2021.
- “Training networks in null space of feature covariance for continual learning,” in Proc. IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit., 2021, pp. 184–193.
- “Efficient lifelong learning with a-GEM,” in Proc. Int. Conf. Learn. Represent., 2019.
- “Balancing stability and plasticity through advanced null space in continual learning,” in Proc. Eur. Conf. Comput. Vis., 2022, pp. 219–236.
- “Rethinking the value of network pruning,” in Proc. Int. Conf. Learn. Represent., 2019.
- “The lottery ticket hypothesis: Finding sparse, trainable neural networks,” in Proc. Int. Conf. Learn. Represent., 2019.
- “Learning efficient convolutional networks through network slimming,” in Proc. IEEE Int. Conf. Comput. Vis., 2017, pp. 2736–2744.
- “Memory aware synapses: Learning what (not) to forget,” in Proc. Eur. Conf. Comput. Vis., 2018, pp. 139–154.
- “Improved schemes for episodic memory-based lifelong learning,” in Proc. Adv. Neural Inf. Process. Systs., 2020, vol. 33, pp. 1023–1035.
- “Continual learning of context-dependent processing in neural networks,” Nat. Mach. Intell., vol. 1, no. 8, pp. 364–372, 2019.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.