Improve training stability and sample efficiency of online continual learning algorithms
Improve the training stability and sample efficiency of online learning algorithms used for continual learning in embodied robotics, ensuring reliable convergence and effective learning under non-stationary, real-world data streams.
References
Open research problems and viable approaches include: 1) mixing different proportions of prior data distribution when fine-tuning on the latest data to alleviate catastrophic forgetting , 2) developing efficient prototypes from prior distributions or curricula for task inference in learning new tasks, 3) improving training stability and sample efficiency of online learning algorithms, 4) identifying principled ways to seamlessly incorporate large-capacity models into control frameworks, potentially through hierarchical learning or slow-fast control, for real-time inference.