Learning the RMALA proposal covariance G0 via reinforcement learning
Develop reinforcement learning techniques to learn the proposal covariance matrix G0 in the Riemannian Metropolis–adjusted Langevin algorithm (RMALA), potentially through reduced-rank covariance parameterizations, to enable effective adaptive tuning of gradient-based MCMC proposals.
References
For future research, we highlight that learning the proposal covariance structure (i.e. G0 in the setting of RMALA) is an open challenge for RL; we speculate that reduced-rank covariance matrix approximations may be useful here, enabling the difficulty of the learning task to be reduced, but our attempts to implement this (not shown) were unsuccessful.
— Harnessing the Power of Reinforcement Learning for Adaptive MCMC
(2507.00671 - Wang et al., 1 Jul 2025) in Section 5 (Discussion)