Papers
Topics
Authors
Recent
Search
2000 character limit reached

Momentum-based Accelerated Q-learning

Published 23 Oct 2019 in eess.SY and cs.SY | (1910.11673v1)

Abstract: This paper studies accelerated algorithms for Q-learning. We propose an acceleration scheme by incorporating the historical iterates of the Q-function. The idea is conceptually inspired by the momentum-based acceleration methods in the optimization theory. Under finite state-action space settings, the proposed accelerated Q-learning algorithm provably converges to the global optimum with a rate of $\mathcal{O}(1/\sqrt{T})$. While sharing a comparable theoretic convergence rate with the existing Speedy Q-learning (SpeedyQ) algorithm, we numerically show that the proposed algorithm outperforms SpeedyQ via playing the FrozenLake grid world game. Furthermore, we generalize the acceleration scheme to the continuous state-action space case where function approximation of the Q-function is necessary. In this case, the algorithms are validated using commonly adopted testing problems in reinforcement learning, including two discrete-time linear quadratic regulation (LQR) problems from the Deepmind Control Suite, and the Atari 2600 games. Simulation results show that the proposed accelerated algorithms can improve the convergence performance compared with the vanilla Q-learning algorithm.

Citations (1)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.