Scalable Volt-VAR Optimization using RLlib-IMPALA Framework: A Reinforcement Learning Approach
Abstract: In the rapidly evolving domain of electrical power systems, the Volt-VAR optimization (VVO) is increasingly critical, especially with the burgeoning integration of renewable energy sources. Traditional approaches to learning-based VVO in expansive and dynamically changing power systems are often hindered by computational complexities. To address this challenge, our research presents a novel framework that harnesses the potential of Deep Reinforcement Learning (DRL), specifically utilizing the Importance Weighted Actor-Learner Architecture (IMPALA) algorithm, executed on the RAY platform. This framework, built upon RLlib-an industry-standard in Reinforcement Learning-ingeniously capitalizes on the distributed computing capabilities and advanced hyperparameter tuning offered by RAY. This design significantly expedites the exploration and exploitation phases in the VVO solution space. Our empirical results demonstrate that our approach not only surpasses existing DRL methods in achieving superior reward outcomes but also manifests a remarkable tenfold reduction in computational requirements. The integration of our DRL agent with the RAY platform facilitates the creation of RLlib-IMPALA, a novel framework that efficiently uses RAY's resources to improve system adaptability and control. RLlib-IMPALA leverages RAY's toolkit to enhance analytical capabilities and significantly speeds up training to become more than 10 times faster than other state-of-the-art DRL methods.
- D. Cao, W. Hu, J. Zhao, G. Zhang, B. Zhang, Z. Liu, Z. Chen, and F. Blaabjerg, “Reinforcement learning and its applications in modern power and energy systems: A review,” Journal of modern power systems and clean energy, vol. 8, no. 6, pp. 1029–1042, 2020.
- Q. Ma and C. Deng, “Simplified Deep Reinforcement Learning Based Volt-Var Control of Topologically Variable Power System,” Journal of Modern Power Systems and Clean Energy, 2023.
- T.-H. Fan, X. Y. Lee, and Y. Wang, “Powergym: A reinforcement learning environment for volt-var control in power distribution systems,” in Learning for Dynamics and Control Conference, pp. 21–33, PMLR, 2022.
- H. Liu, W. Wu, and Y. Wang, “Bi-level off-policy reinforcement learning for two-timescale volt/var control in active distribution networks,” IEEE Trans. Power Systems, vol. 38, no. 1, pp. 385–395, 2022.
- S. Gupta, A. Mehrizi-Sani, S. Chatzivasileiadis, and V. Kekatos, “Deep Learning for Scalable Optimal Design of Incremental Volt/VAR Control Rules,” IEEE Control Systems Letters, 2023.
- S. Manna, T. D. Loeffler, R. Batra, S. Banik, H. Chan, B. Varughese, K. Sasikumar, M. Sternberg, T. Peterka, M. J. Cherukara, et al., “Learning in continuous action space for developing high dimensional potential energy models,” Nature communications, vol. 13, no. 1, p. 368, 2022.
- F. Ming, F. Gao, K. Liu, and C. Zhao, “Cooperative modular reinforcement learning for large discrete action space problem,” Neural Networks, vol. 161, pp. 281–296, 2023.
- K. R. Williams, R. Schlossman, D. Whitten, J. Ingram, S. Musuvathy, J. Pagan, K. A. Williams, S. Green, A. Patel, A. Mazumdar, et al., “Trajectory planning with deep reinforcement learning in high-level action spaces,” IEEE Trans. Aerospace and Electronic Systems, 2022.
- A. Kanervisto, C. Scheller, and V. Hautamäki, “Action space shaping in deep reinforcement learning,” in 2020 IEEE conference on games (CoG), pp. 479–486, IEEE, 2020.
- Ray Team, “Ray Documentation — Libraries.” https://www.ray.io/libraries, 2023. Accessed: 2023-10-01.
- Anyscale Team, “Ray Distributed Computing - Anyscale.” https://www.anyscale.com/ray-distributed-computing, 2023. Accessed: 2023-10-01.
- L. Espeholt, H. Soyer, R. Munos, K. Simonyan, V. Mnih, T. Ward, Y. Doron, V. Firoiu, T. Harley, I. Dunning, et al., “Impala: Scalable distributed deep-rl with importance weighted actor-learner architectures,” in International conference on machine learning, pp. 1407–1416, PMLR, 2018.
- T. Haarnoja, A. Zhou, K. Hartikainen, G. Tucker, S. Ha, J. Tan, V. Kumar, H. Zhu, A. Gupta, P. Abbeel, et al., “Soft actor-critic algorithms and applications,” arXiv preprint arXiv:1812.05905, 2018.
- J. Schulman, F. Wolski, P. Dhariwal, A. Radford, and O. Klimov, “Proximal policy optimization algorithms,” arXiv preprint arXiv:1707.06347, 2017.
- D. Montenegro, M. Hernandez, and G. Ramos, “Real time OpenDSS framework for distribution systems simulation and analysis,” in 2012 Sixth IEEE/PES Transmission and Distribution: Latin America Conference and Exposition (T&D-LA), pp. 1–5, IEEE, 2012.
- A. Selim, J. Zhao, F. Ding, F. Miao, and S.-Y. Park, “Adaptive Deep Reinforcement Learning Algorithm for Distribution System Cyber Attack Defense With High Penetration of DERs,” IEEE Trans. Smart Grid, 2023.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.