Papers
Topics
Authors
Recent
Search
2000 character limit reached

Deep W-Networks: Solving Multi-Objective Optimisation Problems With Deep Reinforcement Learning

Published 9 Nov 2022 in cs.LG and math.OC | (2211.04813v2)

Abstract: In this paper, we build on advances introduced by the Deep Q-Networks (DQN) approach to extend the multi-objective tabular Reinforcement Learning (RL) algorithm W-learning to large state spaces. W-learning algorithm can naturally solve the competition between multiple single policies in multi-objective environments. However, the tabular version does not scale well to environments with large state spaces. To address this issue, we replace underlying Q-tables with DQN, and propose an addition of W-Networks, as a replacement for tabular weights (W) representations. We evaluate the resulting Deep W-Networks (DWN) approach in two widely-accepted multi-objective RL benchmarks: deep sea treasure and multi-objective mountain car. We show that DWN solves the competition between multiple policies while outperforming the baseline in the form of a DQN solution. Additionally, we demonstrate that the proposed algorithm can find the Pareto front in both tested environments.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (21)
  1. Dynamic weights in multi-objective deep reinforcement learning. In International Conference on Machine Learning, pages 11–20. PMLR.
  2. Learning run-time compositions of interacting adaptations. SEAMS ’20, page 108–114, New York, NY, USA. Association for Computing Machinery.
  3. Maximizing renewable energy use with decentralized residential demand response. In 2015 IEEE First International Smart Cities Conference (ISC2), pages 1–6.
  4. A novel joint radio resource management approach with reinforcement learning mechanisms. In IEEE International Performance, Computing, and Communications Conference (IPCCC), pages 621–626. Phoenix, AZ, USA.
  5. Energy Aware Deep Reinforcement Learning Scheduling for Sensors Correlated in Time and Space. IEEE Internet of Things Journal, 9(9):6732–6744.
  6. Humphrys, M. (1995). W-learning: Competition among selfish Q-learners.
  7. Pareto-Based Multiobjective Machine Learning: An Overview and Case Studies. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 38(3):397–415.
  8. Karlsson, J. (1997). Learning to solve multiple goals. University of Rochester.
  9. Kauten, C. (2018). Super Mario Bros for OpenAI Gym. GitHub: github.com/Kautenja/gym-super-mario-bros.
  10. Spatial-temporal traffic flow control on motorways using distributed multi-agent reinforcement learning. Mathematics - Special Issue Advances in Artificial Intelligence: Models, Optimization, and Machine Learning, 9(23).
  11. Deep learning. nature, 521(7553):436–444.
  12. Multiobjective Reinforcement Learning: A Comprehensive Overview. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 45(3):385–398.
  13. Playing Atari With Deep Reinforcement Learning. arXiv preprint arXiv:1312.5602.
  14. Multi-Objective Deep Reinforcement Learning. arXiv preprint arXiv:1610.02707.
  15. A multi-objective deep reinforcement learning framework. Engineering Applications of Artificial Intelligence, 96:103915.
  16. Prioritized experience replay. Presented at International Conference on Learning Representations (ICLR), San Diego, CA, May 7–9, 2015. arXiv preprint 1511.05952.
  17. Multiple-goal reinforcement learning with modular sarsa(0). In 18th Int. Joint Conf. Artif. Intell., page 1445–1447.
  18. Tajmajer, T. (2018). Modular multi-objective deep reinforcement learning with decision values. In 2018 Federated conference on computer science and information systems (FedCSIS), pages 85–93. IEEE.
  19. Empirical evaluation methods for multiobjective reinforcement learning algorithms. Machine learning, 84(1):51–80.
  20. Multi-objective reinforcement learning for infectious disease control with application to COVID-19 spread. arXiv preprint arXiv:2009.04607.
  21. Dueling network architectures for deep reinforcement learning. In Proceedings of Machine Learning Research (PMLR), vol.48, pages 1995–2003. New York, USA.
Citations (1)

Summary

No one has generated a summary of this paper yet.

Paper to Video (Beta)

No one has generated a video about this paper yet.

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.