Artificial Intelligence for Multi-Unit Auction design
Abstract: Understanding bidding behavior in multi-unit auctions remains an ongoing challenge for researchers. Despite their widespread use, theoretical insights into the bidding behavior, revenue ranking, and efficiency of commonly used multi-unit auctions are limited. This paper utilizes artificial intelligence, specifically reinforcement learning, as a model free learning approach to simulate bidding in three prominent multi-unit auctions employed in practice. We introduce six algorithms that are suitable for learning and bidding in multi-unit auctions and compare them using an illustrative example. This paper underscores the significance of using artificial intelligence in auction design, particularly in enhancing the design of multi-unit auctions.
- Ausubel, Lawrence M, Peter Cramton, Marek Pycia, Marzena Rostek, and Marek Weretka (2014) “Demand reduction and inefficiency in multi-unit auctions,” The Review of Economic Studies, 81 (4), 1366–1400.
- Back, Kerry and Jaime F Zender (1993) “Auctions of divisible goods: On the rationale for the treasury experiment,” The Review of Financial Studies, 6 (4), 733–764.
- Banchio, Martino and Andrzej Skrzypacz (2022) “Artificial intelligence and auction design,” in Proceedings of the 23rd ACM Conference on Economics and Computation, 30–31.
- Bellman, Richard (1957) Dynamic Programming: Princeton University Press.
- Bichler, Martin and Jacob K Goeree (2017) Handbook of spectrum auction design: Cambridge University Press.
- Cheng, Wen-Chung (Andy), Zhen Ni, and Xiangnan Zhong (2021) “Experimental Evaluation of Proximal Policy Optimization and Advantage Actor-Critic RL Algorithms using MiniGrid Environment,” in 34th Florida Conference on Recent Advances in Robotics (FCRAR 2021), Florida Atlantic University, Boca Raton, United States.
- Corecco, Samuel, Giorgia Adorni, and Luca Maria Gambardella (2023) “Proximal Policy Optimization-Based Reinforcement Learning and Hybrid Approaches to Explore the Cross Array Task Optimal Solution,” Machine Learning and Knowledge Extraction, 5 (4), 1660–1679, 10.3390/make5040082.
- Edelman, Benjamin, Michael Ostrovsky, and Michael Schwarz (2007) “Internet advertising and the generalized second-price auction: Selling billions of dollars worth of keywords,” American economic review, 97 (1), 242–259.
- Engelbrecht-Wiggans, Richard and Charles M Kahn (1998) “Multi-unit auctions with uniform prices,” Economic theory, 12, 227–258.
- Hailu, Atakelty and Sophie Thoyer (2007) “Designing Multi-unit Multiple Bid Auctions: An Agent-based Computational Model of Uniform, Discriminatory and Generalised Vickrey Auctions,” Economic Record, 83, S57–S72.
- Huang, Shengyi, Anssi Kanervisto, Antonin Raffin, Weixun Wang, Santiago Ontañón, and Rousslan Fernand Julien Dossa (2022) “A2C is a special case of PPO.”
- Khezr, Peyman and Anne Cumpston (2022) “A review of multiunit auctions with homogeneous goods,” Journal of Economic Surveys, 36 (4), 1225–1247.
- Khezr, Peyman and Flavio M Menezes (2017) “A new characterization of equilibrium in multiple-object uniform-price auctions,” Economics letters, 157, 53–55.
- Khezr, Peyman, Vijay Mohan, and Lionel Page (2024) “Strategic Bidding in Knapsack Auctions,” arXiv preprint arXiv:2403.07928.
- Mnih, Volodymyr, Koray Kavukcuoglu, David Silver et al. (2015) “Human-level control through deep reinforcement learning,” Nature, 518 (7540), 529–533, 10.1038/nature14236.
- Morgenstern, Jamie and Tim Roughgarden (2016) “Learning simple auctions,” in Conference on Learning Theory, 1298–1318, PMLR.
- Raffin, Antonin, Ashley Hill, Adam Gleave, Anssi Kanervisto, Maximilian Ernestus, and Noah Dormann (2021) “Stable-Baselines3: Reliable Reinforcement Learning Implementations,” Journal of Machine Learning Research, 22 (268), 1–8, http://jmlr.org/papers/v22/20-1364.html.
- Roughgarden, Tim, Vasilis Syrgkanis, and Eva Tardos (2017) “The price of anarchy in auctions,” Journal of Artificial Intelligence Research, 59, 59–101.
- Schulman, John, Filip Wolski, Prafulla Dhariwal, Alec Radford, and Oleg Klimov (2017) “Proximal policy optimization algorithms,” arXiv preprint arXiv:1707.06347.
- Silver, David, Satinder Singh, Doina Precup, and Richard S Sutton (2021) “Reward is enough,” Artificial Intelligence, 299, 103535.
- Tijsma, Arryon D., Madalina M. Drugan, and Marco A. Wiering (2016) “Comparing exploration strategies for Q-learning in random stochastic mazes,” in 2016 IEEE Symposium Series on Computational Intelligence (SSCI), 1–8, 10.1109/SSCI.2016.7849366.
- Towers, Mark, Jordan K. Terry, Ariel Kwiatkowski et al. (2023) “Gymnasium,” March, 10.5281/zenodo.8127026.
- Vickrey, William (1961) “Counterspeculation, auctions, and competitive sealed tenders,” The Journal of finance, 16 (1), 8–37.
- Williams, Ronald J. (1992) “Simple statistical gradient-following algorithms for connectionist reinforcement learning,” Machine Learning, 8 (3), 229–256, 10.1007/BF00992696.
- Zang, Yifan, Jinmin He, Kai Li, Haobo Fu, Qiang Fu, and Junliang Xing (2023) “Sequential Cooperative Multi-Agent Reinforcement Learning,” in Adaptive Agents and Multi-Agent Systems, https://api.semanticscholar.org/CorpusID:258845637.
- Zhu, Changxi, Mehdi Dastani, and Shihan Wang (2024) “A survey of multi-agent deep reinforcement learning with communication,” Autonomous Agents and Multi-Agent Systems, 38 (1), 4, 10.1007/s10458-023-09633-6.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.