Papers
Topics
Authors
Recent
Search
2000 character limit reached

Multi-Agent Multi-Armed Bandits

Updated 22 January 2026
  • MA-MAB is a framework that extends the classical bandit model to include multiple agents interacting over shared arms with individualized rewards.
  • The model supports centralized or decentralized control, integrating fairness, collision avoidance, and efficiency through tailored regret measures.
  • Advanced algorithms leverage UCB, consensus schemes, and linear programming to optimize performance in dynamic, large-scale, and adversarial environments.

A multi-agent multi-armed bandit (MA-MAB) problem generalizes the classical stochastic bandit model by introducing multiple agents, each of whom may experience individualized rewards and potentially interact, coordinate, or compete according to a variety of information, resource, or fairness constraints. The MA-MAB framework encompasses settings with centralized or decentralized control, diverse sharing and communication topologies, and a spectrum of fairness, efficiency, and robustness targets. Rigorous theoretical results characterize regret—relative to optimal (possibly fair) policies—under realistic operational constraints, and recent advances have produced efficient algorithms with strong empirical and analytic performance guarantees in large-scale, heterogeneous, and dynamic environments.

1. Formal MA-MAB Model and Regret Metrics

In MA-MAB models, nn agents interact with mm arms over a time horizon TT. Each pull yields a random vector Xi,jtX^t_{i,j}, where i=1,,ni = 1, \ldots, n, j=1,,mj = 1, \ldots, m, and Xi,jtD(μi,j)X^t_{i,j} \sim D(\mu_{i,j}). The mean reward matrix A[0,1]n×mA \in [0,1]^{n \times m} with entries Ai,j=μi,jA_{i,j} = \mu_{i,j} is unknown to the learner. The system may be controlled by a centralized mechanism or comprise fully decentralized autonomous agents. Performance is assessed by cumulative regret, either:

  • Social Welfare Regret: RSW(T)=TW(π)t=1TW(πt)R_\text{SW}(T) = T\,W(\pi^*) - \sum_{t=1}^T W(\pi^t), where W(πt)=i=1nj=1mπjtμi,jW(\pi^t) = \sum_{i=1}^n \sum_{j=1}^m \pi^t_j \mu_{i,j}.
  • Per-Agent (Individual) Regret: Ri(T)=E[t=1T(μiμi,πi(t))]R_i(T) = E[\sum_{t=1}^T (\mu^*_{i} - \mu_{i,\pi_i(t)})], where πi(t)\pi_i(t) is the arm for agent ii at time tt.
  • Fairness Regret: Measures such as RF(T)=t=1Ti=1n[CiAiAi,πt]+R_F(T) = \sum_{t=1}^T \sum_{i=1}^n [C_i A_i - \langle A_i,\pi^t \rangle]_+ for explicit per-agent guarantees, or Nash-regret when optimizing Nash social welfare (Manupriya et al., 21 Feb 2025, Hossain et al., 2020, Caiata et al., 15 Jan 2026).

These metrics are chosen based on application-driven priorities—maximizing system throughput, balancing individual outcomes, or enforcing policy-based constraints.

2. Fairness Notions and Algorithmic Frameworks

Multi-agent bandit research has formalized several fairness paradigms:

  • Minimum-Reward Guarantee: Each agent ii must receive at least a CiC_i-fraction of its maximum achievable mean, operationalized via a policy πΔm\pi \in \Delta_m such that AπCAA\pi \geq C \odot A (entrywise). Social welfare and fairness regret track deviation from optimal fair policies (Manupriya et al., 21 Feb 2025).
  • Nash Social Welfare (NSW): NSW(π;A)=i=1nUi(π)1/nNSW(\pi;A) = \prod_{i=1}^n U_i(\pi)^{1/n}, a balance of fairness and efficiency, optimized using UCB-based or explore-first algorithms that deliver O(T)O(\sqrt{T}) Nash-regret (Hossain et al., 2020, Xu et al., 17 Jun 2025).
  • Procedural Fairness: Equal voice in determining the policy, requiring that for every agent, the share of decision-making over their favorite arms matches $1/n$ and is immune to blocking by subcoalitions—a departure from purely outcome-based metrics (Caiata et al., 15 Jan 2026).
  • Collision Fairness: In distributed settings where simultaneous pulls by multiple agents on a single arm result in collisions (zero reward), "fair" assignments must coordinate to allocate top arms without overlap and balance individual and group regret (Zhou et al., 8 Oct 2025).

Algorithms incorporate fairness by explicit constraints in linear programs (RewardFairUCB), weighted social objectives (OFMUP in probing settings), Nash/fairness-optimal policies, or by guarantee structures that align process and coalition stability.

3. Algorithmic Approaches: Centralized, Decentralized, and Communication Constraints

Centralized Approaches

In centralized MA-MAB, a single decision-maker controls arm allocation or sampling distributions for all agents, solving per-round convex or linear programs subject to utility/fairness constraints. UCB and LCB indices support exploration-exploitation, as in RewardFairUCB, where per-arm confidence intervals are integrated into an LP for both social welfare maximization and fairness (Manupriya et al., 21 Feb 2025).

Decentralized and Communication-Constrained Schemes

Decentralized MA-MAB models feature autonomous agents with limited local reward information and restricted communication—over fixed or time-varying (random or protocol-driven) graphs:

  • Consensus-Based UCB: Agents maintain local and shared running estimates using consensus schemes on static graphs; UCB indices drive arm selection, achieving group regrets asymptotically matching centralized policies, with spectral properties of communication topology determining additive regret penalties (Landgren et al., 2016, Landgren et al., 2020). Extensions handle selection with collisions, time-varying graphs, and heavy-tailed rewards (Wang et al., 31 Jan 2025, Sankararaman et al., 2019).
  • UCB with Strategic Communication: Agents select communication partners by UCB-based heuristics, focus messaging on informative (still-exploring) peers, and constrain per-round messaging to control communication load, achieving O(logT)O(\log T) regret with small per-link message size (Pankayaraj et al., 2019, Pankayaraj et al., 2020).
  • Limited or Private Sharing: Regimes where agents share only a subset of information (e.g., non-sensitive arms) require balancing exploration to avoid over/under-compensation among arms and incentivizing participation through explicit payoff mechanisms. Balanced-ETC ensures individual rationality and asymptotic optimality under limited information (Shao et al., 21 Feb 2025).

Robust and Secure Decentralized Learning

Adversarial settings with potentially malicious or Byzantine agents are handled via:

  • Blocking/Filtering: Honest agents detect and block unreliable or malicious peers using local evidence, restoring collaboration benefits and maintaining O(logT)O(\log T) individual regret as long as the adversarial population is sublinear in problem size (Vial et al., 2020).
  • Blockchain-Integrated Consensus: Secure multiparty computation, validator sorting, and digital signatures underpin distributed consensus mechanisms for robust learning, guaranteeing O(logT)O(\log T) regret for honest agents under majority-honest assumptions and explicit incentives enforcing truthful participation (Xu et al., 2024).

4. Extensions: Capacity, Dynamics, and Advanced Models

Advanced MA-MAB variants expand applicability and capture richer dynamics:

  • Stochastic Arm Capacities: Models that include random availability of arm capacity (requests per arm per round) and distributed protocols to consistently converge to optimal arm profiles. Explore-then-commit frameworks, consensus protocols, and constant-round distributed committing enable O(logT)O(\log T) regret and scalable solution synthesis (Xie et al., 2024).
  • Probing Frameworks: Strategic selection of arms to sample before allocation (probing) couples offline submodular maximization (with theoretical approximation guarantees) and online UCB-based scheduling to accelerate high-reward identification and ensure high Nash welfare (Xu et al., 17 Jun 2025).
  • Cournot Games and Ordered Action Spaces: Casting economic oligopoly games as MA-MAB exposes emergent collusive equilibria, and structure-exploiting heuristics (bucket refinement, elimination) reduce regret and support efficient convergence (Taywade et al., 2022).
  • Heavy-Tailed and Sparse Communication: Under heavy-tailed degree distributions, hub-based estimators and information-delay bounds ensure sublinear regret even with infinite-variance rewards and extremely sparse communication networks (Wang et al., 31 Jan 2025).
  • Change Points and Dynamic Nonstationarity: MA-MAB with piecewise-stationary arms deploys Bayesian online change point detection and cooperative restart rules (based on neighborhood voting) to maintain provably low regret in highly nonstationary settings (Cheng et al., 2023).

5. Empirical Results and System-Level Applications

Empirical investigation across synthetic and real datasets (MovieLens, ridesharing, Wi-Fi spatial reuse, digital marketing logs) consistently demonstrates:

  • Significant reductions in regret for algorithms exploiting structured information sharing, consensus, or fairness-aware allocation (Manupriya et al., 21 Feb 2025, Xu et al., 17 Jun 2025).
  • Robustness of collaboration protocols (e.g., blocking, limited sharing) to malicious outliers and privacy-constrained agents (Vial et al., 2020, Shao et al., 21 Feb 2025).
  • Efficient and scalable operation under large agent and arm populations, with communication and computation requirements compatible with real-world large-scale deployment.
  • Concrete system improvements such as throughput gains and fairness in spatial-reuse Wi-Fi scheduling and collusion-aware economic games (Wilhelmi et al., 2024, Taywade et al., 2022).

6. Open Problems and Future Directions

Prominent research challenges include:

7. Comparative Table of Representative MA-MAB Settings

Model/Algorithm Key Features Main Regret Bound
RewardFairUCB (Manupriya et al., 21 Feb 2025) Social welfare + min-guarantee fairness O~(T)\tilde O(\sqrt{T}) (SW), O~(T3/4)\tilde O(T^{3/4}) (fairness)
NashUCB/Explore-first (Hossain et al., 2020) Nash social welfare fairness O~(T)\tilde O(\sqrt{T}) Nash-regret
Procedural Fairness (Caiata et al., 15 Jan 2026) Proportional voice, core stability O(Tγ+(lnT)1/γ)O(T^\gamma + (\ln T)^{1/\gamma})
Balanced-ETC (Shao et al., 21 Feb 2025) Limited/private sharing + incentives O((logT)/Δi)O(\sum (\log T)/\Delta_i) per-agent
SynCD (Collision) (Zhou et al., 8 Oct 2025) Fully distributed, collision-aware O((K/M)logT)O(\sum (K/M)\log T) (individual)
BatchSP2 (Hanna et al., 2023) Erasure channels, scheduling O(poly(logT))O(\text{poly}(\log T)) robust regret
Heavy-tailed Decentralized (Wang et al., 31 Jan 2025) Sparse graph, infinite variance O(M11/αlogT)O(M^{1-1/\alpha}\log T)

This table summarizes configurational diversity in agent–arm structures, communication/interference constraints, fairness paradigms, and corresponding regret guarantees.


Research on MA-MAB is distinguished by a rich interaction between algorithmic innovation, information/communication structure, and explicit social or procedural constraints, addressing both foundational and application-specific challenges across distributed learning, fairness in automated decision making, and strategic interaction under uncertainty.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (19)

Topic to Video (Beta)

No one has generated a video about this topic yet.

Whiteboard

No one has generated a whiteboard explanation for this topic yet.

Follow Topic

Get notified by email when new papers are published related to Multi-Agent Multi-Armed Bandits (MA-MAB).