[2602.12517] Bench-MFG: A Benchmark Suite for Learning in Stationary Mean Field Games

[2602.12517] Bench-MFG: A Benchmark Suite for Learning in Stationary Mean Field Games

arXiv - Machine Learning 4 min read Article

Summary

The paper presents Bench-MFG, a benchmark suite designed to standardize evaluations in learning for stationary Mean Field Games, addressing fragmentation in current methodologies.

Why It Matters

As Mean Field Games (MFGs) intersect with Reinforcement Learning (RL), a lack of standardized evaluation protocols hampers the assessment of new algorithms. Bench-MFG aims to fill this gap, providing a structured approach to evaluate and compare various learning methods, which is crucial for advancing research in multi-agent systems.

Key Takeaways

  • Bench-MFG introduces a comprehensive benchmark suite for evaluating algorithms in stationary Mean Field Games.
  • The suite includes a taxonomy of problem classes to facilitate diverse testing environments.
  • MF-Garnets is proposed for generating random MFG instances, enhancing statistical testing rigor.
  • The paper benchmarks various learning algorithms, including a novel black-box approach for exploitability minimization.
  • Guidelines for standardizing future experimental comparisons are provided based on empirical results.

Computer Science > Machine Learning arXiv:2602.12517 (cs) [Submitted on 13 Feb 2026] Title:Bench-MFG: A Benchmark Suite for Learning in Stationary Mean Field Games Authors:Lorenzo Magnino, Jiacheng Shen, Matthieu Geist, Olivier Pietquin, Mathieu Laurière View a PDF of the paper titled Bench-MFG: A Benchmark Suite for Learning in Stationary Mean Field Games, by Lorenzo Magnino and 4 other authors View PDF HTML (experimental) Abstract:The intersection of Mean Field Games (MFGs) and Reinforcement Learning (RL) has fostered a growing family of algorithms designed to solve large-scale multi-agent systems. However, the field currently lacks a standardized evaluation protocol, forcing researchers to rely on bespoke, isolated, and often simplistic environments. This fragmentation makes it difficult to assess the robustness, generalization, and failure modes of emerging methods. To address this gap, we propose a comprehensive benchmark suite for MFGs (Bench-MFG), focusing on the discrete-time, discrete-space, stationary setting for the sake of clarity. We introduce a taxonomy of problem classes, ranging from no-interaction and monotone games to potential and dynamics-coupled games, and provide prototypical environments for each. Furthermore, we propose MF-Garnets, a method for generating random MFG instances to facilitate rigorous statistical testing. We benchmark a variety of learning algorithms across these environments, including a novel black-box approach (MF-PSO) for exploitab...

Related Articles

Nlp

McKinsey's AI Lie Explains What's Happening to Work

Everyone thinks McKinsey just built 25,000 AI experts. They didn't. They took a 35-year-old internal database, put a natural language int...

Reddit - Artificial Intelligence · 1 min ·
Generative Ai

Midjourney has a new offer on the cancel page there is 20 off for 2 months

submitted by /u/RainDragonfly826 [link] [comments]

Reddit - Artificial Intelligence · 1 min ·
Walmart CEO reportedly brags that company's in-app AI agent is making people spend 35% more money
Nlp

Walmart CEO reportedly brags that company's in-app AI agent is making people spend 35% more money

AI Tools & Products · 4 min ·
Llms

[R] Looking for arXiv cs.LG endorser, inference monitoring using information geometry

Hi r/MachineLearning, I’m looking for an arXiv endorser in cs.LG for a paper on inference-time distribution shift detection for deployed ...

Reddit - Machine Learning · 1 min ·
More in Nlp: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime