[2511.05640] Blind Inverse Game Theory: Jointly Decoding Rewards and Rationality in Entropy-Regularized Competitive Games

[2511.05640] Blind Inverse Game Theory: Jointly Decoding Rewards and Rationality in Entropy-Regularized Competitive Games

arXiv - Machine Learning 4 min read Article

Summary

This paper presents Blind Inverse Game Theory, a novel framework for jointly decoding rewards and rationality in entropy-regularized competitive games, addressing the challenge of unknown rationality parameters.

Why It Matters

The study is significant as it introduces a new statistical framework that resolves the scale ambiguity in Inverse Game Theory, allowing for better parameter recovery in competitive settings. This advancement can enhance the understanding of agent behavior in various applications, including economics and AI.

Key Takeaways

  • Introduces Blind-IGT for joint recovery of reward and rationality parameters.
  • Addresses scale ambiguity in Inverse Game Theory with a normalization constraint.
  • Proposes an efficient Normalized Least Squares estimator with optimal convergence rates.
  • Extends the framework to Markov games, demonstrating strong empirical performance.
  • Provides partial identification guarantees when strong identifiability conditions fail.

Computer Science > Machine Learning arXiv:2511.05640 (cs) [Submitted on 7 Nov 2025 (v1), last revised 21 Feb 2026 (this version, v2)] Title:Blind Inverse Game Theory: Jointly Decoding Rewards and Rationality in Entropy-Regularized Competitive Games Authors:Hamza Virk, Sandro Amaglobeli, Zuhayr Syed View a PDF of the paper titled Blind Inverse Game Theory: Jointly Decoding Rewards and Rationality in Entropy-Regularized Competitive Games, by Hamza Virk and 2 other authors View PDF HTML (experimental) Abstract:Inverse Game Theory (IGT) methods based on the entropy-regularized Quantal Response Equilibrium (QRE) offer a tractable approach for competitive settings, but critically assume the agents' rationality parameter (temperature $\tau$) is known a priori. When $\tau$ is unknown, a fundamental scale ambiguity emerges that couples $\tau$ with the reward parameters ($\theta$), making them statistically unidentifiable. We introduce Blind-IGT, the first statistical framework to jointly recover both $\theta$ and $\tau$ from observed behavior. We analyze this bilinear inverse problem and establish necessary and sufficient conditions for unique identification by introducing a normalization constraint that resolves the scale ambiguity. We propose an efficient Normalized Least Squares (NLS) estimator and prove it achieves the optimal $\mathcal{O}(N^{-1/2})$ convergence rate for joint parameter recovery. When strong identifiability conditions fail, we provide partial identification gua...

Related Articles

Machine Learning

[D] ICML Reviewer Acknowledgement

Hi, I'm a little confused about ICML discussion period Does the period for reviewer acknowledging responses have already ended? One of th...

Reddit - Machine Learning · 1 min ·
Llms

Claude Opus 4.6 API at 40% below Anthropic pricing – try free before you pay anything

Hey everyone I've set up a self-hosted API gateway using [New-API](QuantumNous/new-ap) to manage and distribute Claude Opus 4.6 access ac...

Reddit - Artificial Intelligence · 1 min ·
Machine Learning

[D] ICML reviewer making up false claim in acknowledgement, what to do?

In a rebuttal acknowledgement we received, the reviewer made up a claim that our method performs worse than baselines with some hyperpara...

Reddit - Machine Learning · 1 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
More in Machine Learning: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime