[2510.01051] GEM: A Gym for Agentic LLMs
About this article
Abstract page for arXiv paper 2510.01051: GEM: A Gym for Agentic LLMs
Computer Science > Machine Learning arXiv:2510.01051 (cs) [Submitted on 1 Oct 2025 (v1), last revised 1 Mar 2026 (this version, v2)] Title:GEM: A Gym for Agentic LLMs Authors:Zichen Liu, Anya Sims, Keyu Duan, Changyu Chen, Simon Yu, Xiangxin Zhou, Haotian Xu, Shaopan Xiong, Bo Liu, Chenmien Tan, Chuen Yang Beh, Weixun Wang, Hao Zhu, Weiyan Shi, Diyi Yang, Michael Shieh, Yee Whye Teh, Wee Sun Lee, Min Lin View a PDF of the paper titled GEM: A Gym for Agentic LLMs, by Zichen Liu and 18 other authors View PDF HTML (experimental) Abstract:The training paradigm for large language models (LLMs) is moving from static datasets to experience-based learning, where agents acquire skills via interacting with complex environments. To facilitate this transition we introduce GEM (General Experience Maker), an open-source environment simulator designed for the age of LLMs. Analogous to OpenAI-Gym for traditional reinforcement learning (RL), GEM provides a standardized framework for the environment-agent interface, including asynchronous vectorized execution for high throughput, and flexible wrappers for easy extensibility. GEM also features a diverse suite of environments, robust integrated tools, and single-file example scripts demonstrating using GEM with five popular RL training frameworks. Along with this, we also provide a set of baselines across 24 environments using REINFORCE with Return Batch Normalization (ReBN), which -- unlike GRPO -- is compatible with the full RL setting of d...