[2505.13820] Structured Agent Distillation for Large Language Model
About this article
Abstract page for arXiv paper 2505.13820: Structured Agent Distillation for Large Language Model
Computer Science > Machine Learning arXiv:2505.13820 (cs) [Submitted on 20 May 2025 (v1), last revised 28 Mar 2026 (this version, v4)] Title:Structured Agent Distillation for Large Language Model Authors:Jun Liu, Zhenglun Kong, Peiyan Dong, Changdi Yang, Tianqi Li, Hao Tang, Geng Yuan, Wei Niu, Wenbin Zhang, Pu Zhao, Xue Lin, Dong Huang, Yanzhi Wang View a PDF of the paper titled Structured Agent Distillation for Large Language Model, by Jun Liu and 11 other authors View PDF HTML (experimental) Abstract:Large language models (LLMs) exhibit strong capabilities as decision-making agents by interleaving reasoning and actions, as seen in ReAct-style frameworks. Yet, their practical deployment is constrained by high inference costs and large model sizes. We propose Structured Agent Distillation, a framework that compresses large LLM-based agents into smaller student models while preserving both reasoning fidelity and action consistency. Unlike standard token-level distillation, our method segments trajectories into {[REASON]} and {[ACT]} spans, applying segment-specific losses to align each component with the teacher's behavior. This structure-aware supervision enables compact agents to better replicate the teacher's decision process. Experiments on ALFWorld, HotPotQA-ReAct, and WebShop show that our approach consistently outperforms token-level and imitation learning baselines, achieving significant compression with minimal performance drop. Scaling and ablation results furthe...