[2604.03679] LightThinker++: From Reasoning Compression to Memory Management

[2604.03679] LightThinker++: From Reasoning Compression to Memory Management

arXiv - AI 4 min read

About this article

Abstract page for arXiv paper 2604.03679: LightThinker++: From Reasoning Compression to Memory Management

Computer Science > Computation and Language arXiv:2604.03679 (cs) [Submitted on 4 Apr 2026] Title:LightThinker++: From Reasoning Compression to Memory Management Authors:Yuqi Zhu, Jintian Zhang, Zhenjie Wan, Yujie Luo, Shuofei Qiao, Zhengke Gui, Da Zheng, Lei Liang, Huajun Chen, Ningyu Zhang View a PDF of the paper titled LightThinker++: From Reasoning Compression to Memory Management, by Yuqi Zhu and 9 other authors View PDF HTML (experimental) Abstract:Large language models (LLMs) excel at complex reasoning, yet their efficiency is limited by the surging cognitive overhead of long thought traces. In this paper, we propose LightThinker, a method that enables LLMs to dynamically compress intermediate thoughts into compact semantic representations. However, static compression often struggles with complex reasoning where the irreversible loss of intermediate details can lead to logical bottlenecks. To address this, we evolve the framework into LightThinker++, introducing Explicit Adaptive Memory Management. This paradigm shifts to behavioral-level management by incorporating explicit memory primitives, supported by a specialized trajectory synthesis pipeline to train purposeful memory scheduling. Extensive experiments demonstrate the framework's versatility across three dimensions. (1) LightThinker reduces peak token usage by 70% and inference time by 26% with minimal accuracy loss. (2) In standard reasoning, LightThinker++ slashes peak token usage by 69.9% while yielding a ...

Originally published on April 07, 2026. Curated by AI News.

Related Articles

[2603.07475] A Comparative analysis of Layer-wise Representational Capacity in AR and Diffusion LLMs
Llms

[2603.07475] A Comparative analysis of Layer-wise Representational Capacity in AR and Diffusion LLMs

Abstract page for arXiv paper 2603.07475: A Comparative analysis of Layer-wise Representational Capacity in AR and Diffusion LLMs

arXiv - Machine Learning · 3 min ·
[2601.22925] BEAR: Towards Beam-Search-Aware Optimization for Recommendation with Large Language Models
Llms

[2601.22925] BEAR: Towards Beam-Search-Aware Optimization for Recommendation with Large Language Models

Abstract page for arXiv paper 2601.22925: BEAR: Towards Beam-Search-Aware Optimization for Recommendation with Large Language Models

arXiv - Machine Learning · 4 min ·
[2512.10551] LLM-Auction: Generative Auction towards LLM-Native Advertising
Llms

[2512.10551] LLM-Auction: Generative Auction towards LLM-Native Advertising

Abstract page for arXiv paper 2512.10551: LLM-Auction: Generative Auction towards LLM-Native Advertising

arXiv - Machine Learning · 3 min ·
[2511.17411] SPEAR-1: Scaling Beyond Robot Demonstrations via 3D Understanding
Llms

[2511.17411] SPEAR-1: Scaling Beyond Robot Demonstrations via 3D Understanding

Abstract page for arXiv paper 2511.17411: SPEAR-1: Scaling Beyond Robot Demonstrations via 3D Understanding

arXiv - Machine Learning · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime