[2603.20843] HiCI: Hierarchical Construction-Integration for Long-Context Attention
About this article
Abstract page for arXiv paper 2603.20843: HiCI: Hierarchical Construction-Integration for Long-Context Attention
Computer Science > Computation and Language arXiv:2603.20843 (cs) [Submitted on 21 Mar 2026] Title:HiCI: Hierarchical Construction-Integration for Long-Context Attention Authors:Xiangyu Zeng, Qi Xu, Yunke Wang, Chang Xu View a PDF of the paper titled HiCI: Hierarchical Construction-Integration for Long-Context Attention, by Xiangyu Zeng and 3 other authors View PDF HTML (experimental) Abstract:Long-context language modeling is commonly framed as a scalability challenge of token-level attention, yet local-to-global information structuring remains largely implicit in existing approaches. Drawing on cognitive theories of discourse comprehension, we propose HiCI (Hierarchical Construction--Integration), a hierarchical attention module that constructs segment-level representations, integrates them into a shared global context, and broadcasts both to condition segment-level attention. We validate HiCI through parameter-efficient adaptation of LLaMA-2 with only <5.5% additional parameters, extending context from 4K to 100K tokens (7B) and 64K tokens (13B). Across language modeling, retrieval, and instruction-following benchmarks, HiCI yields consistent improvements over strong baselines, including matching proprietary models on topic retrieval and surpassing GPT-3.5-Turbo-16K on code comprehension. These results demonstrate the effectiveness of explicit hierarchical structuring as an inductive bias for long-context modeling. Comments: Subjects: Computation and Language (cs.CL); A...