[2602.23387] Hello-Chat: Towards Realistic Social Audio Interactions
About this article
Abstract page for arXiv paper 2602.23387: Hello-Chat: Towards Realistic Social Audio Interactions
Computer Science > Sound arXiv:2602.23387 (cs) [Submitted on 16 Feb 2026] Title:Hello-Chat: Towards Realistic Social Audio Interactions Authors:Yueran Hou, Peilei Jia, Zihan Sun, Qihang Lu, Wenbing Yang, Yingming Gao, Ya Li, Jun Gao View a PDF of the paper titled Hello-Chat: Towards Realistic Social Audio Interactions, by Yueran Hou and 7 other authors View PDF HTML (experimental) Abstract:Recent advancements in Large Audio Language Models (LALMs) have demonstrated exceptional performance in speech recognition and translation. However, existing models often suffer from a disconnect between perception and expression, resulting in a robotic "read-speech" style that lacks the spontaneity and emotional resonance of real human interaction. In this report, we introduce Hello-Chat, an end-to-end audio language model designed for realistic social scenarios. By leveraging a massive dataset of real-life conversations and employing a modality-interleaved training strategy, Hello-Chat achieves a breakthrough in anthropomorphic generation. Experimental results show that our model not only reaches state-of-the-art (SOTA) performance on specific audio understanding tasks but also significantly outperforms existing baselines in prosodic naturalness and emotional alignment, paving the way for the next generation of empathetic AI agents. Subjects: Sound (cs.SD); Artificial Intelligence (cs.AI); Computation and Language (cs.CL); Audio and Speech Processing (eess.AS) Cite as: arXiv:2602.23387...