[2603.03596] MEM: Multi-Scale Embodied Memory for Vision Language Action Models
About this article
Abstract page for arXiv paper 2603.03596: MEM: Multi-Scale Embodied Memory for Vision Language Action Models
Computer Science > Robotics arXiv:2603.03596 (cs) [Submitted on 4 Mar 2026] Title:MEM: Multi-Scale Embodied Memory for Vision Language Action Models Authors:Marcel Torne, Karl Pertsch, Homer Walke, Kyle Vedder, Suraj Nair, Brian Ichter, Allen Z. Ren, Haohuan Wang, Jiaming Tang, Kyle Stachowicz, Karan Dhabalia, Michael Equi, Quan Vuong, Jost Tobias Springenberg, Sergey Levine, Chelsea Finn, Danny Driess View a PDF of the paper titled MEM: Multi-Scale Embodied Memory for Vision Language Action Models, by Marcel Torne and 16 other authors View PDF HTML (experimental) Abstract:Conventionally, memory in end-to-end robotic learning involves inputting a sequence of past observations into the learned policy. However, in complex multi-stage real-world tasks, the robot's memory must represent past events at multiple levels of granularity: from long-term memory that captures abstracted semantic concepts (e.g., a robot cooking dinner should remember which stages of the recipe are already done) to short-term memory that captures recent events and compensates for occlusions (e.g., a robot remembering the object it wants to pick up once its arm occludes it). In this work, our main insight is that an effective memory architecture for long-horizon robotic control should combine multiple modalities to capture these different levels of abstraction. We introduce Multi-Scale Embodied Memory (MEM), an approach for mixed-modal long-horizon memory in robot policies. MEM combines video-based short...