[2510.19268] Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models

[2510.19268] Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models

arXiv - Machine Learning 4 min read

About this article

Abstract page for arXiv paper 2510.19268: Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models

Computer Science > Robotics arXiv:2510.19268 (cs) [Submitted on 22 Oct 2025 (v1), last revised 15 Apr 2026 (this version, v2)] Title:Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models Authors:Mingen Li, Houjian Yu, Yixuan Huang, Youngjin Hong, Hantao Ye, Changhyun Choi View a PDF of the paper titled Hierarchical DLO Routing with Reinforcement Learning and In-Context Vision-language Models, by Mingen Li and 5 other authors View PDF HTML (experimental) Abstract:Long-horizon routing tasks of deformable linear objects (DLOs), such as cables and ropes, are common in industrial assembly lines and everyday life. These tasks are particularly challenging because they require robots to manipulate DLO with long-horizon planning and reliable skill execution. Successfully completing such tasks demands adapting to their nonlinear dynamics, decomposing abstract routing goals, and generating multi-step plans composed of multiple skills, all of which require accurate high-level reasoning during execution. In this paper, we propose a fully autonomous hierarchical framework for solving challenging DLO routing tasks. Given an implicit or explicit routing goal expressed in language, our framework leverages vision-language models~(VLMs) for in-context high-level reasoning to synthesize feasible plans, which are then executed by low-level skills trained via reinforcement learning. To improve robustness over long horizons, we further introduce a failure rec...

Originally published on April 16, 2026. Curated by AI News.

Related Articles

[2510.26491] Data-Efficient RLVR via Off-Policy Influence Guidance
Llms

[2510.26491] Data-Efficient RLVR via Off-Policy Influence Guidance

Abstract page for arXiv paper 2510.26491: Data-Efficient RLVR via Off-Policy Influence Guidance

arXiv - Machine Learning · 4 min ·
[2604.13327] Event Tensor: A Unified Abstraction for Compiling Dynamic Megakernel
Llms

[2604.13327] Event Tensor: A Unified Abstraction for Compiling Dynamic Megakernel

Abstract page for arXiv paper 2604.13327: Event Tensor: A Unified Abstraction for Compiling Dynamic Megakernel

arXiv - Machine Learning · 3 min ·
[2604.13275] Better and Worse with Scale: How Contextual Entrainment Diverges with Model Size
Llms

[2604.13275] Better and Worse with Scale: How Contextual Entrainment Diverges with Model Size

Abstract page for arXiv paper 2604.13275: Better and Worse with Scale: How Contextual Entrainment Diverges with Model Size

arXiv - Machine Learning · 4 min ·
[2604.13068] Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive Language Models
Llms

[2604.13068] Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive Language Models

Abstract page for arXiv paper 2604.13068: Before the First Token: Scale-Dependent Emergence of Hallucination Signals in Autoregressive La...

arXiv - Machine Learning · 4 min ·
More in Llms: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime