Dream2Flow: Bridging Video Generation and Robotic Manipulation

Published:Dec 31, 2025 10:25
1 min read
ArXiv

Analysis

This paper introduces Dream2Flow, a novel framework that leverages video generation models to enable zero-shot robotic manipulation. The core idea is to use 3D object flow as an intermediate representation, bridging the gap between high-level video understanding and low-level robotic control. This approach allows the system to manipulate diverse object categories without task-specific demonstrations, offering a promising solution for open-world robotic manipulation.

Reference

Dream2Flow overcomes the embodiment gap and enables zero-shot guidance from pre-trained video models to manipulate objects of diverse categories-including rigid, articulated, deformable, and granular.