Coordinating a team of robots to reposition multiple objects in cluttered environments requires reasoning jointly about where robots should establish contact, how to manipulate objects once contact is made, and how to navigate safely and efficiently at scale. Prior approaches typically fall into two extremes -- either learning the entire task or relying on privileged information and hand-designed planners -- both of which struggle to handle diverse objects in long-horizon tasks. To address these challenges, we present a unified framework for collaborative multi-robot, multi-object non-prehensile manipulation that integrates flow-matching co-generation with anonymous multi-robot motion planning. Within this framework, a generative model co-generates contact formations and manipulation trajectories from visual observations, while a novel motion planner conveys robots at scale. Crucially, the same planner also supports coordination at the object level, assigning manipulated objects to larger target structures and thereby unifying robot- and object-level reasoning within a single algorithmic framework. Experiments in challenging simulated environments demonstrate that our approach outperforms baselines in both motion planning and manipulation tasks, highlighting the benefits of generative co-design and integrated planning for scaling collaborative manipulation to complex multi-agent, multi-object settings. Visit gco-paper.github.io for code and demonstrations.
翻译:在杂乱环境中协调机器人团队对多个物体进行重新定位,需要联合推理机器人应在何处建立接触、建立接触后如何操作物体,以及如何在大规模场景中安全高效地导航。现有方法通常陷入两个极端——要么学习整个任务,要么依赖特权信息和人工设计的规划器——两者都难以在长时程任务中处理多样化的物体。为解决这些挑战,我们提出了一个统一的协作式多机器人、多物体非抓取操作框架,该框架将流匹配协同生成与匿名多机器人运动规划相结合。在此框架中,生成模型根据视觉观测协同生成接触构型与操作轨迹,而新型运动规划器则实现大规模机器人调度。关键的是,该规划器同时支持物体层面的协调,将操作对象分配给更大的目标结构,从而在单一算法框架内统一了机器人层面与物体层面的推理。在具有挑战性的仿真环境中的实验表明,我们的方法在运动规划和操作任务上均优于基线,凸显了生成式协同设计与集成规划在将协作操作扩展到复杂多智能体、多物体场景中的优势。代码与演示请访问 gco-paper.github.io。