Generative robot policies such as Flow Matching offer flexible, multi-modal policy learning but are sample-inefficient. Although object-centric policies improve sample efficiency, it does not resolve this limitation. In this work, we propose Multi-Stream Generative Policy (MSG), an inference-time composition framework that trains multiple object-centric policies and combines them at inference to improve generalization and sample efficiency. MSG is model-agnostic and inference-only, hence widely applicable to various generative policies and training paradigms. We perform extensive experiments both in simulation and on a real robot, demonstrating that our approach learns high-quality generative policies from as few as five demonstrations, resulting in a 95% reduction in demonstrations, and improves policy performance by 89 percent compared to single-stream approaches. Furthermore, we present comprehensive ablation studies on various composition strategies and provide practical recommendations for deployment. Finally, MSG enables zero-shot object instance transfer. We make our code publicly available at https://msg.cs.uni-freiburg.de.
翻译:基于流匹配等生成式机器人策略虽能实现灵活的多模态策略学习,但其样本利用效率较低。尽管以物体为中心的策略提升了样本效率,但并未从根本上解决这一局限。本文提出多流生成策略(MSG),这是一种推理时组合框架,通过训练多个以物体为中心的策略并在推理时将其组合,以提升泛化能力与样本效率。MSG具有模型无关性与纯推理特性,因而可广泛适用于各类生成式策略与训练范式。我们在仿真环境与真实机器人上进行了大量实验,结果表明该方法仅需五条演示样本即可学习到高质量的生成策略,演示需求降低95%,且相比单流方法策略性能提升89%。此外,我们对多种组合策略开展了系统的消融研究,并为实际部署提供了实用建议。最后,MSG能够实现零样本物体实例迁移。相关代码已公开于 https://msg.cs.uni-freiburg.de。