Recently, quadrupedal locomotion has achieved significant success, but their manipulation capabilities, particularly in handling large objects, remain limited, restricting their usefulness in demanding real-world applications such as search and rescue, construction, industrial automation, and room organization. This paper tackles the task of obstacle-aware, long-horizon pushing by multiple quadrupedal robots. We propose a hierarchical multi-agent reinforcement learning framework with three levels of control. The high-level controller integrates an RRT planner and a centralized adaptive policy to generate subgoals, while the mid-level controller uses a decentralized goal-conditioned policy to guide the robots toward these sub-goals. A pre-trained low-level locomotion policy executes the movement commands. We evaluate our method against several baselines in simulation, demonstrating significant improvements over baseline approaches, with 36.0% higher success rates and 24.5% reduction in completion time than the best baseline. Our framework successfully enables long-horizon, obstacle-aware manipulation tasks like Push-Cuboid and Push-T on Go1 robots in the real world.
翻译:近年来,四足机器人运动控制已取得显著进展,但其操作能力,尤其是在处理大型物体方面,仍然有限,这限制了它们在搜救、建筑施工、工业自动化和室内整理等高要求实际应用中的实用性。本文研究了多台四足机器人在障碍物环境下的长时程推动任务。我们提出了一种包含三层控制的分层多智能体强化学习框架:高层控制器集成RRT规划器与集中式自适应策略以生成子目标;中层控制器采用分散式目标条件策略引导机器人朝向这些子目标运动;预训练的低层运动策略则负责执行具体移动指令。我们在仿真环境中将所提方法与多种基线方法进行比较评估,结果表明该方法较基线方法有显著提升——相比最佳基线方法,成功率提高36.0%,任务完成时间减少24.5%。该框架成功在现实世界的Go1机器人上实现了如立方体推动和T型物体推动等长时程、障碍物感知的操作任务。