We introduce Constrained Human-AI Cooperation (CHAIC), an inclusive embodied social intelligence challenge designed to test social perception and cooperation in embodied agents. In CHAIC, the goal is for an embodied agent equipped with egocentric observations to assist a human who may be operating under physical constraints -- e.g., unable to reach high places or confined to a wheelchair -- in performing common household or outdoor tasks as efficiently as possible. To achieve this, a successful helper must: (1) infer the human's intents and constraints by following the human and observing their behaviors (social perception), and (2) make a cooperative plan tailored to the human partner to solve the task as quickly as possible, working together as a team (cooperative planning). To benchmark this challenge, we create four new agents with real physical constraints and eight long-horizon tasks featuring both indoor and outdoor scenes with various constraints, emergency events, and potential risks. We benchmark planning- and learning-based baselines on the challenge and introduce a new method that leverages large language models and behavior modeling. Empirical evaluations demonstrate the effectiveness of our benchmark in enabling systematic assessment of key aspects of machine social intelligence. Our benchmark and code are publicly available at this URL: https://github.com/UMass-Foundation-Model/CHAIC.
翻译:我们提出受限人机协作(CHAIC),这是一项旨在测试具身智能体社会感知与协作能力的包容性具身社会智能挑战。在CHAIC中,目标是通过配备以自我为中心观测能力的具身智能体,协助可能受物理条件限制(例如无法触及高处或被限制在轮椅上)的人类,以尽可能高效的方式完成日常家庭或户外任务。为实现这一目标,成功的协助者必须:(1)通过跟随人类并观察其行为来推断其意图与限制条件(社会感知);(2)针对人类伙伴制定协作计划,以团队合作形式尽可能快速地完成任务(协作规划)。为建立该挑战的基准,我们创建了四个具有真实物理限制的智能体,以及八个包含室内外场景、多种限制条件、紧急事件与潜在风险的长期任务。我们在该挑战上对基于规划与学习的基线方法进行基准测试,并提出一种利用大语言模型与行为建模的新方法。实证评估表明,我们的基准能有效支持对机器社会智能关键方面的系统性评估。我们的基准与代码已通过此网址公开:https://github.com/UMass-Foundation-Model/CHAIC。