We present Premier-TACO, a multitask feature representation learning approach designed to improve few-shot policy learning efficiency in sequential decision-making tasks. Premier-TACO leverages a subset of multitask offline datasets for pretraining a general feature representation, which captures critical environmental dynamics and is fine-tuned using minimal expert demonstrations. It advances the temporal action contrastive learning (TACO) objective, known for state-of-the-art results in visual control tasks, by incorporating a novel negative example sampling strategy. This strategy is crucial in significantly boosting TACO's computational efficiency, making large-scale multitask offline pretraining feasible. Our extensive empirical evaluation in a diverse set of continuous control benchmarks including Deepmind Control Suite, MetaWorld, and LIBERO demonstrate Premier-TACO's effectiveness in pretraining visual representations, significantly enhancing few-shot imitation learning of novel tasks. Our code, pretraining data, as well as pretrained model checkpoints will be released at https://github.com/PremierTACO/premier-taco. Our project webpage is at https://premiertaco.github.io.
翻译:我们提出了Premier-TACO,一种旨在提升序列决策任务中小样本策略学习效率的多任务特征表征学习方法。Premier-TACO利用多任务离线数据集的一个子集来预训练一个通用的特征表征,该表征捕捉了关键的环境动态,并可使用极少量的专家示范进行微调。该方法通过引入一种新颖的负例采样策略,改进了在视觉控制任务中已知能取得最先进结果的时序动作对比学习(TACO)目标。该策略对于显著提升TACO的计算效率至关重要,使得大规模多任务离线预训练变得可行。我们在包括Deepmind Control Suite、MetaWorld和LIBERO在内的多样化连续控制基准测试中进行的广泛实证评估表明,Premier-TACO在预训练视觉表征方面具有高效性,能显著增强新任务的小样本模仿学习能力。我们的代码、预训练数据以及预训练模型检查点将在 https://github.com/PremierTACO/premier-taco 发布。我们的项目网页位于 https://premiertaco.github.io。