HOPMan exhibiting its skills across diverse tasks in unseen scenarios
We develop a framework for generalizable zero-shot manipulation (HOPMan) that can efficiently acquire a wide diversity of non-trivial skills and generalize them to diverse unseen scenarios.
Trained with large datasets of passive human videos, and small paired datasets of human-robot trajectories, HOPMan can exhibit a diverse set of 16 non-trivial manipulation skills (beyond picking/pushing, including articulated object manipulation and object re-orientation) across 100 tasks and can generalize them to diverse unseen scenarios (involving unseen objects, unseen tasks, and to completely unseen kitchens and offices).
@misc{bharadhwaj2023generalizable,
title={Towards Generalizable Zero-Shot Manipulation via Translating Human Interaction Plans},
author={Homanga Bharadhwaj and Abhinav Gupta and Vikash Kumar and Shubham Tulsiani},
year={2023},
eprint={2312.00775},
archivePrefix={arXiv},
primaryClass={cs.RO}
}