Peter Stone's Selected Publications

Classified by TopicClassified by Publication TypeSorted by DateSorted by First Author Last NameClassified by Funding Source


Bottom-Up Skill Discovery from Unsegmented Demonstrations for Long-Horizon Robot Manipulation

Bottom-Up Skill Discovery from Unsegmented Demonstrations for Long-Horizon Robot Manipulation.
Yifeng Zhu, Peter Stone, and Yuke Zhu.
IEEE Robotics and Automation Letters (RA-L), 7:4126–33, April 2022.
Project page
Code

Download

[PDF]9.3MB  

Abstract

We tackle real-world long-horizon robot manipulation tasks through skill discovery. We present a bottom-up approach to learning a library of reusable skills from unsegmented demonstrations and use these skills to synthesize prolonged robot behaviors. Our method starts with constructing a hierarchical task structure from each demonstration through agglomerative clustering. From the task structures of multi-task demonstrations, we identify skills based on the recurring patterns and train goal-conditioned sensorimotor policies with hierarchical imitation learning. Finally, we train a meta controller to compose these skills to solve long-horizon manipulation tasks. The entire model can be trained on a small set of human demonstrations collected within 30 minutes without further annotations, making it amendable to real-world deployment. We systematically evaluated our method in simulation environments and on a real robot. Our method has shown superior performance over state-of-the-art imitation learning methods in multi-stage manipulation tasks. Furthermore, skills discovered from multi-task demonstrations boost the average task success by 8 percents compared to those discovered from individual tasks.

BibTeX Entry

@article{ral2022-zhu,
  author={Yifeng Zhu and Peter Stone and Yuke Zhu},
  journal={IEEE Robotics and Automation Letters (RA-L)}, 
  title={Bottom-Up Skill Discovery from Unsegmented Demonstrations for Long-Horizon Robot Manipulation}, 
  year={2022},
  doi={10.1109/LRA.2022.3146589},
  month="April",
  volume="7",issue="2",
  pages="4126--33",
  abstract={We tackle real-world long-horizon robot manipulation tasks through skill discovery. We present a bottom-up approach to learning a library of reusable skills from unsegmented demonstrations and use these skills to synthesize prolonged robot behaviors. Our method starts with constructing a hierarchical task structure from each demonstration through agglomerative clustering. From the task structures of multi-task demonstrations, we identify skills based on the recurring patterns and train goal-conditioned sensorimotor policies with hierarchical imitation learning. Finally, we train a meta controller to compose these skills to solve long-horizon manipulation tasks. The entire model can be trained on a small set of human demonstrations collected within 30 minutes without further annotations, making it amendable to real-world deployment. We systematically evaluated our method in simulation environments and on a real robot. Our method has shown superior performance over state-of-the-art imitation learning methods in multi-stage manipulation tasks. Furthermore, skills discovered from multi-task demonstrations boost the average task success by 8 percents compared to those discovered from individual tasks.},
  wwwnote={<a href="https://ut-austin-rpl.github.io/rpl-BUDS/" target="_blank">Project page</a><br><a href="https://github.com/UT-Austin-RPL/BUDS">Code</a>}
}

Generated by bib2html.pl (written by Patrick Riley ) on Sun Nov 24, 2024 20:24:50