Improving Behavioural Cloning with Positive Unlabeled Learning.

Qiang Wang, Robert McCarthy,David Cordova Bulens,Kevin McGuinness,Noel E. O'Connor,Francisco Roldan Sanchez, Nico Gürtler, Felix Widmaier, Stephen J. Redmond

arXiv (Cornell University)(2023)

引用 0|浏览3
暂无评分
摘要
Learning control policies offline from pre-recorded datasets is a promising avenue for solving challenging real-world problems. However, available datasets are typically of mixed quality, with a limited number of the trajectories that we would consider as positive examples; i.e., high-quality demonstrations. Therefore, we propose a novel iterative learning algorithm for identifying expert trajectories in unlabeled mixed-quality robotics datasets given a minimal set of positive examples, surpassing existing algorithms in terms of accuracy. We show that applying behavioral cloning to the resulting filtered dataset outperforms several competitive offline reinforcement learning and imitation learning baselines. We perform experiments on a range of simulated locomotion tasks and on two challenging manipulation tasks on a real robotic system; in these experiments, our method showcases state-of-the-art performance. Our website: \url{https://sites.google.com/view/offline-policy-learning-pubc}.
更多
查看译文
关键词
behavioural cloning,positive unlabeled learning
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要