Robust Learning of Tactile Force Estimation through Robot Interaction

2019 International Conference on Robotics and Automation (ICRA)(2019)

引用 61|浏览234
暂无评分
摘要
Current methods for estimating force from tactile sensor signals are either inaccurate analytic models or task-specific learned models. In this paper, we explore learning a robust model that maps tactile sensor signals to force. We specifically explore learning a mapping for the SynTouch BioTac sensor via neural networks. We propose a voxelized input feature layer for spatial signals and leverage information about the sensor surface to regularize the loss function. To learn a robust tactile force model that transfers across tasks, we generate ground truth data from three different sources: (1) the BioTac rigidly mounted to a force torque~(FT) sensor, (2) a robot interacting with a ball rigidly attached to the same FT sensor, and (3) through force inference on a planar pushing task by formalizing the mechanics as a system of particles and optimizing over the object motion. A total of 140k samples were collected from the three sources. We achieve a median angular accuracy of 3.5 degrees in predicting force direction (66% improvement over the current state of the art) and a median magnitude accuracy of 0.06 N (93% improvement) on a test dataset. Additionally, we evaluate the learned force model in a force feedback grasp controller performing object lifting and gentle placement. Our results can be found on https://sites.google.com/view/tactile-force.
更多
查看译文
关键词
learned force model,robust learning,tactile force estimation,robot interaction,analytic models,robust model,SynTouch BioTac sensor,neural networks,voxelized input feature layer,spatial signals,sensor surface,robust tactile force model,force torque sensor,FT sensor,force inference,planar pushing task,force direction,force estimation,tactile sensor signals,force feedback grasp controller
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要