An egocentric perspective on active vision and visual object learning in toddlers.

Joint IEEE International Conference on Development and Learning and Epigenetic Robotics ICDL-EpiRob(2017)

引用 7|浏览10
暂无评分
摘要
Toddlers quickly learn to recognize thousands of everyday objects despite the seemingly suboptimal training conditions of a visually cluttered world. One reason for this success may be that toddlers do not just passively perceive visual information, but actively explore and manipulate objects around them. The work in this paper is based on the idea that active viewing and exploration creates "clean" egocentric scenes that serve as high-quality training data for the visual system. We tested this idea by collecting first-person video data of free toy play between toddler-parent pairs. We use the raw frames from this data, weakly annotated with toy object labels, to train state-of-the-art machine learning models for object recognition (Convolutional Neural Networks, or CNNs). We run several training simulations, varying quantity and quality of the training data. Our results show that scenes captured by parents and toddlers have different properties, and that toddler scenes lead to models that learn more robust visual representations of the toy objects in them.
更多
查看译文
关键词
visual object learning,toddlers,visually cluttered world,visual information,active viewing,clean egocentric scenes,high-quality training data,visual system,first-person video data,free toy play,toddler-parent pairs,toy object labels,object recognition,training simulations,toddler scenes,robust visual representations,toy objects,egocentric perspective,active vision,suboptimal training conditions,convolutional neural networks,CNN
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要