Learning to See by Moving

被引:250
作者
Agrawal, Pulkit [1 ]
Carreira, Joao [1 ]
Malik, Jitendra [1 ]
机构
[1] Univ Calif Berkeley, Berkeley, CA 94720 USA
来源
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV) | 2015年
关键词
CODE;
D O I
10.1109/ICCV.2015.13
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The current dominant paradigm for feature learning in computer vision relies on training neural networks for the task of object recognition using millions of hand labelled images. Is it also possible to learn features for a diverse set of visual tasks using any other form of supervision? In biology, living organisms developed the ability of visual perception for the purpose of moving and acting in the world. Drawing inspiration from this observation, in this work we investigated if the awareness of egomotion (i.e. self motion) can be used as a supervisory signal for feature learning. As opposed to the knowledge of class labels, information about egomotion is freely available to mobile agents. We found that using the same number of training images, features learnt using egomotion as supervision compare favourably to features learnt using class-label as supervision on the tasks of scene recognition, object recognition, visual odometry and keypoint matching.
引用
收藏
页码:37 / 45
页数:9
相关论文
共 37 条
[1]  
Agrawal P, 2014, LECT NOTES COMPUT SC, V8695, P329, DOI 10.1007/978-3-319-10584-0_22
[2]  
[Anonymous], 1989, P ADV NEUR INF PROC
[3]  
[Anonymous], 2014, ADV NEURAL INFORM PR
[4]  
[Anonymous], 2014, CoRR
[5]  
[Anonymous], 2014, ARXIV14117676
[6]  
[Anonymous], 2012, CORR
[7]  
[Anonymous], 2013, 31 INT C MACH LEARN
[8]  
[Anonymous], PERCEPTION EYE MOTIO
[9]  
[Anonymous], 2009, Deep boltzmann machines
[10]  
[Anonymous], 2015, ARXIV150402518