Visualizing Deep Convolutional Neural Networks Using Natural Pre-images

被引:305
作者
Mahendran, Aravindh [1 ]
Vedaldi, Andrea [1 ]
机构
[1] Univ Oxford, Oxford, England
关键词
Visualization; Convolutional neural networks; Pre-image problem;
D O I
10.1007/s11263-016-0911-8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image representations, from SIFT and bag of visual words to convolutional neural networks (CNNs) are a crucial component of almost all computer vision systems. However, our understanding of them remains limited. In this paper we study several landmark representations, both shallow and deep, by a number of complementary visualization techniques. These visualizations are based on the concept of "natural pre-image", namely a natural-looking image whose representation has some notable property. We study in particular three such visualizations: inversion, in which the aim is to reconstruct an image from its representation, activation maximization, in which we search for patterns that maximally stimulate a representation component, and caricaturization, in which the visual patterns that a representation detects in an image are exaggerated. We pose these as a regularized energy-minimization framework and demonstrate its generality and effectiveness. In particular, we show that this method can invert representations such as HOG more accurately than recent alternatives while being applicable to CNNs too. Among our findings, we show that several layers in CNNs retain photographically accurate information about the image, with different degrees of geometric and photometric invariance.
引用
收藏
页码:233 / 255
页数:23
相关论文
共 55 条
[1]  
[Anonymous], 2011, CVPR
[2]  
[Anonymous], 2014, MATCONVNET CNNS MATL
[3]  
[Anonymous], 2014, P BMVC
[4]  
[Anonymous], 2012, ARXIV12125701 CORR
[5]  
[Anonymous], IEEE T NEURAL NETWOR
[6]  
[Anonymous], CVPR
[7]  
[Anonymous], P ICML WORKSH
[8]  
[Anonymous], P COGSCI
[9]  
[Anonymous], ECCV
[10]  
[Anonymous], 2009, TECHNICAL REPORT