The Robustness of Deep Networks A geometrical perspective

被引:108
作者
Fawzi, Alhussein [1 ]
Moosavi-Dezfooli, Seyed-Mohsen [2 ,3 ]
Frossard, Pascal [4 ,5 ]
机构
[1] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90024 USA
[2] Ecole Polytech Fed Lausanne, Signal Proc Lab 4, Lausanne, Switzerland
[3] Ecole Polytech Fed Lausanne, Audiovisual Commun Lab, Lausanne, Switzerland
[4] IBM TJ Watson Res Ctr, Yorktown Hts, NY USA
[5] Ecole Polytech Fed Lausanne, Signal Proc Lab, Lausanne, Switzerland
关键词
D O I
10.1109/MSP.2017.2740965
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep neural networks have recently shown impressive classification performance on a diverse set of visual tasks. When deployed in real-world (noise-prone) environments, it is equally important that these classifiers satisfy robustness guarantees: small perturbations applied to the samples should not yield significant loss to the performance of the predictor. The goal of this article is to discuss the robustness of deep networks to a diverse set of perturbations that may affect the samples in practice, including adversarial perturbations, random noise, and geometric transformations. This article further discusses the recent works that build on the robustness analysis to provide geometric insights on the classifier's decision surface, which help in developing a better understanding of deep networks. Finally, we present recent solutions that attempt to increase the robustness of deep networks. We hope this review article will contribute to shed ding light on the open research challenges in the robustness of deep networks and stir interest in the analysis of their fundamental properties.
引用
收藏
页码:50 / 62
页数:13
相关论文
共 47 条
[1]  
Alemi A. A., 2016, ABS161200410 ARXIV
[2]  
Nguyen A, 2015, PROC CVPR IEEE, P427, DOI 10.1109/CVPR.2015.7298640
[3]  
[Anonymous], 2016, CORR 2016
[4]  
[Anonymous], ACM INT C MACH LEARN
[5]  
[Anonymous], P IEEE C COMP VIS PA
[6]  
[Anonymous], 2017, ARXIV170509552
[7]  
[Anonymous], MACHINE LEARNING
[8]  
[Anonymous], P ACM INT C MULT
[9]  
[Anonymous], P ADV NEURAL INFORM
[10]  
[Anonymous], 2017, Biologically inspired protection of deep networks from adversarial attacks