Deep Neural Networks for No-Reference and Full-Reference Image Quality Assessment

被引:842
作者
Bosse, Sebastian [1 ]
Maniry, Dominique [1 ]
Mueller, Klaus-Robert [2 ,3 ,4 ]
Wiegand, Thomas [5 ,6 ]
Samek, Wojciech [1 ]
机构
[1] Fraunhofer Heinrich Hertz Inst, Dept Video Coding & Analyt, D-10587 Berlin, Germany
[2] Berlin Inst Technol, Machine Learning Lab, D-10587 Berlin, Germany
[3] Korea Univ, Dept Brain & Cognit Engn, Seoul 136713, South Korea
[4] Max Planck Inst Informat, D-66123 Saarbrucken, Germany
[5] Fraunhofer Heinrich Hertz Inst, D-10587 Berlin, Germany
[6] Berlin Inst Technol, Media Technol Lab, D-10587 Berlin, Germany
基金
新加坡国家研究基金会;
关键词
Full-reference image quality assessment; no-reference image quality assessment; neural networks; quality pooling; deep learning; feature extraction; regression; PERCEPTUAL IMAGE; SIMILARITY; INDEX;
D O I
10.1109/TIP.2017.2760518
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present a deep neural network-based approach to image quality assessment (IQA). The network is trained end-to-end and comprises ten convolutional layers and five pooling layers for feature extraction, and two fully connected layers for regression, which makes it significantly deeper than related IQA models. Unique features of the proposed architecture are that: 1) with slight adaptations it can be used in a no-reference (NR) as well as in a full-reference (FR) IQA setting and 2) it allows for joint learning of local quality and local weights, i.e., relative importance of local quality to the global quality estimate, in an unified framework. Our approach is purely data-driven and does not rely on hand-crafted features or other types of prior domain knowledge about the human visual system or image statistics. We evaluate the proposed approach on the LIVE, CISQ, and TID2013 databases as well as the LIVE In the wild image quality challenge database and show superior performance to state-of-the-art NR and FR IQA methods. Finally, cross-database evaluation shows a high ability to generalize between different databases, indicating a high robustness of the learned features.
引用
收藏
页码:206 / 219
页数:14
相关论文
共 52 条
[11]   Automatic Prediction of Perceptual Image and Video Quality [J].
Bovik, Alan Conrad .
PROCEEDINGS OF THE IEEE, 2013, 101 (09) :2008-2024
[12]  
Bromley J., 1993, International Journal of Pattern Recognition and Artificial Intelligence, V7, P669, DOI 10.1142/S0218001493000339
[13]  
Chopra S., 2005, IEEE C COMPUTER VISI, P349, DOI DOI 10.1109/CVPR.2005.202
[14]   Psychophysiology-Based QoE Assessment: A Survey [J].
Engelke, Ulrich ;
Darcy, Daniel P. ;
Mulliken, Grant H. ;
Bosse, Sebastian ;
Martini, Maria G. ;
Arndt, Sebastian ;
Antons, Jan-Niklas ;
Chan, Kit Yan ;
Ramzan, Naeem ;
Brunnstrom, Kjell .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2017, 11 (01) :6-21
[15]   DeepSim: Deep similarity for image quality assessment [J].
Gao, Fei ;
Wang, Yi ;
Li, Panpeng ;
Tan, Min ;
Yu, Jun ;
Zhu, Yani .
NEUROCOMPUTING, 2017, 257 :104-114
[16]   Massive Online Crowdsourced Study of Subjective and Objective Picture Quality [J].
Ghadiyaram, Deepti ;
Bovik, Alan C. .
IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (01) :372-387
[17]  
Girod Bernd, 1993, P207
[18]   Fast R-CNN [J].
Girshick, Ross .
2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, :1440-1448
[19]   Convolutional Neural Networks for No-Reference Image Quality Assessment [J].
Kang, Le ;
Ye, Peng ;
Li, Yi ;
Doermann, David .
2014 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2014, :1733-1740
[20]   Fully Deep Blind Image Quality Predictor [J].
Kim, Jongyoo ;
Lee, Sanghoon .
IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2017, 11 (01) :206-220