TY - GEN
T1 - Comparative study of visual saliency maps in the problem of classification of architectural images with Deep CNNs
AU - Obeso, Abraham Montoya
AU - Benois-Pineau, Jenny
AU - Guissous, Kamel
AU - Gouet-Brunet, Valerie
AU - Garcia Vazquez, Mireya S.
AU - Ramirez Acosta, Alejandro A.
N1 - Publisher Copyright:
© 2018 IEEE.
PY - 2019/1/10
Y1 - 2019/1/10
N2 - Incorporating Human Visual System (HVS) models into building of classifiers has become an intensively researched field in visual content mining. In the variety of models of HVS we are interested in so-called visual saliency maps. Contrarily to scan-paths they model instantaneous attention assigning the degree of interestingness/saliency for humans to each pixel in the image plane. In various tasks of visual content understanding, these maps proved to be efficient stressing contribution of the areas of interest in image plane to classifiers models. In previous works saliency layers have been introduced in Deep CNNs, showing that they allow reducing training time getting similar accuracy and loss values in optimal models. In case of large image collections efficient building of saliency maps is based on predictive models of visual attention. They are generally bottom-up and are not adapted to specific visual tasks. Unless they are built for specific content, such as »urban images»-targeted saliency maps we also compare in this paper. In present research we propose a »bootstrap» strategy of building visual saliency maps for particular tasks of visual data mining. A small collection of images relevant to the visual understanding problem is annotated with gaze fixations. Then the propagation to a large training dataset is ensured and compared with the classical GBVS model and a recent method of saliency for urban image content. The classification results within Deep CNN framework are promising compared to the purely automatic visual saliency prediction.
AB - Incorporating Human Visual System (HVS) models into building of classifiers has become an intensively researched field in visual content mining. In the variety of models of HVS we are interested in so-called visual saliency maps. Contrarily to scan-paths they model instantaneous attention assigning the degree of interestingness/saliency for humans to each pixel in the image plane. In various tasks of visual content understanding, these maps proved to be efficient stressing contribution of the areas of interest in image plane to classifiers models. In previous works saliency layers have been introduced in Deep CNNs, showing that they allow reducing training time getting similar accuracy and loss values in optimal models. In case of large image collections efficient building of saliency maps is based on predictive models of visual attention. They are generally bottom-up and are not adapted to specific visual tasks. Unless they are built for specific content, such as »urban images»-targeted saliency maps we also compare in this paper. In present research we propose a »bootstrap» strategy of building visual saliency maps for particular tasks of visual data mining. A small collection of images relevant to the visual understanding problem is annotated with gaze fixations. Then the propagation to a large training dataset is ensured and compared with the classical GBVS model and a recent method of saliency for urban image content. The classification results within Deep CNN framework are promising compared to the purely automatic visual saliency prediction.
KW - Deep Learning
KW - Mexican Culture
KW - Saliency Maps
UR - http://www.scopus.com/inward/record.url?scp=85061920656&partnerID=8YFLogxK
U2 - 10.1109/IPTA.2018.8608125
DO - 10.1109/IPTA.2018.8608125
M3 - Contribución a la conferencia
AN - SCOPUS:85061920656
T3 - 2018 8th International Conference on Image Processing Theory, Tools and Applications, IPTA 2018 - Proceedings
BT - 2018 8th International Conference on Image Processing Theory, Tools and Applications, IPTA 2018 - Proceedings
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 8th International Conference on Image Processing Theory, Tools and Applications, IPTA 2018
Y2 - 7 November 2018 through 10 November 2018
ER -