Comparative study of visual saliency maps in the problem of classification of architectural images with Deep CNNs

Abraham Montoya Obeso, Jenny Benois-Pineau, Kamel Guissous, Valerie Gouet-Brunet, Mireya S. Garcia Vazquez, Alejandro A. Ramirez Acosta

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

8 Scopus citations

Abstract

Incorporating Human Visual System (HVS) models into building of classifiers has become an intensively researched field in visual content mining. In the variety of models of HVS we are interested in so-called visual saliency maps. Contrarily to scan-paths they model instantaneous attention assigning the degree of interestingness/saliency for humans to each pixel in the image plane. In various tasks of visual content understanding, these maps proved to be efficient stressing contribution of the areas of interest in image plane to classifiers models. In previous works saliency layers have been introduced in Deep CNNs, showing that they allow reducing training time getting similar accuracy and loss values in optimal models. In case of large image collections efficient building of saliency maps is based on predictive models of visual attention. They are generally bottom-up and are not adapted to specific visual tasks. Unless they are built for specific content, such as »urban images»-targeted saliency maps we also compare in this paper. In present research we propose a »bootstrap» strategy of building visual saliency maps for particular tasks of visual data mining. A small collection of images relevant to the visual understanding problem is annotated with gaze fixations. Then the propagation to a large training dataset is ensured and compared with the classical GBVS model and a recent method of saliency for urban image content. The classification results within Deep CNN framework are promising compared to the purely automatic visual saliency prediction.

Original languageEnglish
Title of host publication2018 8th International Conference on Image Processing Theory, Tools and Applications, IPTA 2018 - Proceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781538664278
DOIs
StatePublished - 10 Jan 2019
Event8th International Conference on Image Processing Theory, Tools and Applications, IPTA 2018 - Xi'an, China
Duration: 7 Nov 201810 Nov 2018

Publication series

Name2018 8th International Conference on Image Processing Theory, Tools and Applications, IPTA 2018 - Proceedings

Conference

Conference8th International Conference on Image Processing Theory, Tools and Applications, IPTA 2018
Country/TerritoryChina
CityXi'an
Period7/11/1810/11/18

Keywords

  • Deep Learning
  • Mexican Culture
  • Saliency Maps

Fingerprint

Dive into the research topics of 'Comparative study of visual saliency maps in the problem of classification of architectural images with Deep CNNs'. Together they form a unique fingerprint.

Cite this