Forward-backward visual saliency propagation in Deep NNs vs internal attentional mechanisms

Abraham Montoya Obeso, Jenny Benois-Pineau, Mireya Sarai Garcia Vazquez, Alejandro Alvaro Ramirez Acosta

Research output: Chapter in Book/Report/Conference proceedingConference contributionpeer-review

9 Scopus citations

Abstract

Attention models in deep learning algorithms gained popularity in recent years. In this work, we propose an attention mechanism on the basis of visual saliency maps injected into the Deep Neural Network (DNN) to enhance regions in feature maps during forward-backward propagation in training, and only forward propagation in testing. The key idea is to spatially capture features associated to prominent regions in images and propagate them to deeper layers. During training, first, we take as backbone the well-known AlexNet architecture and then the ResNet architecture to solve the task of building identification of Mexican architecture. Our model equipped with the "external" visual saliency-based attention mechanism outperforms models armed with squeeze-and-excitation units and double-attention blocks.

Original languageEnglish
Title of host publication2019 9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019
PublisherInstitute of Electrical and Electronics Engineers Inc.
ISBN (Electronic)9781728139753
DOIs
StatePublished - Nov 2019
Event9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019 - Istanbul, Turkey
Duration: 6 Nov 20199 Nov 2019

Publication series

Name2019 9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019

Conference

Conference9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019
Country/TerritoryTurkey
CityIstanbul
Period6/11/199/11/19

Keywords

  • Neural Networks
  • Saliency Maps
  • Visual Attention

Fingerprint

Dive into the research topics of 'Forward-backward visual saliency propagation in Deep NNs vs internal attentional mechanisms'. Together they form a unique fingerprint.

Cite this