Forward-backward visual saliency propagation in Deep NNs vs internal attentional mechanisms

Abraham Montoya Obeso, Jenny Benois-Pineau, Mireya Sarai Garcia Vazquez, Alejandro Alvaro Ramirez Acosta

Producción científica: Capítulo del libro/informe/acta de congresoContribución a la conferenciarevisión exhaustiva

9 Citas (Scopus)

Resumen

Attention models in deep learning algorithms gained popularity in recent years. In this work, we propose an attention mechanism on the basis of visual saliency maps injected into the Deep Neural Network (DNN) to enhance regions in feature maps during forward-backward propagation in training, and only forward propagation in testing. The key idea is to spatially capture features associated to prominent regions in images and propagate them to deeper layers. During training, first, we take as backbone the well-known AlexNet architecture and then the ResNet architecture to solve the task of building identification of Mexican architecture. Our model equipped with the "external" visual saliency-based attention mechanism outperforms models armed with squeeze-and-excitation units and double-attention blocks.

Idioma originalInglés
Título de la publicación alojada2019 9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019
EditorialInstitute of Electrical and Electronics Engineers Inc.
ISBN (versión digital)9781728139753
DOI
EstadoPublicada - nov. 2019
Evento9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019 - Istanbul, Turquía
Duración: 6 nov. 20199 nov. 2019

Serie de la publicación

Nombre2019 9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019

Conferencia

Conferencia9th International Conference on Image Processing Theory, Tools and Applications, IPTA 2019
País/TerritorioTurquía
CiudadIstanbul
Período6/11/199/11/19

Huella

Profundice en los temas de investigación de 'Forward-backward visual saliency propagation in Deep NNs vs internal attentional mechanisms'. En conjunto forman una huella única.

Citar esto