Article (Scientific journals)
DeepRare: Generic Unsupervised Visual Attention Models
Kong, Phutphalla; Mancas, Matei; Gosselin, Bernard et al.
2022In Electronics, 11 (11), p. 1696
Peer reviewed
 

Files


Full Text
electronics-11-01696-with-cover.pdf
Author postprint (5.35 MB) Creative Commons License - Attribution
Download

All documents in ORBi UMONS are protected by a user license.

Send to



Details



Keywords :
deep features; eye tracking; odd one out; rarity; saliency; visibility; visual attention prediction; Control and Systems Engineering; Signal Processing; Hardware and Architecture; Computer Networks and Communications; Electrical and Electronic Engineering
Abstract :
[en] Visual attention selects data considered as “interesting” by humans, and it is modeled in the field of engineering by feature-engineered methods finding contrasted/surprising/unusual image data. Deep learning drastically improved the models efficiency on the main benchmark datasets. However, Deep Neural Networks-based (DNN-based) models are counterintuitive: surprising or unusual data are by definition difficult to learn because of their low occurrence probability. In reality, DNN-based models mainly learn top-down features such as faces, text, people, or animals which usually attract human attention, but they have low efficiency in extracting surprising or unusual data in the images. In this article, we propose a new family of visual attention models called DeepRare and especially DeepRare2021 (DR21), which uses the power of DNNs’ feature extraction and the genericity of feature-engineered algorithms. This algorithm is an evolution of a previous version called DeepRare2019 (DR19) based on this common framework. DR21 (1) does not need any additional training other than the default ImageNet training, (2) is fast even on CPU, (3) is tested on four very different eye-tracking datasets showing that DR21 is generic and is always within the top models on all datasets and metrics while no other model exhibits such a regularity and genericity. Finally, DR21 (4) is tested with several network architectures such as VGG16 (V16), VGG19 (V19), and MobileNetV2 (MN2), and (5) it provides explanation and transparency on which parts of the image are the most surprising at different levels despite the use of a DNN-based feature extractor.
Disciplines :
Computer science
Author, co-author :
Kong, Phutphalla ;  Institute of Technology of Cambodia (ITC), Russian Conf. Blvd, Phnom Penh, Cambodia ; Numediart Institute, University of Mons (UMONS), Mons, Belgium
Mancas, Matei  ;  Université de Mons - UMONS
Gosselin, Bernard  ;  Université de Mons - UMONS
Po, Kimtho;  Institute of Technology of Cambodia (ITC), Russian Conf. Blvd, Phnom Penh, Cambodia
Language :
English
Title :
DeepRare: Generic Unsupervised Visual Attention Models
Publication date :
June 2022
Journal title :
Electronics
ISSN :
2079-9292
eISSN :
2079-9292
Publisher :
MDPI
Volume :
11
Issue :
11
Pages :
1696
Peer reviewed :
Peer reviewed
Research institute :
R450 - Institut NUMEDIART pour les Technologies des Arts Numériques
Funding text :
Funding: This research was funded by ARES-CCD (program AI 2014-2019) by Belgian university cooperation.
Available on ORBi UMONS :
since 23 January 2023

Statistics


Number of views
10 (6 by UMONS)
Number of downloads
13 (1 by UMONS)

Scopus citations®
 
0
Scopus citations®
without self-citations
0
OpenCitations
 
0
OpenAlex citations
 
3

Bibliography


Similar publications



Contact ORBi UMONS