2020
DOI: 10.1016/j.imavis.2020.103887
|View full text |Cite
|
Sign up to set email alerts
|

EML-NET: An Expandable Multi-Layer NETwork for saliency prediction

Abstract: Saliency prediction can benefit from training that involves scene understanding that may be tangential to the central task; this may include understanding places, spatial layout, objects or involve different datasets and their bias. One can combine models, but to do this in a sophisticated manner can be complex, and also result in unwieldy networks or produce competing objectives that are hard to balance. In this paper, we propose a scalable system to leverage multiple powerful deep CNN models to better extrac… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
112
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 131 publications
(112 citation statements)
references
References 51 publications
(121 reference statements)
0
112
0
Order By: Relevance
“…The applications of CNNs in salience detection are extensive. Some of the examples of these applications are Deep Gaze 2 [141], which presents two models for fixation predictions and object recognition; EML-NET [142] for salience feature detection using a modular approach; and DeepFix [143], which is sensitive to semantic information at different scales while using large receptive fields for global context analysis. The scope of existing models in the field is so broad that it is impossible to cover all of them in this review, but their accuracy in predicting areas of spatial salience is without question [144][145][146][147][148][149].…”
Section: Deep Learning Classifiersmentioning
confidence: 99%
“…The applications of CNNs in salience detection are extensive. Some of the examples of these applications are Deep Gaze 2 [141], which presents two models for fixation predictions and object recognition; EML-NET [142] for salience feature detection using a modular approach; and DeepFix [143], which is sensitive to semantic information at different scales while using large receptive fields for global context analysis. The scope of existing models in the field is so broad that it is impossible to cover all of them in this review, but their accuracy in predicting areas of spatial salience is without question [144][145][146][147][148][149].…”
Section: Deep Learning Classifiersmentioning
confidence: 99%
“…Compared to visual attention models for 360 • images, those for 2D traditional images have been well developed in recent years [21], [22], [23]. Seminal methods were proposed based on low-level or high-level semantic feature extraction from handcrafted filters [24], [25], [26] or Deep Convolutional Neural Networks (DCNN) [27], [28], [29], [30], [31] thanks to the establishment of several large scale datasets [32], [33], [34]. Unfortunately, these models are not immediately usable for 360 • images because of the severe geometric distortion on the top and bottom areas in equirectangular projection.…”
Section: By Leveraging 360 • Fixation Cues and User's Orientation Sensedmentioning
confidence: 99%
“…1. Static saliency models: eDN [72], DeepGaze I & II [49], Mr-CNN [55], SALICON [33], DeepFix [48], SAM-ResNet [19], and EML-Net [39].…”
Section: Bottom-up Attention Modeling: Deep Learning Eramentioning
confidence: 99%