2019
DOI: 10.3390/electronics8121538
|View full text |Cite
|
Sign up to set email alerts
|

A Feature Integrated Saliency Estimation Model for Omnidirectional Immersive Images

Abstract: Omnidirectional, or 360°, cameras are able to capture the surrounding space, thus providing an immersive experience when the acquired data is viewed using head mounted displays. Such an immersive experience inherently generates an illusion of being in a virtual environment. The popularity of 360 • media has been growing in recent years. However, due to the large amount of data, processing and transmission pose several challenges. To this aim, efforts are being devoted to the identification of regions that can … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3

Relationship

0
7

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 38 publications
0
3
0
Order By: Relevance
“…For both VGG-TL and VGG-MLNet, our framework first need to create various overlapping 2D projections of scene using spherical coordinate at fixed image resolutions (i.e. 512×512 pixels and 480×640 pixels for VGG-TL and VGG-MLNet models respectively) with θ F oV = 120°at θ viewport−center = [0, 15 , ..., 330, 345] and φ viewport−center = [-90, -75, ..., 0, ..., 75, 90] similar to various works on 360°data processing [13][14][15]31]. Then, each 2D image patch from the scene is used in VGG-TL or VGG-MLNet to obtain saliency maps of each extracted view-ports.…”
Section: Saliency Prediction Modelmentioning
confidence: 75%
“…For both VGG-TL and VGG-MLNet, our framework first need to create various overlapping 2D projections of scene using spherical coordinate at fixed image resolutions (i.e. 512×512 pixels and 480×640 pixels for VGG-TL and VGG-MLNet models respectively) with θ F oV = 120°at θ viewport−center = [0, 15 , ..., 330, 345] and φ viewport−center = [-90, -75, ..., 0, ..., 75, 90] similar to various works on 360°data processing [13][14][15]31]. Then, each 2D image patch from the scene is used in VGG-TL or VGG-MLNet to obtain saliency maps of each extracted view-ports.…”
Section: Saliency Prediction Modelmentioning
confidence: 75%
“…where the normalized image or one of the five saliency map models (Itti [28], GBVS [70], Geometry [71], BMS [72], and EBMS [73]) are computed. The image pairs are obtained as follows:…”
Section: A Analysis Of the Relationship Among Distortion Saliency And...mentioning
confidence: 99%
“…at is, the block compression measurement matrix could be used as the forward discrete cosine transform (FDCT) matrix in the JPEG coding, and the inverse discrete cosine transform (IDCT) process is replaced by sparse reconstruction. In addition, multiple feature saliency and noise analysis are introduced to implement adaptive control of the observation matrix and minimal error iterative reconstruction [8,9].…”
Section: Introductionmentioning
confidence: 99%