2021
DOI: 10.1109/access.2021.3115701
|View full text |Cite
|
Sign up to set email alerts
|

Automatic CNN-Based Enhancement of 360° Video Experience With Multisensorial Effects

Abstract: High-resolution audio-visual virtual reality (VR) technologies currently offer satisfying experiences for both sight and hearing senses in the world of multimedia. However, the delivery of truly immersive experiences requires the incorporation of other senses such as touch and smell. Multisensorial effects are usually manually synchronized with videos and data is stored in companion files, which contain timestamps for these effects. This manual task becomes very complex for 360°videos, as the scenes triggering… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
10
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
4
2

Relationship

1
5

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 36 publications
0
10
0
Order By: Relevance
“…The results show a statistically significant benefit for the presence of odor and wind in the QoE. Sexton et al [11] developed an algorithm to automatically add multisensorial information to 360°videos, combining hapics and olfaction. A playback system was designed to improve on the works of Comsa et al [9] and Bi et al [28].…”
Section: A Mulsemediamentioning
confidence: 99%
See 4 more Smart Citations
“…The results show a statistically significant benefit for the presence of odor and wind in the QoE. Sexton et al [11] developed an algorithm to automatically add multisensorial information to 360°videos, combining hapics and olfaction. A playback system was designed to improve on the works of Comsa et al [9] and Bi et al [28].…”
Section: A Mulsemediamentioning
confidence: 99%
“…CNNs can accurately perform action and scene recognition for the automatic generation of multisensory effects from multimedia inputs, as manually adding these effects is a lengthy process. The work described in [11] proposed an initial CNN-based solution for multisensory systems, but left several questions to be answered: how action detection can improve the effects dispensed, what other metrics can be analyzed to indicate the feasibility of the solution (e.g., GFLOPs, FPR, FNR), and how the solution can be generalized to work with a large number of videos, categories, scents and datasets. Several CNNs were also described in this section and a thorough evaluation needs to be performed for the identification of a suitable network to work with the proposed solution and help achieve best results in terms of performance, complexity and accuracy.…”
Section: B Convolutional Neural Network For Video Processingmentioning
confidence: 99%
See 3 more Smart Citations