2019
DOI: 10.1016/j.eswa.2019.03.017
|View full text |Cite
|
Sign up to set email alerts
|

Hybrid feature-based analysis of video’s affective content using protagonist detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
19
0

Year Published

2019
2019
2025
2025

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 18 publications
(19 citation statements)
references
References 37 publications
0
19
0
Order By: Relevance
“…However, obtaining such dynamic labels by subjective reports could be labor-intensive and time consuming [74,84] and automatic tagging using internet-based crowdsourcing methods [85][86][87] or information from other modalities (e.g. video content analysis [88], face expressions [89] and peripheral physiological responses [90] during video watching) could provide feasible alternative options. For another, the lacking of standardized large-scale datasets has made the performance evaluation across different methods difficult.…”
Section: Decoding Affective State Accurately and Continuouslymentioning
confidence: 99%
“…However, obtaining such dynamic labels by subjective reports could be labor-intensive and time consuming [74,84] and automatic tagging using internet-based crowdsourcing methods [85][86][87] or information from other modalities (e.g. video content analysis [88], face expressions [89] and peripheral physiological responses [90] during video watching) could provide feasible alternative options. For another, the lacking of standardized large-scale datasets has made the performance evaluation across different methods difficult.…”
Section: Decoding Affective State Accurately and Continuouslymentioning
confidence: 99%
“…In many cases, computations are conducted over each frame of the video, and the average values of the computational results of the overall video are considered as visual features. Specifically, the color-related features often contain the histogram and variance of color [20,107,177], the proportions of color [82,86], the number of white frame and fades [177], the grayness [20], darkness ratio, color energy [132,170], brightness ratio and saturation [85,86], etc. In addition, the differences of dark and light can be reflected by the lighting key, which is used to evoke emotions in video and draw the attention of viewers by creating an emotional atmosphere [82].…”
Section: Content-related Featuresmentioning
confidence: 99%
“…ZCR [86] is used to separate different types of audio signals, such as music, environmental sound and speech of human. Besides these frequent related features, audio flatness [177], spectral flux [177], delta spectrum magnitude, harmony [86,111,177], band energy ratio, spectral centroid [49,177], and spectral contrast [86] are also utilized.…”
Section: Content-related Featuresmentioning
confidence: 99%
See 2 more Smart Citations