2016
DOI: 10.1007/978-3-319-46454-1_28
|View full text |Cite
|
Sign up to set email alerts
|

Detecting Engagement in Egocentric Video

Abstract: In a wearable camera video, we see what the camera wearer sees. While this makes it easy to know roughly what he chose to look at, it does not immediately reveal when he was engaged with the environment. Specifically, at what moments did his focus linger, as he paused to gather more information about something he saw? Knowing this answer would benefit various applications in video summarization and augmented reality, yet prior work focuses solely on the "what" question (estimating saliency, gaze) without consi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
37
0

Year Published

2016
2016
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 45 publications
(37 citation statements)
references
References 38 publications
0
37
0
Order By: Relevance
“…Aghaei et al [1] released the EgoSocialStyle dataset of 130k images from eight users, for the purpose of characterising the social patterns of a person relying exclusively on visual data. UT EE [37] gathered and crowd source annotated 27 videos from nine users, covering 14 hours in order to detecting the wearer's engagement with the surrounding environment. An event segmentation dataset was published by Gupta and Gurrin [12], consisting of 14k images from ten participants.…”
Section: Personal Data Test Collectionsmentioning
confidence: 99%
“…Aghaei et al [1] released the EgoSocialStyle dataset of 130k images from eight users, for the purpose of characterising the social patterns of a person relying exclusively on visual data. UT EE [37] gathered and crowd source annotated 27 videos from nine users, covering 14 hours in order to detecting the wearer's engagement with the surrounding environment. An event segmentation dataset was published by Gupta and Gurrin [12], consisting of 14k images from ten participants.…”
Section: Personal Data Test Collectionsmentioning
confidence: 99%
“…This method [30] used low temporal resolution image sequences to detect the social interactions. Su et al [50] proposed a video summarization approach to detect the engagement using long-term ego-motion cues (i.e., gaze). This approach [50] consists of three stages that are frame prediction, interval prediction, and classification with the trained model.…”
Section: Performance Comparison Of the Proposed Framework Withmentioning
confidence: 99%
“…Analysis of images captured by egocentric cameras can reveal a lot about the person recording such images, including, intentions, personality, interests, etc. Firstperson gaze prediction is useful in a wide range of applications in health care, education and entertainment, for tasks such as action and event recognition [35], recognition of handled objects [37], discovering important people [16], video re-editing [26], video summarization [45], engagement detection [42], and assistive vision systems [18].…”
Section: Introductionmentioning
confidence: 99%