2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition 2018
DOI: 10.1109/cvpr.2018.00625
|View full text |Cite
|
Sign up to set email alerts
|

Egocentric Activity Recognition on a Budget

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
25
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 38 publications
(25 citation statements)
references
References 31 publications
0
25
0
Order By: Relevance
“…The wide use of CNNs in third-person vision was followed by their extensive application in egocentric action and activity recognition [16], [21], [38], [40], [41], [56]. Earlier approaches handled CNN features as an additional modality to handcrafted features [49] or as a feature combination mechanism on previously extracted egocentric features [16].…”
Section: Video Activity Recognitionmentioning
confidence: 99%
See 1 more Smart Citation
“…The wide use of CNNs in third-person vision was followed by their extensive application in egocentric action and activity recognition [16], [21], [38], [40], [41], [56]. Earlier approaches handled CNN features as an additional modality to handcrafted features [49] or as a feature combination mechanism on previously extracted egocentric features [16].…”
Section: Video Activity Recognitionmentioning
confidence: 99%
“…In [61], [62] optical flow was employed to detect salient regions, which were cropped from the original RGB frames and were given to the network as a second, more focused RGB stream. Other input modalities have been employed including depth [7], [41], egocentric cues comprising hand [63], [64], [65] and object regions [64], [66], [67], head motions [63] and gaze-based saliency maps [63], [65], sensor-based modalities [15], [56], [59] and sound [43], [68], [69]. In [38], [40] object and hand localization and segmentation were intermediate learning steps that forced the network to focus on important egocentric cues prior to action prediction.…”
Section: Video Activity Recognitionmentioning
confidence: 99%
“…Efficient video activity recognition designed for mobile devices has been studied by several research groups. An energy aware training algorithm was proposed in Possas et al (2018), to demonstrate energy efficient video activity recognition on complex problems. In this work, the authors use reinforcement learning to train a network on both video and motion information captured by sensors while penalizing actions that have high energy costs.…”
Section: Related Workmentioning
confidence: 99%
“…The goal of egocentric vision is to analyze the visual information provided by wearable cameras, which have the capability to acquire images from a first person point-of-view. The analysis of these images provides information about the behavior of the user, useful for several complementary topics like social interactions (Aghaei et al, 2018), scene understanding (Singh et al, 2016), time-space-based localization (Yao et al, 2018), action (Fathi et al, 2011;Possas et al, 2018) or activity recognition (Iwashita et al, 2014;Cartas et al, 2017), or nutritional habits analysis (Bolaños et al, 2018b), among others. Thus, enabling us to understand the whole story and behavior of the users behind the pictures (i.e.…”
Section: Captioning Visual Contentmentioning
confidence: 99%