2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2022
DOI: 10.1109/cvpr52688.2022.01931
|View full text |Cite
|
Sign up to set email alerts
|

E2(GO)MOTION: Motion Augmented Event Stream for Egocentric Action Recognition

Abstract: Event cameras are novel bio-inspired sensors, which asynchronously capture pixel-level intensity changes in the form of "events". Due to their sensing mechanism, event cameras have little to no motion blur, a very high temporal resolution and require significantly less power and memory than traditional frame-based cameras. These characteristics make them a perfect fit to several real-world applications such as egocentric action recognition on wearable devices, where fast camera motion and limited power challen… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
10
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
4
3
3

Relationship

1
9

Authors

Journals

citations
Cited by 33 publications
(11 citation statements)
references
References 74 publications
1
10
0
Order By: Relevance
“…Similarly, the transferability of Kinetics to other popular action recognition datasets has been studied directly [18] or indirectly to benchmark architectures [47,13]. Kinetics pretraining has also been applied to other action recognition settings, such as egocentric actions [33], action recognition from drones [9] or actions in the dark [52]. Other studies have used Kinetics to initialize models for more distant video tasks, including sign language recognition [25] or autonomous vehicle decision-making [41].…”
Section: Transfer Learning From Imagenet and Kineticsmentioning
confidence: 99%
“…Similarly, the transferability of Kinetics to other popular action recognition datasets has been studied directly [18] or indirectly to benchmark architectures [47,13]. Kinetics pretraining has also been applied to other action recognition settings, such as egocentric actions [33], action recognition from drones [9] or actions in the dark [52]. Other studies have used Kinetics to initialize models for more distant video tasks, including sign language recognition [25] or autonomous vehicle decision-making [41].…”
Section: Transfer Learning From Imagenet and Kineticsmentioning
confidence: 99%
“…Another important aspect to consider when posing real-time constrains is that, in the context of egocentric vision, many techniques attain notable results only by leveraging non-real-time secondary modalities such as the optical flow. Although this modality is highly successful, it has a high computational cost [36], [37], which prevents its use in real-time applications, and increases the size of the model.…”
Section: Bringing Fpar In the Wildmentioning
confidence: 99%
“…Despite the abundance of conventional frame-like datasets, there is a noticeable scarcity of event-based action recognition datasets. As for the simulated datasets, N-EPIC-Kitchens [75] is an event version of the EPIC-Kitchens generated by the event camera simulator. The event UCF-50 [76] is derived from the UCF-50 action recognition dataset, which was captured by displaying its data on a monitor.…”
Section: Datasets For Action Recognitionmentioning
confidence: 99%