2017 4th IAPR Asian Conference on Pattern Recognition (ACPR) 2017
DOI: 10.1109/acpr.2017.136
|View full text |Cite
|
Sign up to set email alerts
|

Dynamic Vision Sensors for Human Activity Recognition

Abstract: Unlike conventional cameras which capture video at a fixed frame rate, Dynamic Vision Sensors (DVS) record only changes in pixel intensity values. The output of DVS is simply a stream of discrete ON/OFF events based on the polarity of change in its pixel values. DVS has many attractive features such as low power consumption, high temporal resolution, high dynamic range and less storage requirements. All these make DVS a very promising camera for potential applications in wearable platforms where power consumpt… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
17
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
5
2
1

Relationship

0
8

Authors

Journals

citations
Cited by 27 publications
(17 citation statements)
references
References 20 publications
0
17
0
Order By: Relevance
“…This work presents a VLSI architecture for FPGA to accelerate the HOTS algorithm. The system was tested with a gesture recognition dataset [42], obtaining an accuracy loss of 1.2% from the algorithm implemented in [36]. The estimation of the power consumption is 77 mW with Xilinx X-Power, working with a frequency of 100 MHz and implying a 50% toggle rate.…”
Section: Discussionmentioning
confidence: 99%
“…This work presents a VLSI architecture for FPGA to accelerate the HOTS algorithm. The system was tested with a gesture recognition dataset [42], obtaining an accuracy loss of 1.2% from the algorithm implemented in [36]. The estimation of the power consumption is 77 mW with Xilinx X-Power, working with a frequency of 100 MHz and implying a 50% toggle rate.…”
Section: Discussionmentioning
confidence: 99%
“…There is only a little work on exploring the neuromorphic data beyond object detection addressing highly semantic applications, such as, multi class action recognition, which still poses an important challenge. As mentioned in Section I, work on using NVS data for HAR is still in early stages [39]- [43]. Most of these methods start with temporally aggregating the polarities into a collection of NVS data frames by considering a non-overlapping time window corresponding to the frame rate of conventional APS cameras.…”
Section: Related Workmentioning
confidence: 99%
“…A time stamp aggregation algorithm is used to create the frames from the events, where these frames are fed into CNN for classification with a 92.90% of accuracy. In [39], three 2D motion maps (on x-y, x-z and y-z planes) and Motion Boundary Histogram (MBH) are constructed from the events. Speeded Up Robust Features (SURF) are extracted through grid search on the 2D motion maps followed by k-means clustering to create a Bag of visual vocabulary (BoVV) of k words from motion maps and MBH.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Consequently, k-fold CV could overestimate the performance of the classification by relying on correlations within subjects. Stefanie Anna Baby [5] proposed a model which includes a dynamic vision sensor, it processes only the foreground objects that being the humans captured the camera while completely ignoring the unnecessary processing of backgrounds thus decreasing processing requirements and greatly enhancing the detection efficiency of the given model. The difference in intensity of every pixel solely depends upon the texture edge and hence the output extracted from DVS is very sparse.…”
Section: Introductionmentioning
confidence: 99%