2021
DOI: 10.1016/j.sysarc.2021.102041
|View full text |Cite
|
Sign up to set email alerts
|

A survey of accelerator architectures for 3D convolution neural networks

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
13
0
2

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
2
2

Relationship

0
9

Authors

Journals

citations
Cited by 38 publications
(19 citation statements)
references
References 31 publications
0
13
0
2
Order By: Relevance
“…One disadvantage is that this method fails to utilise the spatial distribution of Doppler velocities. Alternatively, 3D convolution can be used to extract features from all three dimensions in a 3D tensor, but requires huge computation and memory overheads [161]. RODNet [47] samples chirp sequences, as described above, to reduce input dimensionality.…”
Section: Pre-cfar Detectormentioning
confidence: 99%
“…One disadvantage is that this method fails to utilise the spatial distribution of Doppler velocities. Alternatively, 3D convolution can be used to extract features from all three dimensions in a 3D tensor, but requires huge computation and memory overheads [161]. RODNet [47] samples chirp sequences, as described above, to reduce input dimensionality.…”
Section: Pre-cfar Detectormentioning
confidence: 99%
“…CNNs have found numerous applications in video processing [32,49,59]. These include LSTM-based networks that perform per-frame encoding [44,50,59] and the extension of 2D convolutions to the temporal dimension, e.g., 3D CNNs such as C3D [48], R2D [43] and R(2+1)D [49].…”
Section: Deep Learning For Video Understandingmentioning
confidence: 99%
“…They demonstrated the success of their approach in relatively simple scenarios while being simple to train and deploy. Du et al [8] proposed more sophisticated models to perform steering angle prediction using 3D-CNN [25] and Long Short-Term Memory (LSTM) modules [12]. State-of-the-art performance was achieved in the Udacity challenge [41] Another line of works incorporates multi-task learning [31], side-task learning, or auxiliary task learning, in order to enhance the performance of a model on its main task.…”
Section: Related Workmentioning
confidence: 99%
“…Consequently, most contributions in this field focus on extending image-based algorithms with the temporal dimension. Examples are SIFT-3D [35], HOG3D [18], Action-Bank [32], and adaptions of convolutional neural networks [25,39,50]. Attempts to adapt the latter include LSTM-based network that firstly perform per-frame encoding [37,42,50].…”
Section: Introductionmentioning
confidence: 99%