2022 IEEE Intelligent Vehicles Symposium (IV) 2022
DOI: 10.1109/iv51971.2022.9827310
|View full text |Cite
|
Sign up to set email alerts
|

Spatiotemporal Transformer Attention Network for 3D Voxel Level Joint Segmentation and Motion Prediction in Point Cloud

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2

Citation Types

0
4
0

Year Published

2024
2024
2024
2024

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(4 citation statements)
references
References 26 publications
0
4
0
Order By: Relevance
“…These methods represent the environment with BEV maps derived from point clouds and aim to predict the 2D displacement vector for each BEV cell along the horizontal plane. MotionNet (Wu, Chen, and Metaxas 2020) and BE-STI (Wei et al 2022) proposes to perform joint category perception and motion prediction from the BEV maps. LSTM-ED (Schreiber, Hoermann, and Dietmayer 2019) introduce convolutional LSTM (Shi et al 2015) to aggregate temporal context.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…These methods represent the environment with BEV maps derived from point clouds and aim to predict the 2D displacement vector for each BEV cell along the horizontal plane. MotionNet (Wu, Chen, and Metaxas 2020) and BE-STI (Wei et al 2022) proposes to perform joint category perception and motion prediction from the BEV maps. LSTM-ED (Schreiber, Hoermann, and Dietmayer 2019) introduce convolutional LSTM (Shi et al 2015) to aggregate temporal context.…”
Section: Related Workmentioning
confidence: 99%
“…However, these approaches may face challenges when handling categories that have not been seen in the training set, mainly due to their reliance on object detection (Wu, Chen, and Metaxas 2020). To address this challenge, class-agnostic motion prediction task (Schreiber, Hoermann, and Dietmayer 2019;Wu, Chen, and Metaxas 2020;Wang et al 2022;Wei et al 2022) is proposed to provide complementary information. These methods take a sequence of previous point clouds as input and predict the future displacements for each Bird's Eye View (BEV) cell.…”
Section: Introductionmentioning
confidence: 99%
“…These methods represent the environment with BEV maps derived from point clouds and aim to predict the 2D displacement vector for each BEV cell along the horizontal plane. MotionNet (Wu, Chen, and Metaxas 2020) and BE-STI (Wei et al 2022) proposes to perform joint category perception and motion prediction from the BEV maps. LSTM-ED (Schreiber, Hoermann, and Dietmayer 2019) introduce convolutional LSTM (Shi et al 2015) to aggregate temporal context.…”
Section: Related Workmentioning
confidence: 99%
“…However, these approaches may face challenges when handling categories that have not been seen in the training set, mainly due to their reliance on object detection (Wu, Chen, and Metaxas 2020). To address this challenge, class-agnostic motion prediction task (Schreiber, Hoermann, and Dietmayer 2019;Wu, Chen, and Metaxas 2020;Wang et al 2022;Wei et al 2022) is proposed to provide complementary information. These methods take a sequence of previous point clouds as input and predict the future displacements for each Bird's Eye View (BEV) cell.…”
Section: Introductionmentioning
confidence: 99%