2020
DOI: 10.48550/arxiv.2005.10863
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

RV-FuseNet: Range View Based Fusion of Time-Series LiDAR Data for Joint 3D Object Detection and Motion Forecasting

Abstract: Autonomous vehicles rely on robust real-time detection and future motion prediction of traffic participants to safely navigate urban environments. We present a novel end-toend approach that uses raw time-series LiDAR data to jointly solve both detection and prediction. We use the range view representation of LiDAR instead of voxelization since it does not discard information and is more efficient due to its compactness. However, for time-series fusion the data needs to be projected to a common viewpoint, and o… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
19
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
2
1

Relationship

1
2

Authors

Journals

citations
Cited by 3 publications
(19 citation statements)
references
References 23 publications
0
19
0
Order By: Relevance
“…These methods, by virtue of operating in BEV, lose out on high-resolution point information and are often limited by range of operation. RV based methods such as [6] and [3] overcome the limitation on operating range but are outperformed in the motion forecasting task by recent BEV based methods. In this work, we improve the joint framework by including multi-view representation in multiple parts of the network and achieve state-of-the-art performance on both object detection and motion forecasting while scaling to large areas of operation in real-time.…”
Section: Motion Forecastingmentioning
confidence: 99%
See 4 more Smart Citations
“…These methods, by virtue of operating in BEV, lose out on high-resolution point information and are often limited by range of operation. RV based methods such as [6] and [3] overcome the limitation on operating range but are outperformed in the motion forecasting task by recent BEV based methods. In this work, we improve the joint framework by including multi-view representation in multiple parts of the network and achieve state-of-the-art performance on both object detection and motion forecasting while scaling to large areas of operation in real-time.…”
Section: Motion Forecastingmentioning
confidence: 99%
“…For multiple views this can be trivially extended by projecting the points in both BEV and RV for aggregation. However, directly projecting all the past LiDAR data into the RV of the most recent sweep leads to significant performance degradation due to heavy data loss in the projection step [3]. Therefore, instead of previous approaches that focus on one-shot projection, we propose a novel sequential multi-view fusion approach to effectively aggregate the temporal LiDAR data.…”
Section: Multi-view Temporal Fusion Networkmentioning
confidence: 99%
See 3 more Smart Citations