2019
DOI: 10.48550/arxiv.1903.11027
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

nuScenes: A multimodal dataset for autonomous driving

Abstract: Robust detection and tracking of objects is crucial for the deployment of autonomous vehicle technology. Imagebased benchmark datasets have driven development in computer vision tasks such as object detection, tracking and segmentation of agents in the environment. Most autonomous vehicles, however, carry a combination of cameras and range sensors such as lidar and radar. As machine learning based methods for detection and tracking become more prevalent, there is a need to train and evaluate such methods on da… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

2
626
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 291 publications
(628 citation statements)
references
References 62 publications
(115 reference statements)
2
626
0
Order By: Relevance
“…However, SC3D uses exhaustive search to generate numerous candidate proposals for template matching, so it performs well with few training samples. For the nuScenes [6] dataset, we directly apply the models, trained on the corresponding categories of the KITTI dataset, to evaluate performance on the nuScenes dataset. Specifically, the corresponding categories between KITTI and nuScenes datasets are Car→Car, Pedestrian→Pedestrian, Van→Truck, and Cyclist→Bicycle, respectively.…”
Section: Resultsmentioning
confidence: 99%
See 2 more Smart Citations
“…However, SC3D uses exhaustive search to generate numerous candidate proposals for template matching, so it performs well with few training samples. For the nuScenes [6] dataset, we directly apply the models, trained on the corresponding categories of the KITTI dataset, to evaluate performance on the nuScenes dataset. Specifically, the corresponding categories between KITTI and nuScenes datasets are Car→Car, Pedestrian→Pedestrian, Van→Truck, and Cyclist→Bicycle, respectively.…”
Section: Resultsmentioning
confidence: 99%
“…We develop a voxel-to-BEV target localization network, which can accurately detect the 3D target's center in the dense BEV space compared to sparse 3D space. Extensive results show that our method has achieved new state-of-the-art results on the KITTI dataset [20], and has a good generalization ability on the nuScenes [6] dataset.…”
Section: Introductionmentioning
confidence: 93%
See 1 more Smart Citation
“…Those two setups were chosen to provide a diverse perception input to the ego vehicle, in terms of environment constellations and object types. In addition to that, we have tested position errors with the NuScenes dataset [21] as an example of non-simulated LiDAR information. Each scenario duration was sufficiently long for the environment to contain more than 1000 relevant object states in scope, in order to guarantee statistical significance.…”
Section: A Experimental Setupmentioning
confidence: 99%
“…Finally, we also explore the error detection capability of our monitor with real LiDAR data from the NuScenes dataset [21]. Since no ground truth velocity data is provided here, we restrict ourselves to the analysis of permanent position faults with sensor checks only.…”
Section: F Position Error Of a Real Sensormentioning
confidence: 99%