2018 IEEE Intelligent Vehicles Symposium (IV) 2018
DOI: 10.1109/ivs.2018.8500395
|View full text |Cite
|
Sign up to set email alerts
|

MB-Net: MergeBoxes for Real-Time 3D Vehicles Detection

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
9
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 13 publications
(9 citation statements)
references
References 24 publications
0
9
0
Order By: Relevance
“…This innovation enables InvPT++ to markedly surpass InvPT while utilizing 22.51% fewer FLOPs in the transformer decoder. We remarkably extend our experiments to investigate the effectiveness of InvPT++ by considering two new and more challenging 2D/3D multi-task scene understanding benchmarks, i.e., the Cityscapes [19] and Cityscapes-3D [20] datasets. Substantial results and analysis in qualitative and quantitative aspects are further provided in the experiments.…”
Section: Global Spatial Interaction Simultaneous All-task Interactionmentioning
confidence: 99%
See 1 more Smart Citation
“…This innovation enables InvPT++ to markedly surpass InvPT while utilizing 22.51% fewer FLOPs in the transformer decoder. We remarkably extend our experiments to investigate the effectiveness of InvPT++ by considering two new and more challenging 2D/3D multi-task scene understanding benchmarks, i.e., the Cityscapes [19] and Cityscapes-3D [20] datasets. Substantial results and analysis in qualitative and quantitative aspects are further provided in the experiments.…”
Section: Global Spatial Interaction Simultaneous All-task Interactionmentioning
confidence: 99%
“…Fig. 7: Qualitative comparison between our predictions and the ground-truth labels (GT) for joint 2D-3D multi-task scene understanding using the Cityscapes-3D dataset [20].…”
Section: Selective Attentionmentioning
confidence: 99%
“…The projection of the predicted 3D BBox should be consistent with the predicted 2D BBox. This is utilized to build geometric constraints in [27,14,26] to determine the depth. Thanks to the promising performance of convolutional neural networks in 2D object detection, more and more approaches [3,20,31,30,29,24,12,16] have been proposed to directly predict 3D BBoxes using well-designed convolutional neural network for monocular 3D object detection.…”
Section: Related Workmentioning
confidence: 99%
“… Deeplearning: Voxelnet [49]; Second [50]; MVX-Net [51]  Other: Particle filter [52]  Original 3D data information can be retained  SSD [64]; YOLOv3 [65]; MB-Net [66]; EZ-Net [67];…”
Section: Volumetricmentioning
confidence: 99%