2021
DOI: 10.48550/arxiv.2107.11355
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Unsupervised Domain Adaptive 3D Detection with Multi-Level Consistency

Abstract: Figure 1: Visualization of detection results for domain adaptation from KITTI to Waymo dataset. Left: Predictions of baseline model trained on KITTI dataset and directly tested on Waymo dataset. The model can classify and localize the objects, but produces inaccurate box scale due to geometric mismatch. The predicted boxes are therefore noticeably smaller than the ground truth. Right: Predictions of our domain-adaptive MLC-Net, which demonstrates accurate bounding box scale even though MLC-Net is trained witho… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
6
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
4

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(6 citation statements)
references
References 36 publications
(68 reference statements)
0
6
0
Order By: Relevance
“…As a result, when the detector is adapted to a sparse domain, it can not generate enough highquality pseudo labels to provide sufficient knowledge in the selftraining stage. TABLE 4 Unsupervised adaptation results of SF-UDA 3D [23], Dreaming [24], MLC-Net [26] and our ST3D++. We report AP 3D of car at IoU 0.7 and 0.5 on nuScenes → KITTI.…”
Section: Main Results and Analysismentioning
confidence: 99%
See 3 more Smart Citations
“…As a result, when the detector is adapted to a sparse domain, it can not generate enough highquality pseudo labels to provide sufficient knowledge in the selftraining stage. TABLE 4 Unsupervised adaptation results of SF-UDA 3D [23], Dreaming [24], MLC-Net [26] and our ST3D++. We report AP 3D of car at IoU 0.7 and 0.5 on nuScenes → KITTI.…”
Section: Main Results and Analysismentioning
confidence: 99%
“…We report AP 3D of car at IoU 0.7 and 0.5 on nuScenes → KITTI. To further demonstrate the advancement of our ST3D++, we compare it with SF-UDA 3D [23], Dreaming [24] and MLC-Net [26] on nuScenes → KITTI (i.e. the only common adaptation 2.…”
Section: Main Results and Analysismentioning
confidence: 99%
See 2 more Smart Citations
“…The second category is image translation-based which adapts image appearance to mitigate domain gaps [7,14,241,295]. The third category is self-training-based which predicts pseudo labels or minimizes entropy to guide iterative learning from target samples [13,141,183,184,242,296].…”
Section: Discussionmentioning
confidence: 99%