2022
DOI: 10.48550/arxiv.2205.15997
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

TransFuser: Imitation with Transformer-Based Sensor Fusion for Autonomous Driving

Abstract: How should we integrate representations from complementary sensors for autonomous driving? Geometry-based fusion has shown promise for perception (e.g. object detection, motion forecasting). However, in the context of end-to-end driving, we find that imitation learning based on existing sensor fusion methods underperforms in complex driving scenarios with a high density of dynamic agents. Therefore, we propose TransFuser, a mechanism to integrate image and LiDAR representations using self-attention. Our approa… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
5
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
2
2

Relationship

0
4

Authors

Journals

citations
Cited by 4 publications
(7 citation statements)
references
References 49 publications
0
5
0
Order By: Relevance
“…To comprehensively address the multiperspective requirements of our model and the following framework mention in Section 3.4, we use three typical end-to-end autonomous driving model: neat (Chitta et al, 2021), transfuser (Chitta et al, 2022) and LAV (Chen and Krähenbühl, 2022) as the standard navigation model in our evaluation. Neat presents an overall approach based on a multilayer perceptron neural attention field to conduct fully endto-end navigation, which first takes image patches from front, left, right three perspectives in 180-degree range, combined with velocity features as inputs and then summed and fed into a transformer with attention mechanism for final navigation tasks.…”
Section: Navigation Robustness Enhancement Evaluationmentioning
confidence: 99%
See 2 more Smart Citations
“…To comprehensively address the multiperspective requirements of our model and the following framework mention in Section 3.4, we use three typical end-to-end autonomous driving model: neat (Chitta et al, 2021), transfuser (Chitta et al, 2022) and LAV (Chen and Krähenbühl, 2022) as the standard navigation model in our evaluation. Neat presents an overall approach based on a multilayer perceptron neural attention field to conduct fully endto-end navigation, which first takes image patches from front, left, right three perspectives in 180-degree range, combined with velocity features as inputs and then summed and fed into a transformer with attention mechanism for final navigation tasks.…”
Section: Navigation Robustness Enhancement Evaluationmentioning
confidence: 99%
“…As mentioned above, we select three multiperspective navigation methods neat (Chitta et al, 2021), transfuser (Chitta et al, 2022) and LAV (Chen and Krähenbühl, 2022) combined with our reconstruction model to fully evaluate the augmentation of the robustness from our model in autonomous driving with various driving scenarios. In our evaluation, we consider two different routes town02 and town05 in CARLA as our test scenarios under various disturbances and attacks.…”
Section: Navigation Robustness Comparisonmentioning
confidence: 99%
See 1 more Smart Citation
“…To satisfy this time-varying power balance constraint, the safety layer provides a feasible solution space, which enables the RL agent to explore safely in it. Such a safety layer is common in many other similar tasks with high-security requirements [31,32].…”
Section: Das Model-free Rl Gridzeromentioning
confidence: 99%
“…LAV [45] trains the driving agent with the dataset collected from all the vehicles that it observes. Transfuser [8,46] is an imitation learning method where the agent uses transformer to fuse information from the front camera image and LiDAR information. The entries "Latent Transfuser" and "Trans-fuser+" are variants of Transfuser.…”
Section: Route Completionmentioning
confidence: 99%