GLOBECOM 2017 - 2017 IEEE Global Communications Conference 2017
DOI: 10.1109/glocom.2017.8254480
|View full text |Cite
|
Sign up to set email alerts
|

VANETs Meet Autonomous Vehicles: A Multimodal 3D Environment Learning Approach

Abstract: In this paper, we design a multimodal framework for object detection, recognition and mapping based on the fusion of stereo camera frames, point cloud Velodyne Lidar scans, and Vehicle-to-Vehicle (V2V) Basic Safety Messages (BSMs) exchanged using Dedicated Short Range Communication (DSRC). We merge the key features of rich texture descriptions of objects from 2D images, depth and distance between objects provided by 3D point cloud and awareness of hidden vehicles from BSMs' 3D information. We present a joint p… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
5
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4
2

Relationship

2
8

Authors

Journals

citations
Cited by 20 publications
(5 citation statements)
references
References 12 publications
0
5
0
Order By: Relevance
“…More recently, [40] and [22] incorporated deep learning into the V2V system: multiple intelligent vehicles share the intermediate features output by the neural network to promote the vehicle's perception capability. As for the dataset, [5,24,42] simulated the V2V scenarios with different frames from KITTI [11]. Yet, they were unrealistic for not capturing the measurements at the same time.…”
Section: Related Workmentioning
confidence: 99%
“…More recently, [40] and [22] incorporated deep learning into the V2V system: multiple intelligent vehicles share the intermediate features output by the neural network to promote the vehicle's perception capability. As for the dataset, [5,24,42] simulated the V2V scenarios with different frames from KITTI [11]. Yet, they were unrealistic for not capturing the measurements at the same time.…”
Section: Related Workmentioning
confidence: 99%
“…XIV. POST ALIGNMENT AND FUSION Maalej et al presented in [160] a multimodal framework for object detection, recognition and mapping based on the fusion of stereo camera frames, point cloud Velodyne LIDAR scans, and V2V exchanged BSMs over DSRC. Based on the adapted Darknet's CNN, the pixel-wise adjacency coordinates of moments are derived from the bounding boxes of recognized objects of KITTI frames.…”
Section: Part 4: Towards Vcs Meeting Autonomous Vehiclesmentioning
confidence: 99%
“…1. WAVE recommends that, for a fixed SI equal to 100ms, every vehicle needs to be tuned to the CCH in order to exchange their safety related messages with neighboring vehicles as detailed in [30] and in [31]. WAVE differs between various Quality of Service (QoS) levels by using specific fields in the packets that are sent over CCH or SCH.…”
Section: A Wave 80211p Macmentioning
confidence: 99%