BEV-V2X: Cooperative Birds-Eye-View Fusion and Grid Occupancy Prediction via V2X-Based Data Sharing

Birds-Eye-View (BEV) perception can naturally represent natural scenes, which is conducive to multimodal data processing and fusion. BEV data contain rich semantics and integrate the information of driving scenes, which play an important role in researches related to autonomous driving. However, BEV constructed by single vehicle perception encounter certain issues, such as low accuracy and insufficient range, and thus cannot be well applied to scenario understanding and driving situation prediction. To address the challenges, this article proposes a novel data-driven approach based on vehicle-to-everything (V2X) communication. The roadside unit or cloud center collects local BEV data from all connected and automated vehicles (CAVs) within the control area, then fuses and predicts the future global BEV occupancy grid map. It provides powerful support for driving safety warning, cooperative driving planning, cooperative traffic control and other applications. More precisely, the authors develop an attention-based cooperative BEV fusion and prediction model called BEV-V2X. The authors also compare the performance of BEV-V2X with that of single vehicle prediction. Experimental results demonstrate that the authors' proposed method achieves higher accuracy. Even in cases where not all vehicles are CAVs, the model can still comprehensively estimate and predict global spatiotemporal changes. The authors also discuss the impact of the CAV rate, single vehicle perception ability, and grid size on the fusion and prediction results.

Language

  • English

Media Info

Subject/Index Terms

Filing Info

  • Accession Number: 01905975
  • Record Type: Publication
  • Files: TRIS
  • Created Date: Jan 26 2024 10:02AM