Paper ThAT4.3
Teufel, Sven (University of Tübingen), Gamerdinger, Jörg (Eberhard Karls Universität Tübingen), Volk, Georg (Eberhard Karls Universität Tübingen), Bringmann, Oliver (Eberhard Karls Universität Tübingen)
MR3D-Net: Dynamic Multi-Resolution 3D Sparse Voxel Grid Fusion for LiDAR-Based Collective Perception
Scheduled for presentation during the Regular Session "Collective perception and localization" (ThAT4), Thursday, September 26, 2024,
11:10−11:30, Salon 7
2024 IEEE 27th International Conference on Intelligent Transportation Systems (ITSC), September 24- 27, 2024, Edmonton, Canada
This information is tentative and subject to change. Compiled on December 26, 2024
|
|
Keywords Cooperative Techniques and Systems, Multi-autonomous Vehicle Studies, Models, Techniques and Simulations, Sensing, Vision, and Perception
Abstract
The safe operation of automated vehicles depends on their ability to perceive the environment comprehensively. However, occlusion, sensor range, and environmental factors limit their perception capabilities. To overcome these limitations, collective perception enables vehicles to exchange information. However, fusing this exchanged information is a challenging task. Early fusion approaches require large amounts of bandwidth, while intermediate fusion approaches face interchangeability issues. Late fusion of shared detections is currently the only feasible approach. However, it often results in inferior performance due to information loss. To address this issue, we propose MR3D-Net, a dynamic multi-resolution 3D sparse voxel grid fusion backbone architecture for LiDAR-based collective perception. We show that sparse voxel grids at varying resolutions provide a meaningful and compact environment representation that can adapt to the communication bandwidth. MR3D-Net achieves state-of-the-art performance on the OPV2V 3D object detection benchmark while reducing the required bandwidth by up to 94% compared to early fusion.
|
|