ITSC 2024 Paper Abstract

Close

Paper WeBT13.8

Pu, Yuxuan (Hunan University), Li, Yang (State key laboratory of automotive safety and energy, Tsinghua U), Xia, Beihao (Huazhong University of Science and Technology), Wang, Xiaowei (Hunan University), qin, hongmao (hunan university), Zhu, Lei (Mogo ai)

A Lightweight Lane-Guided Vector Transformer for Multi-Agent Trajectory Prediction in Autonomous Driving

Scheduled for presentation during the Poster Session "Transformer networks" (WeBT13), Wednesday, September 25, 2024, 14:30−16:30, Foyer

2024 IEEE 27th International Conference on Intelligent Transportation Systems (ITSC), September 24- 27, 2024, Edmonton, Canada

This information is tentative and subject to change. Compiled on December 26, 2024

Keywords Sensing, Vision, and Perception

Abstract

Multi-agent trajectory prediction has been crucial for the safe decision-making of autonomous driving in complex transportation scenarios. Making accurate and real-time predictions is of great significance, and interaction modeling has shown great potential in improving prediction accuracy, but it is a non-trivial task to capture the complicated interaction mechanisms between agents. Existing learning-based methods that consider interaction modeling often have large model sizes, sacrificing the inference speed in pursuit of superior accuracy, making them hard for practical implementations. To tackle this issue, we present a lightweight Lane-Guided Vector Transformer (LGVT) to balance the tradeoff between the inference speed and model accuracy. Specifically, we employ the Attention Free Transformer module that eliminates the need for dot product self-attention to efficiently learn the spatial-temporal dependency and leverages a motion encoder to encode agent motion features. In addition, we use a lane-guided representation that defines a region centered along the ground truth trajectory to align the agent’s motion states with lane constraints, which helps filter out the irrelevant lane segments and accelerate the learning process. We conduct extensive experiments on the Argoverse 1 dataset and compare our model with several baselines. Results indicate that our model can outperform the baseline HiVT-64 regarding the prediction accuracy metrics minADE, minFDE, and MR. We also achieve an approximate accuracy of the large version HiVT-128 with a 70% decrease in model parameters and a 20% increase in the inference speed.

 

 

All Content © PaperCept, Inc.


This site is protected by copyright and trademark laws under US and International law.
All rights reserved. © 2002-2024 PaperCept, Inc.
Page generated 2024-12-26  13:19:06 PST  Terms of use