In contrast to traditional single-vehicle intelligent autonomous driving systems, which can only make predictions about the future based on their own perception of the environment, intelligent connected autonomous driving systems have the capability to enhance predictions by incorporating additional dynamic information about the surrounding road environment through V2X technology. Building upon the foundation of single-vehicle intelligent trajectory prediction, a specialized encoder was employd to enable the trajectory prediction model to seamlessly fuse its own perceptual information with dynamic road data obtained via V2X communication. The experimental results on the CARLA simulation dataset demonstrate that using V2X technology to obtain dynamic information of the surrounding road environment can more accurately predict vehicle trajectories compared to trajectory prediction algorithms that do not use dynamic environment information.
ChaiY, SappB, BansalM, et al. MultiPath: multiple probabilistic anchor trajectory hypotheses for behavior prediction[J/OL]. [2023-07-13].
[2]
CuiH G, RadosavljevicV, ChouF C, et al. Multimodal trajectory predictions for autonomous driving using deep convolutional networks[J/OL]. [2023-07-13].
[3]
CasasS, LuoW, UrtasunR. Intentnet: learning to predict intention from raw sensor data[C]∥Conference on Robot Learning, Zürich, Switzerland, 2018: 947-956.
[4]
Lee, N, ChoiW, VernazaP, et al. Desire: Distant future prediction in dynamic scenes with interacting agents[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 2017: 336-345.
[5]
HongJ, SappB, PhilbinJ. Rules of the road: predicting driving behavior with a convolutional model of semantic interactions[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 2019: 8454-8462.
[6]
JainA, CasasS, LiaoR, et al. Discrete residual flow for probabilistic pedestrian behavior prediction[DB/OL]. [2023-07-13].
[7]
DeoN, TrivediM M. Trajectory forecasts in unknown environments conditioned on grid-based plans[DB/OL]. [2023-07-13].
[8]
KhandelwalS, QiW, SinghJ, et al. What-if motion prediction for autonomous driving[DB/OL]. [2023-07-13].
[9]
MangalamK, AnY, GiraseH, et al. From goals, waypoints & paths to long term human trajectory forecasting[C]∥IEEE/CVF International Conference on Computer Vision, Nashville, TN, USA, 2021: 15233-15242.
[10]
Phan-MinhT, GrigoreE C, BoultonF A, et al. Covernet: multimodal behavior prediction using trajectory sets[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 2020: 14074-14083.
[11]
RhinehartN, McAllisterR, KitaniK, et al. Precog: prediction conditioned on goals in visual multiagent settings[C]∥IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 2019: 2821-2830.
[12]
RidelD, DeoN, WolfD, et al. Scene compliant trajectory forecast with agent-centric spatiotemporal grids[J]. IEEE Robotics and Automation Letters, 2020, 2(5): 2816-2823.
[13]
CasasS, GulinoC, LiaoR, alet, Spagnn: Spatially-aware graph neural networks for relational behavior forecasting from sensor data[C]∥IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 2020: 9491-9497.
[14]
HuangY, BiH, LiZ, alet, Stgat: Modeling spatial-temporal interactions for human trajectory prediction[C]∥IEEE/CVF International Conference on Computer Vision, Long Beach, CA, USA, 2019: 6272-6281.
[15]
LiangM, YangB, HuR, alet, Learning lane graph representations for motion forecasting[C]∥Computer Vision-ECCV 2020, Glasgow, UK, 2020: 541-556.
[16]
MohamedA, QianK, ElhoseinyM, alet, Social-stgcnn: A social spatio-temporal graph convolutional neural network for human trajectory prediction[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 2020: 14424-14432.
[17]
GaoJ, SunC, ZhaoH, alet, Vectornet: Encoding hd maps and agent dynamics from vectorized representation[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 2020: 11525-11533.
[18]
ZhaoH, GaoJ, LanT, et al. Tnt: Targetdriven trajectory prediction[C]∥Conference on Robot Learning, London, UK, 2021: 895-904.
[19]
ZengW, LiangM, LiaoR, et al. Lanercnn: distributed representations for graph-centric motion forecasting[C]∥IEEE/RSJ International Conference on Intelligent Robots and Systems, Prague, Czech Republic, 2021: 532-539.
[20]
AlahiA, GoelK, RamanathanV, alet, Social lstm: Human trajectory prediction in crowded spaces[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Las Vegas, Nevada, USA, 2016: 961-971.
[21]
DeoN, TrivediM M. Convolutional social pooling for vehicle trajectory prediction[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 2018: 1468-1476.
[22]
GuptaA, JohnsonJ, LiF F, et al. Social GAN: socially acceptable trajectories with generative adversarial networks[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 2018: 2255-2264.
[23]
MangalamK, GiraseH, AgarwalS, alet, It is not the journey but the destination: endpoint conditioned trajectory prediction[C]∥Computer Vision-ECCV 2020, Glasgow, UK, 2020: 759-776.
KosarajuV, SadeghianA, Mart´ın-Mart´ınR, alet, Social-bigat: multimodal trajectory forecasting using bicycle-GAN and graph attention networks[C]∥Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 2019, 32: 137-146.
[26]
LiL L, YangB, LiangM, alet, End-to-end contextual perception and prediction with interaction transformer[C]∥IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, USA, 2020: 5784-5791.
[27]
LiuY, ZhangJ, FangL, alet, Multimodal motion prediction with stacked transformers[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 2021: 7577-7586.
[28]
MercatJ, GillesT, ZoghbyN E, alet, Multi-head attention for multimodal joint vehicle motion forecasting[C]∥IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 2020: 9638-9644.
[29]
NgiamJ, CaineB, VasudevanV, et al. Scene transformer: a unified architecture for predicting multiple agent trajectories [DB/OL]. [2023-07-13].
[30]
YuC, MaX, RenJ, et al. Spatio-temporal graph transformer networks for pedestrian trajectory prediction[DB/OL]. [2023-07-13].
[31]
SalzmannT, IvanovicB, ChakravartyP, et al. Trajectron++: multi-agent generative trajectory forecasting with heterogeneous data for control[DB/OL]. [2023-07-13].
[32]
YeL, WangZ, ChenX, et al. GSAN: graph self-attention network for learning spatial-temporal interaction representation in autonomous driving[J]. IEEE Internet of Things Journal, 2021, 9(12): 9190-9204.
[33]
LuoC, SunL, DabiriD, et al. Probabilistic multi-modal trajectory prediction with lane attention for autonomous vehicles[C]∥IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Las Vegas, NV, USA, 2020: 2370-2376.
[34]
MessaoudK, DeoN, TrivediM M, et al. Multi-head attention with joint agent-map representation for trajectory prediction in autonomous driving[DB/OL]. [2023-07-13].
[35]
GillesT, SabatiniS, TsishkouD, et al. HOME: heatmap output for future motion estimation[J/OL]. [2023-07-13].
[36]
VaswaniA, ShazeerN, ParmarN, et al. Attention is all you need[J/OL]. [2023-07-13].
[37]
BrownT B, MannB, RyderN, et al. Language models are few-shot learners[DB/OL]. [2023-07-13].
[38]
CarionN, MassaF, SynnaeveG, et al. End-to-end object detection with transformers[C]∥Computer Vision-ECCV 2020: 16th European Conference, Berlin: Springer, 2020: 213-229.
[39]
DevlinJ, ChangM W, LeeK, et al. BERT: pre-training of deep bidirectional transformers for language understanding[DB/OL]. [2023-07-13].
[40]
DosovitskiyA, BeyerL, KolesnikovA, et al. An image is worth 16×16 words: Transformers for image recognition at scale [DB/OL]. [2023-07-13].
[41]
ArnabA, DehghaniM, HeigoldG, et al. Vivit: a video vision transformer[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 2021: 6836-6846.
[42]
HoJ, KalchbrennerN, WeissenbornD, et al. Axial attention in multidimensional transformers [DB/OL]. [2023-07-13].
[43]
LeeJ, LeeY, KimJ, et al. Set transformer: a framework for attention-based permutation-invariant neural networks[C]∥International Conference on Machine Learning, Long Beach, California, USA, 2019: 3744-3753.
[44]
BelloI, ZophB, VaswaniA, et al. Attention augmented convolutional networks[C]∥IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 2019: 3286-3295.
[45]
SrinivasA, LinT Y, ParmarN, et al. Bottleneck transformers for visual recognition[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 2021: 16519-16529.
[46]
HungW C, KretzschmarH, LinT Y, et al. Soda: Multi-object tracking with soft data association[DB/OL]. [2023-07-13].
[47]
RamachandranP, ParmarN, VaswaniA, et al. Stand-alone self-attention in vision models[DB/OL]. [2023-07-13].
[48]
TayY, DehghaniM, AbnarS, et al. Long range arena: a benchmark for efficient transformers [DB/OL]. [2023-07-13].
[49]
TayY, DehghaniM, BahriD, et al. Efficient transformers: a survey[J]. ACM Computing Surveys, 2022, 55(6): 1-28.
[50]
HeK, ChenX, XieS, et al. Masked autoencoders are scalable vision learners[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 2022: 16000-16009.
[51]
GiuliariF, HasanI, CristaniM, et al. Transformer networks for trajectory forecasting[C]∥International conference on pattern recognition (ICPR), Montréal, QC, Canada, 2021: 10335-10342.
[52]
ZhouZ, YeL, WangJ, alet, Hivt: hierarchical vector transformer for multi-agent motion prediction[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 2022: 8823-8833.
[53]
JungC, LeeD, LeeS, et al. V2X communication-aided autonomous driving: system design and experimental validation[J]. Sensors, 2020, 20(10): No.2903.
[54]
DengR, DiB, SongL. Cooperative collision avoidance for overtaking maneuvers in cellular V2X-based autonomous driving[J]. IEEE Transactions on Vehicular Technology, 2019, 68(5): 4434-4446.
[55]
ChangM F, LambertJ, SangkloyP, alet, Argoverse: 3D tracking and forecasting with rich maps[C]∥IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 2019: 8748-8757.
[56]
HoustonJ, ZuidhofG, BergaminiL, alet, One thousand and one hours: self-driving motion prediction dataset[C]∥Conference on Robot Learning, London, UK, 2021: 409-418.
[57]
EttingerS, ChengS, CaineB, et al. Large scale interactive motion forecasting for autonomous driving: the waymo open motion dataset[C]∥IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 2021: 9710-9719.
[58]
CaesarH, BankitiV, LangA H, et al.Nuscenes: a multimodal dataset for autonomous driving[C]∥IEEE/CVF Conference on Computer Cision and Pattern Recognition, Seattle, WA, USA, 2020: 11621-11631.
[59]
DosovitskiyA, RosG, CodevillaF, et al. CARLA: an open urban driving simulator[C]∥Conference on Robot Learning, Mountain View, California, 2017: 1-16.