Li Ding, Zhang Qichao, Lu Shuai, Pan Yifeng, Zhao Dongbin
IEEE Trans Neural Netw Learn Syst. 2024 Dec;35(12):18758-18770. doi: 10.1109/TNNLS.2023.3321564. Epub 2024 Dec 2.
Predicting future trajectories of pairwise traffic agents in highly interactive scenarios, such as cut-in, yielding, and merging, is challenging for autonomous driving. The existing works either treat such a problem as a marginal prediction task or perform single-axis factorized joint prediction, where the former strategy produces individual predictions without considering future interaction, while the latter strategy conducts conditional trajectory-oriented prediction via agentwise interaction or achieves conditional rollout-oriented prediction via timewise interaction. In this article, we propose a novel double-axis factorized joint prediction pipeline, namely, conditional goal-oriented trajectory prediction (CGTP) framework, which models future interaction both along the agent and time axes to achieve goal and trajectory interactive prediction. First, a goals-of-interest network (GoINet) is designed to extract fine-grained features of goal candidates via hierarchical vectorized representation. Furthermore, we propose a conditional goal prediction network (CGPNet) to produce multimodal goal pairs in an agentwise conditional manner, along with a newly designed goal interactive loss to better learn the joint distribution of the intermediate interpretable modes. Explicitly guided by the goal-pair predictions, we propose a goal-oriented trajectory rollout network (GTRNet) to predict scene-compliant trajectory pairs via timewise interactive rollouts. Extensive experimental results confirm that the proposed CGTP outperforms the state-of-the-art (SOTA) prediction models on the Waymo open motion dataset (WOMD), Argoverse motion forecasting dataset, and In-house cut-in dataset. Code is available at https://github.com/LiDinga/CGTP/.
在诸如切入、让路和合并等高交互场景中预测成对交通代理的未来轨迹,对于自动驾驶来说具有挑战性。现有工作要么将此类问题视为边缘预测任务,要么执行单轴分解联合预测,其中前一种策略在不考虑未来交互的情况下生成单独的预测,而后一种策略通过智能体间交互进行面向条件轨迹的预测,或者通过时间交互实现面向条件展开的预测。在本文中,我们提出了一种新颖的双轴分解联合预测管道,即条件目标导向轨迹预测(CGTP)框架,该框架沿智能体和时间轴对未来交互进行建模,以实现目标和轨迹的交互预测。首先,设计了一个感兴趣目标网络(GoINet),通过分层矢量化表示来提取目标候选的细粒度特征。此外,我们提出了一个条件目标预测网络(CGPNet),以智能体条件方式生成多模态目标对,同时还设计了一种新的目标交互损失,以更好地学习中间可解释模式的联合分布。在目标对预测的明确指导下,我们提出了一个目标导向轨迹展开网络(GTRNet),通过时间交互展开来预测符合场景的轨迹对。大量实验结果证实,所提出的CGTP在Waymo开放运动数据集(WOMD)、Argoverse运动预测数据集和内部切入数据集上优于现有最先进(SOTA)预测模型。代码可在https://github.com/LiDinga/CGTP/获取。