Accurate vehicle trajectory prediction is crucial for ensuring safe and efficient autonomous driving. This work explores the integration of Transformer based model with Long Short-Term Memory (LSTM) based technique to enhance spatial and temporal feature learning in vehicle trajectory prediction. Here, a hybrid model that combines LSTMs for temporal encoding with a Transformer encoder for capturing complex interactions between vehicles is proposed. Spatial trajectory features of the neighboring vehicles are processed and goes through a masked scatter mechanism in a grid based environment, which is then combined with temporal trajectory of the vehicles. This combined trajectory data are learned by sequential LSTM encoding and Transformer based attention layers. The proposed model is benchmarked against predecessor LSTM based methods, including STA-LSTM, SA-LSTM, CS-LSTM, and NaiveLSTM. Our results, while not outperforming it's predecessor, demonstrate the potential of integrating Transformers with LSTM based technique to build interpretable trajectory prediction model. Future work will explore alternative architectures using Transformer applications to further enhance performance. This study provides a promising direction for improving trajectory prediction models by leveraging transformer based architectures, paving the way for more robust and interpretable vehicle trajectory prediction system.
翻译:精确的车辆轨迹预测对于确保自动驾驶的安全性和效率至关重要。本研究探索将基于Transformer的模型与基于长短期记忆(LSTM)的技术相结合,以增强车辆轨迹预测中的空间和时间特征学习。本文提出了一种混合模型,该模型结合了用于时间编码的LSTM和用于捕获车辆间复杂交互的Transformer编码器。邻车的空间轨迹特征在基于网格的环境中被处理并通过掩码散射机制,随后与车辆的时间轨迹相结合。这些组合的轨迹数据通过序列LSTM编码和基于Transformer的注意力层进行学习。所提出的模型与基于LSTM的先前方法进行了基准测试,包括STA-LSTM、SA-LSTM、CS-LSTM和NaiveLSTM。我们的结果虽然未超越其前身,但证明了将Transformer与基于LSTM的技术相结合以构建可解释轨迹预测模型的潜力。未来的工作将探索使用Transformer应用的替代架构以进一步提升性能。本研究通过利用基于Transformer的架构,为改进轨迹预测模型提供了一个有前景的方向,为构建更鲁棒和可解释的车辆轨迹预测系统铺平了道路。