LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

  • Mengmeng Liu
  • Hao Cheng
  • Lin Chen
  • Hellward Broszio
  • Jiangtao Li
  • Runjiang Zhao
  • Monika Sester
  • Michael Ying Yang

External Research Organisations

  • University of Twente
  • VISCODA GmbH
  • PhiGent Robotics
  • University of Bath
View graph of relations

Details

Original languageEnglish
Title of host publication2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
PublisherIEEE Computer Society
Pages2039-2049
Number of pages11
ISBN (electronic)9798350365474
ISBN (print)979-8-3503-6548-1
Publication statusPublished - 17 Jun 2024
Event2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024 - Seattle, United States
Duration: 16 Jun 202422 Jun 2024

Publication series

NameIEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
ISSN (Print)2160-7508
ISSN (electronic)2160-7516

Abstract

Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.

Keywords

    lane-aware selection, motion refinement, multimodal, Trajectory prediction

ASJC Scopus subject areas

Cite this

LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. / Liu, Mengmeng; Cheng, Hao; Chen, Lin et al.
2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society, 2024. p. 2039-2049 (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Liu, M, Cheng, H, Chen, L, Broszio, H, Li, J, Zhao, R, Sester, M & Yang, MY 2024, LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. in 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, IEEE Computer Society, pp. 2039-2049, 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024, Seattle, United States, 16 Jun 2024. https://doi.org/10.48550/arXiv.2302.13933, https://doi.org/10.1109/CVPRW63382.2024.00209
Liu, M., Cheng, H., Chen, L., Broszio, H., Li, J., Zhao, R., Sester, M., & Yang, M. Y. (2024). LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. In 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (pp. 2039-2049). (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops). IEEE Computer Society. https://doi.org/10.48550/arXiv.2302.13933, https://doi.org/10.1109/CVPRW63382.2024.00209
Liu M, Cheng H, Chen L, Broszio H, Li J, Zhao R et al. LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. In 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society. 2024. p. 2039-2049. (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops). doi: 10.48550/arXiv.2302.13933, 10.1109/CVPRW63382.2024.00209
Liu, Mengmeng ; Cheng, Hao ; Chen, Lin et al. / LAformer : Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society, 2024. pp. 2039-2049 (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops).
Download
@inproceedings{9631cb9c26594b4d88bd1d86fae3400c,
title = "LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints",
abstract = "Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.",
keywords = "lane-aware selection, motion refinement, multimodal, Trajectory prediction",
author = "Mengmeng Liu and Hao Cheng and Lin Chen and Hellward Broszio and Jiangtao Li and Runjiang Zhao and Monika Sester and Yang, {Michael Ying}",
note = "Publisher Copyright: {\textcopyright} 2024 IEEE.; 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024 ; Conference date: 16-06-2024 Through 22-06-2024",
year = "2024",
month = jun,
day = "17",
doi = "10.48550/arXiv.2302.13933",
language = "English",
isbn = "979-8-3503-6548-1",
series = "IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops",
publisher = "IEEE Computer Society",
pages = "2039--2049",
booktitle = "2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)",
address = "United States",

}

Download

TY - GEN

T1 - LAformer

T2 - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024

AU - Liu, Mengmeng

AU - Cheng, Hao

AU - Chen, Lin

AU - Broszio, Hellward

AU - Li, Jiangtao

AU - Zhao, Runjiang

AU - Sester, Monika

AU - Yang, Michael Ying

N1 - Publisher Copyright: © 2024 IEEE.

PY - 2024/6/17

Y1 - 2024/6/17

N2 - Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.

AB - Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.

KW - lane-aware selection

KW - motion refinement

KW - multimodal

KW - Trajectory prediction

UR - http://www.scopus.com/inward/record.url?scp=85206383198&partnerID=8YFLogxK

U2 - 10.48550/arXiv.2302.13933

DO - 10.48550/arXiv.2302.13933

M3 - Conference contribution

SN - 979-8-3503-6548-1

T3 - IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops

SP - 2039

EP - 2049

BT - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)

PB - IEEE Computer Society

Y2 - 16 June 2024 through 22 June 2024

ER -

By the same author(s)