LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

  • Mengmeng Liu
  • Hao Cheng
  • Lin Chen
  • Hellward Broszio
  • Jiangtao Li
  • Runjiang Zhao
  • Monika Sester
  • Michael Ying Yang

Externe Organisationen

  • University of Twente
  • VISCODA GmbH
  • PhiGent Robotics
  • University of Bath
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des Sammelwerks2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)
Herausgeber (Verlag)IEEE Computer Society
Seiten2039-2049
Seitenumfang11
ISBN (elektronisch)9798350365474
ISBN (Print)979-8-3503-6548-1
PublikationsstatusVeröffentlicht - 17 Juni 2024
Veranstaltung2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024 - Seattle, USA / Vereinigte Staaten
Dauer: 16 Juni 202422 Juni 2024

Publikationsreihe

NameIEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops
ISSN (Print)2160-7508
ISSN (elektronisch)2160-7516

Abstract

Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.

ASJC Scopus Sachgebiete

Zitieren

LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. / Liu, Mengmeng; Cheng, Hao; Chen, Lin et al.
2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society, 2024. S. 2039-2049 (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops).

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Liu, M, Cheng, H, Chen, L, Broszio, H, Li, J, Zhao, R, Sester, M & Yang, MY 2024, LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. in 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops, IEEE Computer Society, S. 2039-2049, 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024, Seattle, USA / Vereinigte Staaten, 16 Juni 2024. https://doi.org/10.48550/arXiv.2302.13933, https://doi.org/10.1109/CVPRW63382.2024.00209
Liu, M., Cheng, H., Chen, L., Broszio, H., Li, J., Zhao, R., Sester, M., & Yang, M. Y. (2024). LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. In 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW) (S. 2039-2049). (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops). IEEE Computer Society. https://doi.org/10.48550/arXiv.2302.13933, https://doi.org/10.1109/CVPRW63382.2024.00209
Liu M, Cheng H, Chen L, Broszio H, Li J, Zhao R et al. LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. in 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society. 2024. S. 2039-2049. (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops). doi: 10.48550/arXiv.2302.13933, 10.1109/CVPRW63382.2024.00209
Liu, Mengmeng ; Cheng, Hao ; Chen, Lin et al. / LAformer : Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints. 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW). IEEE Computer Society, 2024. S. 2039-2049 (IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops).
Download
@inproceedings{9631cb9c26594b4d88bd1d86fae3400c,
title = "LAformer: Trajectory Prediction for Autonomous Driving with Lane-Aware Scene Constraints",
abstract = "Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.",
keywords = "lane-aware selection, motion refinement, multimodal, Trajectory prediction",
author = "Mengmeng Liu and Hao Cheng and Lin Chen and Hellward Broszio and Jiangtao Li and Runjiang Zhao and Monika Sester and Yang, {Michael Ying}",
note = "Publisher Copyright: {\textcopyright} 2024 IEEE.; 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024 ; Conference date: 16-06-2024 Through 22-06-2024",
year = "2024",
month = jun,
day = "17",
doi = "10.48550/arXiv.2302.13933",
language = "English",
isbn = "979-8-3503-6548-1",
series = "IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops",
publisher = "IEEE Computer Society",
pages = "2039--2049",
booktitle = "2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)",
address = "United States",

}

Download

TY - GEN

T1 - LAformer

T2 - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, CVPRW 2024

AU - Liu, Mengmeng

AU - Cheng, Hao

AU - Chen, Lin

AU - Broszio, Hellward

AU - Li, Jiangtao

AU - Zhao, Runjiang

AU - Sester, Monika

AU - Yang, Michael Ying

N1 - Publisher Copyright: © 2024 IEEE.

PY - 2024/6/17

Y1 - 2024/6/17

N2 - Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.

AB - Existing trajectory prediction methods for autonomous driving typically rely on one-stage trajectory prediction models, which condition future trajectories on observed trajectories combined with fused scene information. However, they often struggle with complex scene constraints, such as those encountered at intersections. To this end, we present a novel method, called LAformer. It uses an attention-based temporally dense lane-aware estimation module to continuously estimate the likelihood of the alignment between motion dynamics and scene information extracted from an HD map. Additionally, unlike one-stage prediction models, LAformer utilizes predictions from the first stage as anchor trajectories. It leverages a second-stage motion refinement module to further explore temporal consistency across the complete time horizon. Extensive experiments on nuScenes and Argoverse 1 demonstrate that LAformer achieves excellent generalized performance for multimodal trajectory prediction. The source code of LAformer is available at https://github.com/mengmengliu1998/LAformer.

KW - lane-aware selection

KW - motion refinement

KW - multimodal

KW - Trajectory prediction

UR - http://www.scopus.com/inward/record.url?scp=85206383198&partnerID=8YFLogxK

U2 - 10.48550/arXiv.2302.13933

DO - 10.48550/arXiv.2302.13933

M3 - Conference contribution

SN - 979-8-3503-6548-1

T3 - IEEE Computer Society Conference on Computer Vision and Pattern Recognition Workshops

SP - 2039

EP - 2049

BT - 2024 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW)

PB - IEEE Computer Society

Y2 - 16 June 2024 through 22 June 2024

ER -

Von denselben Autoren