Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving

Research output: Contribution to journalArticleResearchpeer review

Authors

View graph of relations

Details

Original languageEnglish
Pages (from-to)3054 - 3061
Number of pages8
JournalIEEE Robotics and Automation Letters
Volume7
Issue number2
Publication statusPublished - 14 Jan 2022

Abstract

Sharing collective perception messages (CPM) between vehicles is investigated to decrease occlusions, so as to improve perception accuracy and safety of autonomous driving. However, highly accurate data sharing and low communication overhead is a big challenge for collective perception, especially when real-time communication is required among connected and automated vehicles. In this paper, we propose an efficient and effective keypoints-based deep feature fusion framework, called FPV-RCNN, for collective perception, which is built on top of the 3D object detector PV-RCNN. We introduce a bounding box proposal matching module and a keypoints selection strategy to compress the CPM size and solve the multi-vehicle data fusion problem. Compared to a bird's-eye view (BEV) keypoints feature fusion, FPV-RCNN achieves improved detection accuracy by about 14% at a high evaluation criterion (IoU 0.7) on a synthetic dataset COMAP dedicated to collective perception. Also, its performance is comparable to two raw data fusion baselines that have no data loss in sharing. Moreover, our method also significantly decreases the CPM size to less than 0.3KB, which is about 50 times smaller than the BEV feature map sharing used in previous works. Even with a further decreased number of CPM feature channels, i.e., from 128 to 32, the detection performance only drops about 1%. The code of our method is available at https://github.com/YuanYunshuang/FPV_RCNN.

Keywords

    Data integration, Feature extraction, Location awareness, Object Detection, Point cloud compression, Proposals, Segmentation and Categorization, Sensor Fusion, Sensor Networks, Three-dimensional displays, Vehicle detection, Sensor fusion, object detection, sensor networks, segmentation and categorization

ASJC Scopus subject areas

Cite this

Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving. / Yuan, Yunshuang; Cheng, Hao; Sester, Monika.
In: IEEE Robotics and Automation Letters, Vol. 7, No. 2, 14.01.2022, p. 3054 - 3061.

Research output: Contribution to journalArticleResearchpeer review

Yuan, Y, Cheng, H & Sester, M 2022, 'Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving', IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 3054 - 3061. https://doi.org/10.1109/LRA.2022.3143299
Yuan Y, Cheng H, Sester M. Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving. IEEE Robotics and Automation Letters. 2022 Jan 14;7(2):3054 - 3061. doi: 10.1109/LRA.2022.3143299
Yuan, Yunshuang ; Cheng, Hao ; Sester, Monika. / Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving. In: IEEE Robotics and Automation Letters. 2022 ; Vol. 7, No. 2. pp. 3054 - 3061.
Download
@article{766570d9926d49008fd309292d922d04,
title = "Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving",
abstract = " Sharing collective perception messages (CPM) between vehicles is investigated to decrease occlusions, so as to improve perception accuracy and safety of autonomous driving. However, highly accurate data sharing and low communication overhead is a big challenge for collective perception, especially when real-time communication is required among connected and automated vehicles. In this paper, we propose an efficient and effective keypoints-based deep feature fusion framework, called FPV-RCNN, for collective perception, which is built on top of the 3D object detector PV-RCNN. We introduce a bounding box proposal matching module and a keypoints selection strategy to compress the CPM size and solve the multi-vehicle data fusion problem. Compared to a bird's-eye view (BEV) keypoints feature fusion, FPV-RCNN achieves improved detection accuracy by about 14% at a high evaluation criterion (IoU 0.7) on a synthetic dataset COMAP dedicated to collective perception. Also, its performance is comparable to two raw data fusion baselines that have no data loss in sharing. Moreover, our method also significantly decreases the CPM size to less than 0.3KB, which is about 50 times smaller than the BEV feature map sharing used in previous works. Even with a further decreased number of CPM feature channels, i.e., from 128 to 32, the detection performance only drops about 1%. The code of our method is available at https://github.com/YuanYunshuang/FPV_RCNN. ",
keywords = "Data integration, Feature extraction, Location awareness, Object Detection, Point cloud compression, Proposals, Segmentation and Categorization, Sensor Fusion, Sensor Networks, Three-dimensional displays, Vehicle detection, Sensor fusion, object detection, sensor networks, segmentation and categorization",
author = "Yunshuang Yuan and Hao Cheng and Monika Sester",
note = "Funding Information: This letter was recommended for publication by Associate Editor G. Costante and Editor E. Marchand upon evaluation of the reviewers' comments. This work was supported by the Projects DFG RTC1931 SocialCars and DFG GRK2159 i.c.sens.",
year = "2022",
month = jan,
day = "14",
doi = "10.1109/LRA.2022.3143299",
language = "English",
volume = "7",
pages = "3054 -- 3061",
number = "2",

}

Download

TY - JOUR

T1 - Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving

AU - Yuan, Yunshuang

AU - Cheng, Hao

AU - Sester, Monika

N1 - Funding Information: This letter was recommended for publication by Associate Editor G. Costante and Editor E. Marchand upon evaluation of the reviewers' comments. This work was supported by the Projects DFG RTC1931 SocialCars and DFG GRK2159 i.c.sens.

PY - 2022/1/14

Y1 - 2022/1/14

N2 - Sharing collective perception messages (CPM) between vehicles is investigated to decrease occlusions, so as to improve perception accuracy and safety of autonomous driving. However, highly accurate data sharing and low communication overhead is a big challenge for collective perception, especially when real-time communication is required among connected and automated vehicles. In this paper, we propose an efficient and effective keypoints-based deep feature fusion framework, called FPV-RCNN, for collective perception, which is built on top of the 3D object detector PV-RCNN. We introduce a bounding box proposal matching module and a keypoints selection strategy to compress the CPM size and solve the multi-vehicle data fusion problem. Compared to a bird's-eye view (BEV) keypoints feature fusion, FPV-RCNN achieves improved detection accuracy by about 14% at a high evaluation criterion (IoU 0.7) on a synthetic dataset COMAP dedicated to collective perception. Also, its performance is comparable to two raw data fusion baselines that have no data loss in sharing. Moreover, our method also significantly decreases the CPM size to less than 0.3KB, which is about 50 times smaller than the BEV feature map sharing used in previous works. Even with a further decreased number of CPM feature channels, i.e., from 128 to 32, the detection performance only drops about 1%. The code of our method is available at https://github.com/YuanYunshuang/FPV_RCNN.

AB - Sharing collective perception messages (CPM) between vehicles is investigated to decrease occlusions, so as to improve perception accuracy and safety of autonomous driving. However, highly accurate data sharing and low communication overhead is a big challenge for collective perception, especially when real-time communication is required among connected and automated vehicles. In this paper, we propose an efficient and effective keypoints-based deep feature fusion framework, called FPV-RCNN, for collective perception, which is built on top of the 3D object detector PV-RCNN. We introduce a bounding box proposal matching module and a keypoints selection strategy to compress the CPM size and solve the multi-vehicle data fusion problem. Compared to a bird's-eye view (BEV) keypoints feature fusion, FPV-RCNN achieves improved detection accuracy by about 14% at a high evaluation criterion (IoU 0.7) on a synthetic dataset COMAP dedicated to collective perception. Also, its performance is comparable to two raw data fusion baselines that have no data loss in sharing. Moreover, our method also significantly decreases the CPM size to less than 0.3KB, which is about 50 times smaller than the BEV feature map sharing used in previous works. Even with a further decreased number of CPM feature channels, i.e., from 128 to 32, the detection performance only drops about 1%. The code of our method is available at https://github.com/YuanYunshuang/FPV_RCNN.

KW - Data integration

KW - Feature extraction

KW - Location awareness

KW - Object Detection

KW - Point cloud compression

KW - Proposals

KW - Segmentation and Categorization

KW - Sensor Fusion

KW - Sensor Networks

KW - Three-dimensional displays

KW - Vehicle detection

KW - Sensor fusion

KW - object detection

KW - sensor networks

KW - segmentation and categorization

UR - http://www.scopus.com/inward/record.url?scp=85123314530&partnerID=8YFLogxK

U2 - 10.1109/LRA.2022.3143299

DO - 10.1109/LRA.2022.3143299

M3 - Article

VL - 7

SP - 3054

EP - 3061

JO - IEEE Robotics and Automation Letters

JF - IEEE Robotics and Automation Letters

SN - 2377-3766

IS - 2

ER -

By the same author(s)