Personalized 3D Human Pose and Shape Refinement

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

Externe Organisationen

  • Epic Games, Inc.
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des Sammelwerks2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW)
Herausgeber (Verlag)Institute of Electrical and Electronics Engineers Inc.
Seiten4191-4201
Seitenumfang11
ISBN (elektronisch)9798350307443
ISBN (Print)9798350307450
PublikationsstatusVeröffentlicht - 2023
Veranstaltung2023 IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2023 - Paris, Frankreich
Dauer: 2 Okt. 20236 Okt. 2023

Abstract

Recently, regression-based methods have dominated the field of 3D human pose and shape estimation. Despite their promising results, a common issue is the misalignment between predictions and image observations, often caused by minor joint rotation errors that accumulate along the kinematic chain. To address this issue, we propose to construct dense correspondences between initial human model estimates and the corresponding images that can be used to refine the initial predictions. To this end, we utilize renderings of the 3D models to predict per-pixel 2D displacements between the synthetic renderings and the RGB images. This allows us to effectively integrate and exploit appearance information of the persons. Our per-pixel displacements can be efficiently transformed to per-visible-vertex displacements and then used for 3D model refinement by minimizing a reprojection loss. To demonstrate the effectiveness of our approach, we refine the initial 3D human mesh predictions of multiple models using different refinement procedures on 3DPW and RICH. We show that our approach not only consistently leads to better image-model alignment, but also to improved 3D accuracy.

ASJC Scopus Sachgebiete

Zitieren

Personalized 3D Human Pose and Shape Refinement. / Wehrbein, Tom; Rosenhahn, Bodo; Matthews, Iain et al.
2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW). Institute of Electrical and Electronics Engineers Inc., 2023. S. 4191-4201.

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Wehrbein, T, Rosenhahn, B, Matthews, I & Stoll, C 2023, Personalized 3D Human Pose and Shape Refinement. in 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW). Institute of Electrical and Electronics Engineers Inc., S. 4191-4201, 2023 IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2023, Paris, Frankreich, 2 Okt. 2023. https://doi.org/10.1109/ICCVW60793.2023.00453
Wehrbein, T., Rosenhahn, B., Matthews, I., & Stoll, C. (2023). Personalized 3D Human Pose and Shape Refinement. In 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW) (S. 4191-4201). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.1109/ICCVW60793.2023.00453
Wehrbein T, Rosenhahn B, Matthews I, Stoll C. Personalized 3D Human Pose and Shape Refinement. in 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW). Institute of Electrical and Electronics Engineers Inc. 2023. S. 4191-4201 doi: 10.1109/ICCVW60793.2023.00453
Wehrbein, Tom ; Rosenhahn, Bodo ; Matthews, Iain et al. / Personalized 3D Human Pose and Shape Refinement. 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW). Institute of Electrical and Electronics Engineers Inc., 2023. S. 4191-4201
Download
@inproceedings{4626c97728984b5f95465b434e6cd349,
title = "Personalized 3D Human Pose and Shape Refinement",
abstract = "Recently, regression-based methods have dominated the field of 3D human pose and shape estimation. Despite their promising results, a common issue is the misalignment between predictions and image observations, often caused by minor joint rotation errors that accumulate along the kinematic chain. To address this issue, we propose to construct dense correspondences between initial human model estimates and the corresponding images that can be used to refine the initial predictions. To this end, we utilize renderings of the 3D models to predict per-pixel 2D displacements between the synthetic renderings and the RGB images. This allows us to effectively integrate and exploit appearance information of the persons. Our per-pixel displacements can be efficiently transformed to per-visible-vertex displacements and then used for 3D model refinement by minimizing a reprojection loss. To demonstrate the effectiveness of our approach, we refine the initial 3D human mesh predictions of multiple models using different refinement procedures on 3DPW and RICH. We show that our approach not only consistently leads to better image-model alignment, but also to improved 3D accuracy.",
author = "Tom Wehrbein and Bodo Rosenhahn and Iain Matthews and Carsten Stoll",
note = "Funding Information: This work was supported by the Federal Ministry of Education and Research (BMBF), Germany, under the project LeibnizKILabor (grant no. 01DD20003) and the AI service center KISSKI (grant no. 01IS22093C), the Center for Digital Innovations (ZDIN) and the Deutsche Forschungsgemeinschaft (DFG) under Germany's Excellence Strategy within the Cluster of Excellence PhoenixD (EXC 2122). ; 2023 IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2023 ; Conference date: 02-10-2023 Through 06-10-2023",
year = "2023",
doi = "10.1109/ICCVW60793.2023.00453",
language = "English",
isbn = "9798350307450",
pages = "4191--4201",
booktitle = "2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW)",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
address = "United States",

}

Download

TY - GEN

T1 - Personalized 3D Human Pose and Shape Refinement

AU - Wehrbein, Tom

AU - Rosenhahn, Bodo

AU - Matthews, Iain

AU - Stoll, Carsten

N1 - Funding Information: This work was supported by the Federal Ministry of Education and Research (BMBF), Germany, under the project LeibnizKILabor (grant no. 01DD20003) and the AI service center KISSKI (grant no. 01IS22093C), the Center for Digital Innovations (ZDIN) and the Deutsche Forschungsgemeinschaft (DFG) under Germany's Excellence Strategy within the Cluster of Excellence PhoenixD (EXC 2122).

PY - 2023

Y1 - 2023

N2 - Recently, regression-based methods have dominated the field of 3D human pose and shape estimation. Despite their promising results, a common issue is the misalignment between predictions and image observations, often caused by minor joint rotation errors that accumulate along the kinematic chain. To address this issue, we propose to construct dense correspondences between initial human model estimates and the corresponding images that can be used to refine the initial predictions. To this end, we utilize renderings of the 3D models to predict per-pixel 2D displacements between the synthetic renderings and the RGB images. This allows us to effectively integrate and exploit appearance information of the persons. Our per-pixel displacements can be efficiently transformed to per-visible-vertex displacements and then used for 3D model refinement by minimizing a reprojection loss. To demonstrate the effectiveness of our approach, we refine the initial 3D human mesh predictions of multiple models using different refinement procedures on 3DPW and RICH. We show that our approach not only consistently leads to better image-model alignment, but also to improved 3D accuracy.

AB - Recently, regression-based methods have dominated the field of 3D human pose and shape estimation. Despite their promising results, a common issue is the misalignment between predictions and image observations, often caused by minor joint rotation errors that accumulate along the kinematic chain. To address this issue, we propose to construct dense correspondences between initial human model estimates and the corresponding images that can be used to refine the initial predictions. To this end, we utilize renderings of the 3D models to predict per-pixel 2D displacements between the synthetic renderings and the RGB images. This allows us to effectively integrate and exploit appearance information of the persons. Our per-pixel displacements can be efficiently transformed to per-visible-vertex displacements and then used for 3D model refinement by minimizing a reprojection loss. To demonstrate the effectiveness of our approach, we refine the initial 3D human mesh predictions of multiple models using different refinement procedures on 3DPW and RICH. We show that our approach not only consistently leads to better image-model alignment, but also to improved 3D accuracy.

UR - http://www.scopus.com/inward/record.url?scp=85182935057&partnerID=8YFLogxK

U2 - 10.1109/ICCVW60793.2023.00453

DO - 10.1109/ICCVW60793.2023.00453

M3 - Conference contribution

AN - SCOPUS:85182935057

SN - 9798350307450

SP - 4191

EP - 4201

BT - 2023 IEEE/CVF International Conference on Computer Vision Workshops (ICCVW)

PB - Institute of Electrical and Electronics Engineers Inc.

T2 - 2023 IEEE/CVF International Conference on Computer Vision Workshops, ICCVW 2023

Y2 - 2 October 2023 through 6 October 2023

ER -

Von denselben Autoren