Relative Pose Consistency for Semi-Supervised Head Pose Estimation

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des Sammelwerks16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021)
Herausgeber/-innenVitomir Struc, Marija Ivanovska
Seiten1-8
Seitenumfang8
ISBN (elektronisch)978-1-6654-3176-7
PublikationsstatusVeröffentlicht - 2021

Publikationsreihe

NameProceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021

Abstract

Human head pose estimation from images plays a vital role in applications like driver assistance systems and human behavior analysis. Head pose estimation networks are typically trained in a supervised manner. Unfortunately, manual/sensor-based annotations of head poses are prone to errors. A solution is supervised training on synthetic training data generated from 3D face models which can provide an infinite amount of perfect labels. However, computer generated face images only provide an approximation of real-world images which results in a domain gap between training and application domain. To date, domain adaptation is rarely addressed in current work on head pose estimation. In this work we propose relative pose consistency, a semi-supervised learning strategy for head pose estimation based on consistency regularization. It allows simultaneous learning on labeled synthetic data and unlabeled real-world data to overcome the domain gap, while keeping the advantages of synthetic data. Consistency regularization enforces consistent network predictions under random image augmentations. We address pose-preserving and pose-altering augmentations. Naturally, pose-altering augmentations cannot be used on unlabeled data. We therefore propose a strategy to exploit the relative pose introduced by pose-altering augmentations between augmented image pairs. This allows the network to benefit from relative pose labels during training on the unlabeled, real-world images. We evaluate our approach on a widely used benchmark (Biwi Kinect Head Pose) and outperform domain-adaptation SOTA. We are the first to present a consistency regularization framework for head pose estimation. Our experiments show that our approach improves head pose estimation accuracy for real-world images despite using only labels from synthetic images.

ASJC Scopus Sachgebiete

Zitieren

Relative Pose Consistency for Semi-Supervised Head Pose Estimation. / Kuhnke, Felix; Ihler, Sontje; Ostermann, Jörn.
16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021). Hrsg. / Vitomir Struc; Marija Ivanovska. 2021. S. 1-8 (Proceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021).

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Kuhnke, F, Ihler, S & Ostermann, J 2021, Relative Pose Consistency for Semi-Supervised Head Pose Estimation. in V Struc & M Ivanovska (Hrsg.), 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021). Proceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021, S. 1-8. https://doi.org/10.1109/FG52635.2021.9666992
Kuhnke, F., Ihler, S., & Ostermann, J. (2021). Relative Pose Consistency for Semi-Supervised Head Pose Estimation. In V. Struc, & M. Ivanovska (Hrsg.), 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021) (S. 1-8). (Proceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021). https://doi.org/10.1109/FG52635.2021.9666992
Kuhnke F, Ihler S, Ostermann J. Relative Pose Consistency for Semi-Supervised Head Pose Estimation. in Struc V, Ivanovska M, Hrsg., 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021). 2021. S. 1-8. (Proceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021). doi: 10.1109/FG52635.2021.9666992
Kuhnke, Felix ; Ihler, Sontje ; Ostermann, Jörn. / Relative Pose Consistency for Semi-Supervised Head Pose Estimation. 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021). Hrsg. / Vitomir Struc ; Marija Ivanovska. 2021. S. 1-8 (Proceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021).
Download
@inproceedings{9b05dbedf2074c259022de84a0463cca,
title = "Relative Pose Consistency for Semi-Supervised Head Pose Estimation",
abstract = "Human head pose estimation from images plays a vital role in applications like driver assistance systems and human behavior analysis. Head pose estimation networks are typically trained in a supervised manner. Unfortunately, manual/sensor-based annotations of head poses are prone to errors. A solution is supervised training on synthetic training data generated from 3D face models which can provide an infinite amount of perfect labels. However, computer generated face images only provide an approximation of real-world images which results in a domain gap between training and application domain. To date, domain adaptation is rarely addressed in current work on head pose estimation. In this work we propose relative pose consistency, a semi-supervised learning strategy for head pose estimation based on consistency regularization. It allows simultaneous learning on labeled synthetic data and unlabeled real-world data to overcome the domain gap, while keeping the advantages of synthetic data. Consistency regularization enforces consistent network predictions under random image augmentations. We address pose-preserving and pose-altering augmentations. Naturally, pose-altering augmentations cannot be used on unlabeled data. We therefore propose a strategy to exploit the relative pose introduced by pose-altering augmentations between augmented image pairs. This allows the network to benefit from relative pose labels during training on the unlabeled, real-world images. We evaluate our approach on a widely used benchmark (Biwi Kinect Head Pose) and outperform domain-adaptation SOTA. We are the first to present a consistency regularization framework for head pose estimation. Our experiments show that our approach improves head pose estimation accuracy for real-world images despite using only labels from synthetic images.",
author = "Felix Kuhnke and Sontje Ihler and J{\"o}rn Ostermann",
year = "2021",
doi = "10.1109/FG52635.2021.9666992",
language = "English",
isbn = "978-1-6654-3177-4",
series = "Proceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021",
pages = "1--8",
editor = "Vitomir Struc and Marija Ivanovska",
booktitle = "16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021)",

}

Download

TY - GEN

T1 - Relative Pose Consistency for Semi-Supervised Head Pose Estimation

AU - Kuhnke, Felix

AU - Ihler, Sontje

AU - Ostermann, Jörn

PY - 2021

Y1 - 2021

N2 - Human head pose estimation from images plays a vital role in applications like driver assistance systems and human behavior analysis. Head pose estimation networks are typically trained in a supervised manner. Unfortunately, manual/sensor-based annotations of head poses are prone to errors. A solution is supervised training on synthetic training data generated from 3D face models which can provide an infinite amount of perfect labels. However, computer generated face images only provide an approximation of real-world images which results in a domain gap between training and application domain. To date, domain adaptation is rarely addressed in current work on head pose estimation. In this work we propose relative pose consistency, a semi-supervised learning strategy for head pose estimation based on consistency regularization. It allows simultaneous learning on labeled synthetic data and unlabeled real-world data to overcome the domain gap, while keeping the advantages of synthetic data. Consistency regularization enforces consistent network predictions under random image augmentations. We address pose-preserving and pose-altering augmentations. Naturally, pose-altering augmentations cannot be used on unlabeled data. We therefore propose a strategy to exploit the relative pose introduced by pose-altering augmentations between augmented image pairs. This allows the network to benefit from relative pose labels during training on the unlabeled, real-world images. We evaluate our approach on a widely used benchmark (Biwi Kinect Head Pose) and outperform domain-adaptation SOTA. We are the first to present a consistency regularization framework for head pose estimation. Our experiments show that our approach improves head pose estimation accuracy for real-world images despite using only labels from synthetic images.

AB - Human head pose estimation from images plays a vital role in applications like driver assistance systems and human behavior analysis. Head pose estimation networks are typically trained in a supervised manner. Unfortunately, manual/sensor-based annotations of head poses are prone to errors. A solution is supervised training on synthetic training data generated from 3D face models which can provide an infinite amount of perfect labels. However, computer generated face images only provide an approximation of real-world images which results in a domain gap between training and application domain. To date, domain adaptation is rarely addressed in current work on head pose estimation. In this work we propose relative pose consistency, a semi-supervised learning strategy for head pose estimation based on consistency regularization. It allows simultaneous learning on labeled synthetic data and unlabeled real-world data to overcome the domain gap, while keeping the advantages of synthetic data. Consistency regularization enforces consistent network predictions under random image augmentations. We address pose-preserving and pose-altering augmentations. Naturally, pose-altering augmentations cannot be used on unlabeled data. We therefore propose a strategy to exploit the relative pose introduced by pose-altering augmentations between augmented image pairs. This allows the network to benefit from relative pose labels during training on the unlabeled, real-world images. We evaluate our approach on a widely used benchmark (Biwi Kinect Head Pose) and outperform domain-adaptation SOTA. We are the first to present a consistency regularization framework for head pose estimation. Our experiments show that our approach improves head pose estimation accuracy for real-world images despite using only labels from synthetic images.

UR - http://www.scopus.com/inward/record.url?scp=85125094373&partnerID=8YFLogxK

U2 - 10.1109/FG52635.2021.9666992

DO - 10.1109/FG52635.2021.9666992

M3 - Conference contribution

SN - 978-1-6654-3177-4

T3 - Proceedings - 2021 16th IEEE International Conference on Automatic Face and Gesture Recognition, FG 2021

SP - 1

EP - 8

BT - 16th IEEE International Conference on Automatic Face and Gesture Recognition (FG 2021)

A2 - Struc, Vitomir

A2 - Ivanovska, Marija

ER -

Von denselben Autoren