Learning disentangled representations via independent subspaces

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des Sammelwerks2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)
UntertitelProceedings
Herausgeber (Verlag)Institute of Electrical and Electronics Engineers Inc.
Seiten560-568
Seitenumfang9
ISBN (elektronisch)978-1-7281-5023-9
ISBN (Print)978-1-7281-5024-6
PublikationsstatusVeröffentlicht - Okt. 2019
Veranstaltung2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW) - Seoul, Südkorea
Dauer: 27 Okt. 201928 Okt. 2019

Publikationsreihe

NameIEEE International Conference on Computer Vision Workshops
ISSN (Print)2473-9936
ISSN (elektronisch)2473-9944

Abstract

Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.

ASJC Scopus Sachgebiete

Zitieren

Learning disentangled representations via independent subspaces. / Awiszus, Maren; Ackermann, Hanno; Rosenhahn, Bodo.
2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. S. 560-568 9022161 (IEEE International Conference on Computer Vision Workshops).

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Awiszus, M, Ackermann, H & Rosenhahn, B 2019, Learning disentangled representations via independent subspaces. in 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings., 9022161, IEEE International Conference on Computer Vision Workshops, Institute of Electrical and Electronics Engineers Inc., S. 560-568, 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW), Seoul, Südkorea, 27 Okt. 2019. https://doi.org/10.48550/arXiv.1908.08989, https://doi.org/10.1109/ICCVW.2019.00069
Awiszus, M., Ackermann, H., & Rosenhahn, B. (2019). Learning disentangled representations via independent subspaces. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings (S. 560-568). Artikel 9022161 (IEEE International Conference on Computer Vision Workshops). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.48550/arXiv.1908.08989, https://doi.org/10.1109/ICCVW.2019.00069
Awiszus M, Ackermann H, Rosenhahn B. Learning disentangled representations via independent subspaces. in 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings. Institute of Electrical and Electronics Engineers Inc. 2019. S. 560-568. 9022161. (IEEE International Conference on Computer Vision Workshops). doi: 10.48550/arXiv.1908.08989, 10.1109/ICCVW.2019.00069
Awiszus, Maren ; Ackermann, Hanno ; Rosenhahn, Bodo. / Learning disentangled representations via independent subspaces. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. S. 560-568 (IEEE International Conference on Computer Vision Workshops).
Download
@inproceedings{b911c5db935047d183b12efbd116a0b2,
title = "Learning disentangled representations via independent subspaces",
abstract = "Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.",
keywords = "Autoencoders, Face image editing, Independent subspace analysis, Latent space editing, Machine learning",
author = "Maren Awiszus and Hanno Ackermann and Bodo Rosenhahn",
note = "Funding information: The work is inspired by BIAS (”Bias and Discrimination in Big Data and Algorithmic Processing. Philosophical Assessments, Legal Dimensions, and Technical Solutions”), a project funded by the Volkswagen Foundation within the initiative ”AI and the Society of the Future” for which the last author is a Principal Investigator.; 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW), ICCVW ; Conference date: 27-10-2019 Through 28-10-2019",
year = "2019",
month = oct,
doi = "10.48550/arXiv.1908.08989",
language = "English",
isbn = "978-1-7281-5024-6",
series = "IEEE International Conference on Computer Vision Workshops",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "560--568",
booktitle = "2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)",
address = "United States",

}

Download

TY - GEN

T1 - Learning disentangled representations via independent subspaces

AU - Awiszus, Maren

AU - Ackermann, Hanno

AU - Rosenhahn, Bodo

N1 - Funding information: The work is inspired by BIAS (”Bias and Discrimination in Big Data and Algorithmic Processing. Philosophical Assessments, Legal Dimensions, and Technical Solutions”), a project funded by the Volkswagen Foundation within the initiative ”AI and the Society of the Future” for which the last author is a Principal Investigator.

PY - 2019/10

Y1 - 2019/10

N2 - Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.

AB - Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.

KW - Autoencoders

KW - Face image editing

KW - Independent subspace analysis

KW - Latent space editing

KW - Machine learning

UR - http://www.scopus.com/inward/record.url?scp=85082453249&partnerID=8YFLogxK

U2 - 10.48550/arXiv.1908.08989

DO - 10.48550/arXiv.1908.08989

M3 - Conference contribution

AN - SCOPUS:85082453249

SN - 978-1-7281-5024-6

T3 - IEEE International Conference on Computer Vision Workshops

SP - 560

EP - 568

BT - 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)

PB - Institute of Electrical and Electronics Engineers Inc.

T2 - 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW)

Y2 - 27 October 2019 through 28 October 2019

ER -

Von denselben Autoren