Loading [MathJax]/extensions/tex2jax.js

Learning disentangled representations via independent subspaces

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

Details

Original languageEnglish
Title of host publication2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)
Subtitle of host publicationProceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages560-568
Number of pages9
ISBN (electronic)978-1-7281-5023-9
ISBN (print)978-1-7281-5024-6
Publication statusPublished - Oct 2019
Event2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW) - Seoul, Korea, Republic of
Duration: 27 Oct 201928 Oct 2019

Publication series

NameIEEE International Conference on Computer Vision Workshops
ISSN (Print)2473-9936
ISSN (electronic)2473-9944

Abstract

Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.

Keywords

    Autoencoders, Face image editing, Independent subspace analysis, Latent space editing, Machine learning

ASJC Scopus subject areas

Cite this

Learning disentangled representations via independent subspaces. / Awiszus, Maren; Ackermann, Hanno; Rosenhahn, Bodo.
2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. p. 560-568 9022161 (IEEE International Conference on Computer Vision Workshops).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Awiszus, M, Ackermann, H & Rosenhahn, B 2019, Learning disentangled representations via independent subspaces. in 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings., 9022161, IEEE International Conference on Computer Vision Workshops, Institute of Electrical and Electronics Engineers Inc., pp. 560-568, 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, Republic of, 27 Oct 2019. https://doi.org/10.48550/arXiv.1908.08989, https://doi.org/10.1109/ICCVW.2019.00069
Awiszus, M., Ackermann, H., & Rosenhahn, B. (2019). Learning disentangled representations via independent subspaces. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings (pp. 560-568). Article 9022161 (IEEE International Conference on Computer Vision Workshops). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.48550/arXiv.1908.08989, https://doi.org/10.1109/ICCVW.2019.00069
Awiszus M, Ackermann H, Rosenhahn B. Learning disentangled representations via independent subspaces. In 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings. Institute of Electrical and Electronics Engineers Inc. 2019. p. 560-568. 9022161. (IEEE International Conference on Computer Vision Workshops). doi: 10.48550/arXiv.1908.08989, 10.1109/ICCVW.2019.00069
Awiszus, Maren ; Ackermann, Hanno ; Rosenhahn, Bodo. / Learning disentangled representations via independent subspaces. 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW): Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. pp. 560-568 (IEEE International Conference on Computer Vision Workshops).
Download
@inproceedings{b911c5db935047d183b12efbd116a0b2,
title = "Learning disentangled representations via independent subspaces",
abstract = "Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.",
keywords = "Autoencoders, Face image editing, Independent subspace analysis, Latent space editing, Machine learning",
author = "Maren Awiszus and Hanno Ackermann and Bodo Rosenhahn",
note = "Funding information: The work is inspired by BIAS (”Bias and Discrimination in Big Data and Algorithmic Processing. Philosophical Assessments, Legal Dimensions, and Technical Solutions”), a project funded by the Volkswagen Foundation within the initiative ”AI and the Society of the Future” for which the last author is a Principal Investigator.; 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW), ICCVW ; Conference date: 27-10-2019 Through 28-10-2019",
year = "2019",
month = oct,
doi = "10.48550/arXiv.1908.08989",
language = "English",
isbn = "978-1-7281-5024-6",
series = "IEEE International Conference on Computer Vision Workshops",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "560--568",
booktitle = "2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)",
address = "United States",

}

Download

TY - GEN

T1 - Learning disentangled representations via independent subspaces

AU - Awiszus, Maren

AU - Ackermann, Hanno

AU - Rosenhahn, Bodo

N1 - Funding information: The work is inspired by BIAS (”Bias and Discrimination in Big Data and Algorithmic Processing. Philosophical Assessments, Legal Dimensions, and Technical Solutions”), a project funded by the Volkswagen Foundation within the initiative ”AI and the Society of the Future” for which the last author is a Principal Investigator.

PY - 2019/10

Y1 - 2019/10

N2 - Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.

AB - Image generating neural networks are mostly viewed as black boxes, where any change in the input can have a number of globally effective changes on the output. In this work, we propose a method for learning disentangled representations to allow for localized image manipulations. We use face images as our example of choice. Depending on the image region, identity and other facial attributes can be modified. The proposed network can transfer parts of a face such as shape and color of eyes, hair, mouth, etc.directly between persons while all other parts of the face remain unchanged. The network allows to generate modified images which appear like realistic images. Our model learns disentangled representations by weak supervision. We propose a localized resnet autoencoder optimized using several loss functions including a loss based on the semantic segmentation, which we interpret as masks, and a loss which enforces disentanglement by decomposition of the latent space into statistically independent subspaces. We evaluate the proposed solution w.r.t. disentanglement and generated image quality. Convincing results are demonstrated using the CelebA dataset.

KW - Autoencoders

KW - Face image editing

KW - Independent subspace analysis

KW - Latent space editing

KW - Machine learning

UR - http://www.scopus.com/inward/record.url?scp=85082453249&partnerID=8YFLogxK

U2 - 10.48550/arXiv.1908.08989

DO - 10.48550/arXiv.1908.08989

M3 - Conference contribution

AN - SCOPUS:85082453249

SN - 978-1-7281-5024-6

T3 - IEEE International Conference on Computer Vision Workshops

SP - 560

EP - 568

BT - 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW)

PB - Institute of Electrical and Electronics Engineers Inc.

T2 - 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW)

Y2 - 27 October 2019 through 28 October 2019

ER -

By the same author(s)