Structuring autoencoders

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Authors

View graph of relations

Details

Original languageEnglish
Title of host publication2019 International Conference on Computer Vision Workshop, ICCVW 2019
Subtitle of host publicationProceedings
PublisherInstitute of Electrical and Electronics Engineers Inc.
Pages615-623
Number of pages9
ISBN (electronic)978-1-7281-5023-9
ISBN (print)978-1-7281-5024-6
Publication statusPublished - Oct 2019
Event2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW) - Seoul, Korea, Republic of
Duration: 27 Oct 201928 Oct 2019

Publication series

NameIEEE International Conference on Computer Vision Workshops
ISSN (Print)2473-9936
ISSN (electronic)2473-9944

Abstract

In this paper we propose Structuring AutoEncoders (SAE). SAEs are neural networks which learn a low dimensional representation of data and are additionally enriched with a desired structure in this low dimensional space. While traditional Autoencoders have proven to structure data naturally they fail to discover semantic structure that is hard to recognize in the raw data. The SAE solves the problem by enhancing a traditional Autoencoder using weak supervision to form a structured latent space. In the experiments we demonstrate, that the structured latent space allows for a much more efficient data representation for further tasks such as classification for sparsely labeled data, an efficient choice of data to label, and morphing between classes. To demonstrate the general applicability of our method, we show experiments on the benchmark image datasets MNIST, Fashion-MNIST, DeepFashion2 and on a dataset of 3D human shapes.

Keywords

    Autoencoder, Subspaces

ASJC Scopus subject areas

Cite this

Structuring autoencoders. / Rudolph, Marco; Wandt, Bastian; Rosenhahn, Bodo.
2019 International Conference on Computer Vision Workshop, ICCVW 2019: Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. p. 615-623 (IEEE International Conference on Computer Vision Workshops).

Research output: Chapter in book/report/conference proceedingConference contributionResearchpeer review

Rudolph, M, Wandt, B & Rosenhahn, B 2019, Structuring autoencoders. in 2019 International Conference on Computer Vision Workshop, ICCVW 2019: Proceedings. IEEE International Conference on Computer Vision Workshops, Institute of Electrical and Electronics Engineers Inc., pp. 615-623, 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW), Seoul, Korea, Republic of, 27 Oct 2019. https://doi.org/10.48550/arXiv.1908.02626, https://doi.org/10.1109/ICCVW.2019.00075
Rudolph, M., Wandt, B., & Rosenhahn, B. (2019). Structuring autoencoders. In 2019 International Conference on Computer Vision Workshop, ICCVW 2019: Proceedings (pp. 615-623). (IEEE International Conference on Computer Vision Workshops). Institute of Electrical and Electronics Engineers Inc.. https://doi.org/10.48550/arXiv.1908.02626, https://doi.org/10.1109/ICCVW.2019.00075
Rudolph M, Wandt B, Rosenhahn B. Structuring autoencoders. In 2019 International Conference on Computer Vision Workshop, ICCVW 2019: Proceedings. Institute of Electrical and Electronics Engineers Inc. 2019. p. 615-623. (IEEE International Conference on Computer Vision Workshops). doi: 10.48550/arXiv.1908.02626, 10.1109/ICCVW.2019.00075
Rudolph, Marco ; Wandt, Bastian ; Rosenhahn, Bodo. / Structuring autoencoders. 2019 International Conference on Computer Vision Workshop, ICCVW 2019: Proceedings. Institute of Electrical and Electronics Engineers Inc., 2019. pp. 615-623 (IEEE International Conference on Computer Vision Workshops).
Download
@inproceedings{b3b1aa34d24849808de59c5d938b470f,
title = "Structuring autoencoders",
abstract = "In this paper we propose Structuring AutoEncoders (SAE). SAEs are neural networks which learn a low dimensional representation of data and are additionally enriched with a desired structure in this low dimensional space. While traditional Autoencoders have proven to structure data naturally they fail to discover semantic structure that is hard to recognize in the raw data. The SAE solves the problem by enhancing a traditional Autoencoder using weak supervision to form a structured latent space. In the experiments we demonstrate, that the structured latent space allows for a much more efficient data representation for further tasks such as classification for sparsely labeled data, an efficient choice of data to label, and morphing between classes. To demonstrate the general applicability of our method, we show experiments on the benchmark image datasets MNIST, Fashion-MNIST, DeepFashion2 and on a dataset of 3D human shapes.",
keywords = "Autoencoder, Subspaces",
author = "Marco Rudolph and Bastian Wandt and Bodo Rosenhahn",
note = "Funding information: This work was funded by the Deutsche Forschungsgemeinschaft (DFG,GermanResearchFoundation) underGermany{\textquoteright}s Excellence Strategy within the Cluster of ExcellencePhoenixD (EXC2122).; 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW), ICCVW ; Conference date: 27-10-2019 Through 28-10-2019",
year = "2019",
month = oct,
doi = "10.48550/arXiv.1908.02626",
language = "English",
isbn = "978-1-7281-5024-6",
series = "IEEE International Conference on Computer Vision Workshops",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
pages = "615--623",
booktitle = "2019 International Conference on Computer Vision Workshop, ICCVW 2019",
address = "United States",

}

Download

TY - GEN

T1 - Structuring autoencoders

AU - Rudolph, Marco

AU - Wandt, Bastian

AU - Rosenhahn, Bodo

N1 - Funding information: This work was funded by the Deutsche Forschungsgemeinschaft (DFG,GermanResearchFoundation) underGermany’s Excellence Strategy within the Cluster of ExcellencePhoenixD (EXC2122).

PY - 2019/10

Y1 - 2019/10

N2 - In this paper we propose Structuring AutoEncoders (SAE). SAEs are neural networks which learn a low dimensional representation of data and are additionally enriched with a desired structure in this low dimensional space. While traditional Autoencoders have proven to structure data naturally they fail to discover semantic structure that is hard to recognize in the raw data. The SAE solves the problem by enhancing a traditional Autoencoder using weak supervision to form a structured latent space. In the experiments we demonstrate, that the structured latent space allows for a much more efficient data representation for further tasks such as classification for sparsely labeled data, an efficient choice of data to label, and morphing between classes. To demonstrate the general applicability of our method, we show experiments on the benchmark image datasets MNIST, Fashion-MNIST, DeepFashion2 and on a dataset of 3D human shapes.

AB - In this paper we propose Structuring AutoEncoders (SAE). SAEs are neural networks which learn a low dimensional representation of data and are additionally enriched with a desired structure in this low dimensional space. While traditional Autoencoders have proven to structure data naturally they fail to discover semantic structure that is hard to recognize in the raw data. The SAE solves the problem by enhancing a traditional Autoencoder using weak supervision to form a structured latent space. In the experiments we demonstrate, that the structured latent space allows for a much more efficient data representation for further tasks such as classification for sparsely labeled data, an efficient choice of data to label, and morphing between classes. To demonstrate the general applicability of our method, we show experiments on the benchmark image datasets MNIST, Fashion-MNIST, DeepFashion2 and on a dataset of 3D human shapes.

KW - Autoencoder

KW - Subspaces

UR - http://www.scopus.com/inward/record.url?scp=85082491587&partnerID=8YFLogxK

U2 - 10.48550/arXiv.1908.02626

DO - 10.48550/arXiv.1908.02626

M3 - Conference contribution

AN - SCOPUS:85082491587

SN - 978-1-7281-5024-6

T3 - IEEE International Conference on Computer Vision Workshops

SP - 615

EP - 623

BT - 2019 International Conference on Computer Vision Workshop, ICCVW 2019

PB - Institute of Electrical and Electronics Engineers Inc.

T2 - 2019 IEEE/CVF 17th International Conference on Computer Vision Workshop (ICCVW)

Y2 - 27 October 2019 through 28 October 2019

ER -

By the same author(s)