Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Autoren

  • Matthias Springstein
  • Eric Müller-Budack
  • Ralph Ewerth

Organisationseinheiten

Externe Organisationen

  • Technische Informationsbibliothek (TIB) Leibniz-Informationszentrum Technik und Naturwissenschaften und Universitätsbibliothek
Forschungs-netzwerk anzeigen

Details

OriginalspracheEnglisch
Titel des SammelwerksMMPT 2021
UntertitelProceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding
Seiten46-54
Seitenumfang9
ISBN (elektronisch)9781450385305
PublikationsstatusVeröffentlicht - 27 Aug. 2021
Veranstaltung1st International Joint Workshop on Multi-Modal Pre-Training for Multimedia Understanding, MMPT 2021 - Taipei, Taiwan
Dauer: 21 Aug. 2021 → …

Publikationsreihe

NameMMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding

Abstract

The recognition of handwritten mathematical expressions in images and video frames is a difficult and unsolved problem yet. Deep convectional neural networks are basically a promising approach, but typically require a large amount of labeled training data. However, such a large training dataset does not exist for the task of handwritten formula recognition. In this paper, we introduce a system that creates a large set of synthesized training examples of mathematical expressions which are derived from LaTeX documents. For this purpose, we propose a novel attention-based generative adversarial network to translate rendered equations to handwritten formulas. The datasets generated by this approach contain hundreds of thousands of formulas, making it ideal for pretraining or the design of more complex models. We evaluate our synthesized dataset and the recognition approach on the CROHME 2014 benchmark dataset. Experimental results demonstrate the feasibility of the approach.

ASJC Scopus Sachgebiete

Zitieren

Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention. / Springstein, Matthias; Müller-Budack, Eric; Ewerth, Ralph.
MMPT 2021 : Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding. 2021. S. 46-54 (MMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding).

Publikation: Beitrag in Buch/Bericht/Sammelwerk/KonferenzbandAufsatz in KonferenzbandForschungPeer-Review

Springstein, M, Müller-Budack, E & Ewerth, R 2021, Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention. in MMPT 2021 : Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding. MMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding, S. 46-54, 1st International Joint Workshop on Multi-Modal Pre-Training for Multimedia Understanding, MMPT 2021, Taipei, Taiwan, 21 Aug. 2021. https://doi.org/10.48550/arXiv.2106.09432, https://doi.org/10.1145/3463945.3469059
Springstein, M., Müller-Budack, E., & Ewerth, R. (2021). Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention. In MMPT 2021 : Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding (S. 46-54). (MMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding). https://doi.org/10.48550/arXiv.2106.09432, https://doi.org/10.1145/3463945.3469059
Springstein M, Müller-Budack E, Ewerth R. Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention. in MMPT 2021 : Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding. 2021. S. 46-54. (MMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding). doi: https://doi.org/10.48550/arXiv.2106.09432, 10.1145/3463945.3469059
Springstein, Matthias ; Müller-Budack, Eric ; Ewerth, Ralph. / Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention. MMPT 2021 : Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding. 2021. S. 46-54 (MMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding).
Download
@inproceedings{92e79591b6934694992747f0ebe587c6,
title = "Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention",
abstract = "The recognition of handwritten mathematical expressions in images and video frames is a difficult and unsolved problem yet. Deep convectional neural networks are basically a promising approach, but typically require a large amount of labeled training data. However, such a large training dataset does not exist for the task of handwritten formula recognition. In this paper, we introduce a system that creates a large set of synthesized training examples of mathematical expressions which are derived from LaTeX documents. For this purpose, we propose a novel attention-based generative adversarial network to translate rendered equations to handwritten formulas. The datasets generated by this approach contain hundreds of thousands of formulas, making it ideal for pretraining or the design of more complex models. We evaluate our synthesized dataset and the recognition approach on the CROHME 2014 benchmark dataset. Experimental results demonstrate the feasibility of the approach.",
keywords = "datasets, formula recognition, generative adversarial network",
author = "Matthias Springstein and Eric M{\"u}ller-Budack and Ralph Ewerth",
year = "2021",
month = aug,
day = "27",
doi = "https://doi.org/10.48550/arXiv.2106.09432",
language = "English",
series = "MMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding",
pages = "46--54",
booktitle = "MMPT 2021",
note = "1st International Joint Workshop on Multi-Modal Pre-Training for Multimedia Understanding, MMPT 2021 ; Conference date: 21-08-2021",

}

Download

TY - GEN

T1 - Unsupervised Training Data Generation of Handwritten Formulas using Generative Adversarial Networks with Self-Attention

AU - Springstein, Matthias

AU - Müller-Budack, Eric

AU - Ewerth, Ralph

PY - 2021/8/27

Y1 - 2021/8/27

N2 - The recognition of handwritten mathematical expressions in images and video frames is a difficult and unsolved problem yet. Deep convectional neural networks are basically a promising approach, but typically require a large amount of labeled training data. However, such a large training dataset does not exist for the task of handwritten formula recognition. In this paper, we introduce a system that creates a large set of synthesized training examples of mathematical expressions which are derived from LaTeX documents. For this purpose, we propose a novel attention-based generative adversarial network to translate rendered equations to handwritten formulas. The datasets generated by this approach contain hundreds of thousands of formulas, making it ideal for pretraining or the design of more complex models. We evaluate our synthesized dataset and the recognition approach on the CROHME 2014 benchmark dataset. Experimental results demonstrate the feasibility of the approach.

AB - The recognition of handwritten mathematical expressions in images and video frames is a difficult and unsolved problem yet. Deep convectional neural networks are basically a promising approach, but typically require a large amount of labeled training data. However, such a large training dataset does not exist for the task of handwritten formula recognition. In this paper, we introduce a system that creates a large set of synthesized training examples of mathematical expressions which are derived from LaTeX documents. For this purpose, we propose a novel attention-based generative adversarial network to translate rendered equations to handwritten formulas. The datasets generated by this approach contain hundreds of thousands of formulas, making it ideal for pretraining or the design of more complex models. We evaluate our synthesized dataset and the recognition approach on the CROHME 2014 benchmark dataset. Experimental results demonstrate the feasibility of the approach.

KW - datasets

KW - formula recognition

KW - generative adversarial network

UR - http://www.scopus.com/inward/record.url?scp=85114808291&partnerID=8YFLogxK

U2 - https://doi.org/10.48550/arXiv.2106.09432

DO - https://doi.org/10.48550/arXiv.2106.09432

M3 - Conference contribution

AN - SCOPUS:85114808291

T3 - MMPT 2021 - Proceedings of the 2021 Workshop on Multi-Modal Pre-Training for Multimedia Understanding

SP - 46

EP - 54

BT - MMPT 2021

T2 - 1st International Joint Workshop on Multi-Modal Pre-Training for Multimedia Understanding, MMPT 2021

Y2 - 21 August 2021

ER -