Using Semantically Paired Images to Improve Domain Adaptation for the Semantic Segmentation of Aerial Images

Research output: Contribution to journalConference articleResearchpeer review

Authors

Research Organisations

View graph of relations

Details

Original languageEnglish
Pages (from-to)483-492
Number of pages10
JournalISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
Publication statusPublished - 3 Aug 2020
Event2020 24th ISPRS Congress on Technical Commission II - Nice, Virtual, France
Duration: 31 Aug 20202 Sept 2020

Abstract

Modern machine learning, especially deep learning, which is used in a variety of applications, requires a lot of labelled data for model training. Having an insufficient amount of training examples leads to models which do not generalize well to new input instances. This is a particular significant problem for tasks involving aerial images: Often training data is only available for a limited geographical area and a narrow time window, thus leading to models which perform poorly in different regions, at different times of day, or during different seasons. Domain adaptation can mitigate this issue by using labelled source domain training examples and unlabeled target domain images to train a model which performs well on both domains. Modern adversarial domain adaptation approaches use unpaired data. We propose using pairs of semantically similar images, i.e., whose segmentations are accurate predictions of each other, for improved model performance. In this paper we show that, as an upper limit based on ground truth, using semantically paired aerial images during training almost always increases model performance with an average improvement of 4.2% accuracy and .036 mean intersection-over-union (mIoU). Using a practical estimate of semantic similarity, we still achieve improvements in more than half of all cases, with average improvements of 2.5% accuracy and .017 mIoU in those cases.

Keywords

    Aerial Images, Deep Learning, Domain Adaptation, Neural Networks, Semantic Segmentation, Transfer Learning

ASJC Scopus subject areas

Cite this

Using Semantically Paired Images to Improve Domain Adaptation for the Semantic Segmentation of Aerial Images. / Gritzner, D.; Ostermann, J.
In: ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, 03.08.2020, p. 483-492.

Research output: Contribution to journalConference articleResearchpeer review

Gritzner D, Ostermann J. Using Semantically Paired Images to Improve Domain Adaptation for the Semantic Segmentation of Aerial Images. ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences. 2020 Aug 3;483-492. doi: 10.5194/isprs-annals-V-2-2020-483-2020, 10.15488/10877
Gritzner, D. ; Ostermann, J. / Using Semantically Paired Images to Improve Domain Adaptation for the Semantic Segmentation of Aerial Images. In: ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences. 2020 ; pp. 483-492.
Download
@article{01b7a28f573447febb3593c253907b6b,
title = "Using Semantically Paired Images to Improve Domain Adaptation for the Semantic Segmentation of Aerial Images",
abstract = "Modern machine learning, especially deep learning, which is used in a variety of applications, requires a lot of labelled data for model training. Having an insufficient amount of training examples leads to models which do not generalize well to new input instances. This is a particular significant problem for tasks involving aerial images: Often training data is only available for a limited geographical area and a narrow time window, thus leading to models which perform poorly in different regions, at different times of day, or during different seasons. Domain adaptation can mitigate this issue by using labelled source domain training examples and unlabeled target domain images to train a model which performs well on both domains. Modern adversarial domain adaptation approaches use unpaired data. We propose using pairs of semantically similar images, i.e., whose segmentations are accurate predictions of each other, for improved model performance. In this paper we show that, as an upper limit based on ground truth, using semantically paired aerial images during training almost always increases model performance with an average improvement of 4.2% accuracy and .036 mean intersection-over-union (mIoU). Using a practical estimate of semantic similarity, we still achieve improvements in more than half of all cases, with average improvements of 2.5% accuracy and .017 mIoU in those cases.",
keywords = "Aerial Images, Deep Learning, Domain Adaptation, Neural Networks, Semantic Segmentation, Transfer Learning",
author = "D. Gritzner and J. Ostermann",
year = "2020",
month = aug,
day = "3",
doi = "10.5194/isprs-annals-V-2-2020-483-2020",
language = "English",
pages = "483--492",
note = "2020 24th ISPRS Congress on Technical Commission II ; Conference date: 31-08-2020 Through 02-09-2020",

}

Download

TY - JOUR

T1 - Using Semantically Paired Images to Improve Domain Adaptation for the Semantic Segmentation of Aerial Images

AU - Gritzner, D.

AU - Ostermann, J.

PY - 2020/8/3

Y1 - 2020/8/3

N2 - Modern machine learning, especially deep learning, which is used in a variety of applications, requires a lot of labelled data for model training. Having an insufficient amount of training examples leads to models which do not generalize well to new input instances. This is a particular significant problem for tasks involving aerial images: Often training data is only available for a limited geographical area and a narrow time window, thus leading to models which perform poorly in different regions, at different times of day, or during different seasons. Domain adaptation can mitigate this issue by using labelled source domain training examples and unlabeled target domain images to train a model which performs well on both domains. Modern adversarial domain adaptation approaches use unpaired data. We propose using pairs of semantically similar images, i.e., whose segmentations are accurate predictions of each other, for improved model performance. In this paper we show that, as an upper limit based on ground truth, using semantically paired aerial images during training almost always increases model performance with an average improvement of 4.2% accuracy and .036 mean intersection-over-union (mIoU). Using a practical estimate of semantic similarity, we still achieve improvements in more than half of all cases, with average improvements of 2.5% accuracy and .017 mIoU in those cases.

AB - Modern machine learning, especially deep learning, which is used in a variety of applications, requires a lot of labelled data for model training. Having an insufficient amount of training examples leads to models which do not generalize well to new input instances. This is a particular significant problem for tasks involving aerial images: Often training data is only available for a limited geographical area and a narrow time window, thus leading to models which perform poorly in different regions, at different times of day, or during different seasons. Domain adaptation can mitigate this issue by using labelled source domain training examples and unlabeled target domain images to train a model which performs well on both domains. Modern adversarial domain adaptation approaches use unpaired data. We propose using pairs of semantically similar images, i.e., whose segmentations are accurate predictions of each other, for improved model performance. In this paper we show that, as an upper limit based on ground truth, using semantically paired aerial images during training almost always increases model performance with an average improvement of 4.2% accuracy and .036 mean intersection-over-union (mIoU). Using a practical estimate of semantic similarity, we still achieve improvements in more than half of all cases, with average improvements of 2.5% accuracy and .017 mIoU in those cases.

KW - Aerial Images

KW - Deep Learning

KW - Domain Adaptation

KW - Neural Networks

KW - Semantic Segmentation

KW - Transfer Learning

UR - http://www.scopus.com/inward/record.url?scp=85091065244&partnerID=8YFLogxK

U2 - 10.5194/isprs-annals-V-2-2020-483-2020

DO - 10.5194/isprs-annals-V-2-2020-483-2020

M3 - Conference article

AN - SCOPUS:85091065244

SP - 483

EP - 492

JO - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences

JF - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences

SN - 2194-9042

T2 - 2020 24th ISPRS Congress on Technical Commission II

Y2 - 31 August 2020 through 2 September 2020

ER -

By the same author(s)