Details
Original language | English |
---|---|
Pages (from-to) | 483-492 |
Number of pages | 10 |
Journal | ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences |
Publication status | Published - 3 Aug 2020 |
Event | 2020 24th ISPRS Congress on Technical Commission II - Nice, Virtual, France Duration: 31 Aug 2020 → 2 Sept 2020 |
Abstract
Modern machine learning, especially deep learning, which is used in a variety of applications, requires a lot of labelled data for model training. Having an insufficient amount of training examples leads to models which do not generalize well to new input instances. This is a particular significant problem for tasks involving aerial images: Often training data is only available for a limited geographical area and a narrow time window, thus leading to models which perform poorly in different regions, at different times of day, or during different seasons. Domain adaptation can mitigate this issue by using labelled source domain training examples and unlabeled target domain images to train a model which performs well on both domains. Modern adversarial domain adaptation approaches use unpaired data. We propose using pairs of semantically similar images, i.e., whose segmentations are accurate predictions of each other, for improved model performance. In this paper we show that, as an upper limit based on ground truth, using semantically paired aerial images during training almost always increases model performance with an average improvement of 4.2% accuracy and .036 mean intersection-over-union (mIoU). Using a practical estimate of semantic similarity, we still achieve improvements in more than half of all cases, with average improvements of 2.5% accuracy and .017 mIoU in those cases.
Keywords
- Aerial Images, Deep Learning, Domain Adaptation, Neural Networks, Semantic Segmentation, Transfer Learning
ASJC Scopus subject areas
- Earth and Planetary Sciences(all)
- Earth and Planetary Sciences (miscellaneous)
- Environmental Science(all)
- Environmental Science (miscellaneous)
- Physics and Astronomy(all)
- Instrumentation
Cite this
- Standard
- Harvard
- Apa
- Vancouver
- BibTeX
- RIS
In: ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences, 03.08.2020, p. 483-492.
Research output: Contribution to journal › Conference article › Research › peer review
}
TY - JOUR
T1 - Using Semantically Paired Images to Improve Domain Adaptation for the Semantic Segmentation of Aerial Images
AU - Gritzner, D.
AU - Ostermann, J.
PY - 2020/8/3
Y1 - 2020/8/3
N2 - Modern machine learning, especially deep learning, which is used in a variety of applications, requires a lot of labelled data for model training. Having an insufficient amount of training examples leads to models which do not generalize well to new input instances. This is a particular significant problem for tasks involving aerial images: Often training data is only available for a limited geographical area and a narrow time window, thus leading to models which perform poorly in different regions, at different times of day, or during different seasons. Domain adaptation can mitigate this issue by using labelled source domain training examples and unlabeled target domain images to train a model which performs well on both domains. Modern adversarial domain adaptation approaches use unpaired data. We propose using pairs of semantically similar images, i.e., whose segmentations are accurate predictions of each other, for improved model performance. In this paper we show that, as an upper limit based on ground truth, using semantically paired aerial images during training almost always increases model performance with an average improvement of 4.2% accuracy and .036 mean intersection-over-union (mIoU). Using a practical estimate of semantic similarity, we still achieve improvements in more than half of all cases, with average improvements of 2.5% accuracy and .017 mIoU in those cases.
AB - Modern machine learning, especially deep learning, which is used in a variety of applications, requires a lot of labelled data for model training. Having an insufficient amount of training examples leads to models which do not generalize well to new input instances. This is a particular significant problem for tasks involving aerial images: Often training data is only available for a limited geographical area and a narrow time window, thus leading to models which perform poorly in different regions, at different times of day, or during different seasons. Domain adaptation can mitigate this issue by using labelled source domain training examples and unlabeled target domain images to train a model which performs well on both domains. Modern adversarial domain adaptation approaches use unpaired data. We propose using pairs of semantically similar images, i.e., whose segmentations are accurate predictions of each other, for improved model performance. In this paper we show that, as an upper limit based on ground truth, using semantically paired aerial images during training almost always increases model performance with an average improvement of 4.2% accuracy and .036 mean intersection-over-union (mIoU). Using a practical estimate of semantic similarity, we still achieve improvements in more than half of all cases, with average improvements of 2.5% accuracy and .017 mIoU in those cases.
KW - Aerial Images
KW - Deep Learning
KW - Domain Adaptation
KW - Neural Networks
KW - Semantic Segmentation
KW - Transfer Learning
UR - http://www.scopus.com/inward/record.url?scp=85091065244&partnerID=8YFLogxK
U2 - 10.5194/isprs-annals-V-2-2020-483-2020
DO - 10.5194/isprs-annals-V-2-2020-483-2020
M3 - Conference article
AN - SCOPUS:85091065244
SP - 483
EP - 492
JO - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
JF - ISPRS Annals of the Photogrammetry, Remote Sensing and Spatial Information Sciences
SN - 2194-9042
T2 - 2020 24th ISPRS Congress on Technical Commission II
Y2 - 31 August 2020 through 2 September 2020
ER -