Details
Originalsprache | Englisch |
---|---|
Titel des Sammelwerks | Proc. SPIE 12019, AI and Optical Data Sciences III |
Herausgeber/-innen | Bahram Jalali, Ken-ichi Kitayama |
ISBN (elektronisch) | 9781510649095 |
Publikationsstatus | Veröffentlicht - 2022 |
Veranstaltung | SPIE OPTO - San Francisco, USA / Vereinigte Staaten Dauer: 22 Jan. 2022 → 27 Jan. 2022 |
Publikationsreihe
Name | Proceedings of SPIE - The International Society for Optical Engineering |
---|---|
Band | 12019 |
ISSN (Print) | 0277-786X |
ISSN (elektronisch) | 1996-756X |
Abstract
an exchange of weights and optimizer states from the GANs to the segmentation networks, which improves both convergence certainty and accuracy in training. The usage of unsupervised domains in training with the GANs leads to a better generalization of the networks and tackles the challenges caused by the domain gap. Furthermore, a test series is presented that demonstrates the impact of these methods compared to standard supervised training and transfer learning methods based on common datasets. Finally, the developed CNNs are compared to larger state-of-the-art segmentation networks in terms of feed-forward computational time, accuracy and training duration.
ASJC Scopus Sachgebiete
- Werkstoffwissenschaften (insg.)
- Elektronische, optische und magnetische Materialien
- Physik und Astronomie (insg.)
- Physik der kondensierten Materie
- Mathematik (insg.)
- Angewandte Mathematik
- Ingenieurwesen (insg.)
- Elektrotechnik und Elektronik
- Informatik (insg.)
- Angewandte Informatik
Zitieren
- Standard
- Harvard
- Apa
- Vancouver
- BibTex
- RIS
Proc. SPIE 12019, AI and Optical Data Sciences III. Hrsg. / Bahram Jalali; Ken-ichi Kitayama. 2022. 1201902 (Proceedings of SPIE - The International Society for Optical Engineering; Band 12019).
Publikation: Beitrag in Buch/Bericht/Sammelwerk/Konferenzband › Aufsatz in Konferenzband › Forschung
}
TY - GEN
T1 - Damage segmentation using small convolutional neuronal networks and adversarial training methods on low-quality RGB video data
AU - Hedrich, Kolja
AU - Hinz, Lennart
AU - Reithmeier, Eduard
N1 - Funding Information: The underlying project of this conference contribution was funded by the German Federal Ministry of Education and Research as part of the Aviation Research and Technology Program of the Niedersachsen Ministry of Economic Affairs, Employment, Transport and Digitalisation. The project is carried out in cooperation with the MTU Maintenance Hannover GmbH.
PY - 2022
Y1 - 2022
N2 - Within the aviation industry, considerable interest exists in minimizing possible maintenance expenses. In particular, the examination of critical components such as aircraft engines is of significant relevance. Currently, many inspection processes are still performed manually using hand-held endoscopes to detect coating damages in confined spaces and therefore require a high level of individual expertise. Particularly due to the often poorly illuminated video data, these manual inspections are susceptible to uncertainties. This motivates an automated defect detection to provide defined and comparable results and also enable significant cost savings. For such a hand-held application with video data of poor quality, small and fast Convolutional Neural Networks (CNNs) for the segmentation of coating damages are suitable and further examined in this work. Due to high efforts required in image annotation and a significant lack of broadly divergent image data (domain gap), only few expressive annotated images are available. This necessitates extensive training methods to utilize unsupervised domains and further exploit the sparsely annotated data. We propose novel training methods, which implement Generative Adversarial Networks (GAN) to improve the training of segmentation networks by optimizing weights and generating synthetic annotated RGB image data for further training procedures. For this, small individual encoder and decoder structures are designed to resemble the implemented structures of the GANs. This enablesan exchange of weights and optimizer states from the GANs to the segmentation networks, which improves both convergence certainty and accuracy in training. The usage of unsupervised domains in training with the GANs leads to a better generalization of the networks and tackles the challenges caused by the domain gap. Furthermore, a test series is presented that demonstrates the impact of these methods compared to standard supervised training and transfer learning methods based on common datasets. Finally, the developed CNNs are compared to larger state-of-the-art segmentation networks in terms of feed-forward computational time, accuracy and training duration.
AB - Within the aviation industry, considerable interest exists in minimizing possible maintenance expenses. In particular, the examination of critical components such as aircraft engines is of significant relevance. Currently, many inspection processes are still performed manually using hand-held endoscopes to detect coating damages in confined spaces and therefore require a high level of individual expertise. Particularly due to the often poorly illuminated video data, these manual inspections are susceptible to uncertainties. This motivates an automated defect detection to provide defined and comparable results and also enable significant cost savings. For such a hand-held application with video data of poor quality, small and fast Convolutional Neural Networks (CNNs) for the segmentation of coating damages are suitable and further examined in this work. Due to high efforts required in image annotation and a significant lack of broadly divergent image data (domain gap), only few expressive annotated images are available. This necessitates extensive training methods to utilize unsupervised domains and further exploit the sparsely annotated data. We propose novel training methods, which implement Generative Adversarial Networks (GAN) to improve the training of segmentation networks by optimizing weights and generating synthetic annotated RGB image data for further training procedures. For this, small individual encoder and decoder structures are designed to resemble the implemented structures of the GANs. This enablesan exchange of weights and optimizer states from the GANs to the segmentation networks, which improves both convergence certainty and accuracy in training. The usage of unsupervised domains in training with the GANs leads to a better generalization of the networks and tackles the challenges caused by the domain gap. Furthermore, a test series is presented that demonstrates the impact of these methods compared to standard supervised training and transfer learning methods based on common datasets. Finally, the developed CNNs are compared to larger state-of-the-art segmentation networks in terms of feed-forward computational time, accuracy and training duration.
KW - CNN
KW - GAN
KW - Semantic segmentation
KW - U-Net
KW - adversarial methods
KW - damage inspection
KW - endoscopic inspection
KW - semi-supervised learning
KW - transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85129869028&partnerID=8YFLogxK
U2 - 10.1117/12.2610123
DO - 10.1117/12.2610123
M3 - Conference contribution
T3 - Proceedings of SPIE - The International Society for Optical Engineering
BT - Proc. SPIE 12019, AI and Optical Data Sciences III
A2 - Jalali, Bahram
A2 - Kitayama, Ken-ichi
T2 - SPIE OPTO
Y2 - 22 January 2022 through 27 January 2022
ER -