h1

h2

h3

h4

h5
h6
http://join2-wiki.gsi.de/foswiki/pub/Main/Artwork/join2_logo100x88.png

Model-based analysis of the image generation quality of adversarial latent autoencoders for industrial machine vision



VerantwortlichkeitsangabeRuslan Yermakov

ImpressumAachen : RWTH Aachen University 2023

Umfang1 Online-Ressource : Illustrationen, Diagramme


Masterarbeit, RWTH Aachen University, 2022

Veröffentlicht auf dem Publikationsserver der RWTH Aachen University


Genehmigende Fakultät
Fak01

Hauptberichter/Gutachter
; ;

Tag der mündlichen Prüfung/Habilitation
2022-10-12

Online
DOI: 10.18154/RWTH-2023-03250
URL: https://publications.rwth-aachen.de/record/954566/files/954566.pdf

Einrichtungen

  1. Juniorprofessur für Mathematische Bild- und Signalverarbeitung (112430)
  2. Fachgruppe Mathematik (110000)
  3. Aachen Institute for Advanced Study in Computational Engineering Science (080003)

Inhaltliche Beschreibung (Schlagwörter)
deep learning (frei) ; feature extraction (frei) ; generative modeling (frei) ; image quality (frei) ; machine learning (frei) ; machine vision (frei)

Thematische Einordnung (Klassifikation)
DDC: 510

Kurzfassung
In industrial machine vision applications, generative models have an advantage over discriminative models because they enable interpretable feature extraction and overcome the limitations of non-transparent and inexplicable decisions associated with the latter. The state-of-the-art style-based generative adversarial networks (StyleGANs) generate high-quality images mapped from a latent feature vector space by learning the approximation of the high-dimensional training data distribution. Thereby, learned representations can be identified by interpreting the latent space and used to control the properties of the synthesized images. StyleGANs in combination with an embedding algorithm - adversarial latent autoencoder (ALAE) - enable the assessment of the properties of embedded images through their latent space representations. However, in order to achieve the best possible approximation of the training data distribution, it is necessary to optimize the network's design parameters based on its effectiveness to learn the quality characteristics of industrial machine vision data. This requires a quantitative evaluation of the quality of generated, and reconstructed images in comparison to the quality of original images. This work presents an evaluation framework to assess the capabilities of generative ALAE models to learn the features and characteristics of the original data consistently and reliably. Feature consistency is proposed as an evaluation criterium to estimate the performance of the generative models. The quality of images generated by ALAE is quantitatively evaluated, with a focus on determining how the latent space size affects the outcome. Latent space size systematically varied during training on industrially relevant datasets with representative features. Based on the application-specific and human-interpretable features, the image quality of multiple ALAEs with varying latent space dimensionalities is compared using statistical tests to select the most favorable latent space dimension.

OpenAccess:
Download fulltext PDF

Dokumenttyp
Master Thesis

Format
online

Sprache
English

Interne Identnummern
RWTH-2023-03250
Datensatz-ID: 954566

Beteiligte Länder
Germany

 GO


OpenAccess

QR Code for this record

The record appears in these collections:
Document types > Theses > Master Theses
Faculty of Mathematics and Natural Sciences (Fac.1) > Department of Mathematics
Publication server / Open Access
Central and Other Institutions
Public records
Publications database
080003
110000
112430

 Record created 2023-03-29, last modified 2025-10-14


OpenAccess:
Download fulltext PDF
Rate this document:

Rate this document:
1
2
3
 
(Not yet reviewed)