Investigation of generative adversarial network training: The effect of hyperparameters on training time and stability
2021 (English)Independent thesis Basic level (degree of Bachelor), 20 credits / 30 HE credits
Student thesis
Abstract [en]
Generative Adversarial Networks (GAN) is a technique used to learn the distribution of some dataset in order to generate similar data. GAN models are notoriously difficult to train, which has caused limited deployment in the industry. The results of this study can be used to accelerate the process of making GANs production ready.
An experiment was conducted where multiple GAN models were trained, with the hyperparameters Leaky ReLU alpha, convolutional filters, learning rate and batch size as independent variables. A Mann-Whitney U-test was used to compare the training time and training stability of each model to the others’.
Except for the Leaky ReLU alpha, changes to the investigated hyperparameters had a significant effect on the training time and stability. This study is limited to a few hyperparameters and values, a single dataset and few data points, further research in the area could look at the generalisability of the results or investigate more hyperparameters.
Place, publisher, year, edition, pages
2021. , p. 53, xi
Keywords [en]
Generative adversarial networks, hyperparameters, training, neural networks, deep learning, EMNIST
National Category
Information Systems, Social aspects
Identifiers
URN: urn:nbn:se:his:diva-19847OAI: oai:DiVA.org:his-19847DiVA, id: diva2:1567525
Subject / course
Informationsteknologi
Educational program
Computer Science - Specialization in Systems Development
Supervisors
Examiners
2021-06-162021-06-162021-06-16Bibliographically approved