Investigation of generative adversarial network training: The effect of hyperparameters on training time and stability
2021 (Engelska)Självständigt arbete på grundnivå (kandidatexamen), 20 poäng / 30 hp
Studentuppsats (Examensarbete)
Abstract [en]
Generative Adversarial Networks (GAN) is a technique used to learn the distribution of some dataset in order to generate similar data. GAN models are notoriously difficult to train, which has caused limited deployment in the industry. The results of this study can be used to accelerate the process of making GANs production ready.
An experiment was conducted where multiple GAN models were trained, with the hyperparameters Leaky ReLU alpha, convolutional filters, learning rate and batch size as independent variables. A Mann-Whitney U-test was used to compare the training time and training stability of each model to the others’.
Except for the Leaky ReLU alpha, changes to the investigated hyperparameters had a significant effect on the training time and stability. This study is limited to a few hyperparameters and values, a single dataset and few data points, further research in the area could look at the generalisability of the results or investigate more hyperparameters.
Ort, förlag, år, upplaga, sidor
2021. , s. 53, xi
Nyckelord [en]
Generative adversarial networks, hyperparameters, training, neural networks, deep learning, EMNIST
Nationell ämneskategori
Systemvetenskap, informationssystem och informatik med samhällsvetenskaplig inriktning
Identifikatorer
URN: urn:nbn:se:his:diva-19847OAI: oai:DiVA.org:his-19847DiVA, id: diva2:1567525
Ämne / kurs
Informationsteknologi
Utbildningsprogram
Datavetenskap - inriktning systemutveckling
Handledare
Examinatorer
2021-06-162021-06-162021-06-16Bibliografiskt granskad