his.sePublications
Change search
Link to record
Permanent link

Direct link
BETA
Huhnstock, Nikolas Alexander
Publications (2 of 2) Show all publications
Huhnstock, N. A., Karlsson, A., Riveiro, M. & Steinhauer, H. J. (2019). An Infinite Replicated Softmax Model for Topic Modeling. In: Vicenç Torra, Yasuo Narukawa, Gabriella Pasi, Marco Viviani (Ed.), Modeling Decisions for Artificial Intelligence: 16th International Conference, MDAI 2019, Milan, Italy, September 4–6, 2019, Proceedings. Paper presented at 16th International Conference, MDAI 2019, Milan, Italy, September 4–6, 2019 (pp. 307-318). Springer
Open this publication in new window or tab >>An Infinite Replicated Softmax Model for Topic Modeling
2019 (English)In: Modeling Decisions for Artificial Intelligence: 16th International Conference, MDAI 2019, Milan, Italy, September 4–6, 2019, Proceedings / [ed] Vicenç Torra, Yasuo Narukawa, Gabriella Pasi, Marco Viviani, Springer, 2019, p. 307-318Conference paper, Published paper (Refereed)
Abstract [en]

In this paper, we describe the infinite replicated Softmax model (iRSM) as an adaptive topic model, utilizing the combination of the infinite restricted Boltzmann machine (iRBM) and the replicated Softmax model (RSM). In our approach, the iRBM extends the RBM by enabling its hidden layer to adapt to the data at hand, while the RSM allows for modeling low-dimensional latent semantic representation from a corpus. The combination of the two results is a method that is able to self-adapt to the number of topics within the document corpus and hence, renders manual identification of the correct number of topics superfluous. We propose a hybrid training approach to effectively improve the performance of the iRSM. An empirical evaluation is performed on a standard data set and the results are compared to the results of a baseline topic model. The results show that the iRSM adapts its hidden layer size to the data and when trained in the proposed hybrid manner outperforms the base RSM model.

Place, publisher, year, edition, pages
Springer, 2019
Series
Lecture Notes in Computer Science, ISSN 0302-9743, E-ISSN 1611-3349 ; 11676
Keywords
Restricted Boltzmann machine, Unsupervised learning, Topic modeling, Adaptive Neural Network
National Category
Computer Sciences Language Technology (Computational Linguistics)
Research subject
Skövde Artificial Intelligence Lab (SAIL)
Identifiers
urn:nbn:se:his:diva-17664 (URN)10.1007/978-3-030-26773-5_27 (DOI)978-3-030-26772-8 (ISBN)978-3-030-26773-5 (ISBN)
Conference
16th International Conference, MDAI 2019, Milan, Italy, September 4–6, 2019
Available from: 2019-09-09 Created: 2019-09-10 Last updated: 2019-11-08Bibliographically approved
Huhnstock, N. A., Karlsson, A., Riveiro, M. & Steinhauer, H. J. (2018). On the behavior of the infinite restricted boltzmann machine for clustering. In: Hisham M. Haddad, Roger L. Wainwright, Richard Chbeir (Ed.), SAC '18 Proceedings of the 33rd Annual ACM Symposium on Applied Computing: . Paper presented at SAC 18 The 33rd Annual ACM Symposium on Applied Computing, Pau, France, April 9-13, 2018 (pp. 461-470). New York, NY, USA: Association for Computing Machinery (ACM)
Open this publication in new window or tab >>On the behavior of the infinite restricted boltzmann machine for clustering
2018 (English)In: SAC '18 Proceedings of the 33rd Annual ACM Symposium on Applied Computing / [ed] Hisham M. Haddad, Roger L. Wainwright, Richard Chbeir, New York, NY, USA: Association for Computing Machinery (ACM), 2018, p. 461-470Conference paper, Published paper (Refereed)
Abstract [en]

Clustering is a core problem within a wide range of research disciplines ranging from machine learning and data mining to classical statistics. A group of clustering approaches so-called nonparametric methods, aims to cluster a set of entities into a beforehand unspecified and unknown number of clusters, making potentially expensive pre-analysis of data obsolete. In this paper, the recently, by Cote and Larochelle introduced infinite Restricted Boltzmann Machine that has the ability to self-regulate its number of hidden parameters is adapted to the problem of clustering by the introduction of two basic cluster membership assumptions. A descriptive study of the influence of several regularization and sparsity settings on the clustering behavior is presented and results are discussed. The results show that sparsity is a key adaption when using the iRBM for clustering that improves both the clustering performances as well as the number of identified clusters.

Place, publisher, year, edition, pages
New York, NY, USA: Association for Computing Machinery (ACM), 2018
Keywords
clustering, unsupervised, machine learning, restricted boltzmann machine
National Category
Computer Sciences
Research subject
Skövde Artificial Intelligence Lab (SAIL); INF301 Data Science
Identifiers
urn:nbn:se:his:diva-16505 (URN)10.1145/3167132.3167183 (DOI)000455180700067 ()2-s2.0-85050522612 (Scopus ID)978-1-4503-5191-1 (ISBN)
Conference
SAC 18 The 33rd Annual ACM Symposium on Applied Computing, Pau, France, April 9-13, 2018
Available from: 2018-12-17 Created: 2018-12-17 Last updated: 2019-02-14Bibliographically approved
Organisations

Search in DiVA

Show all publications