Högskolan i Skövde

his.sePublications
Change search
Link to record
Permanent link

Direct link
König, Rikard
Alternative names
Publications (10 of 13) Show all publications
Johansson, U., König, R. & Niklasson, L. (2009). Genetically Evolved kNN Ensembles (1ed.). In: Robert Stahlbock, Sven F. Crone, Stefan Lessmann (Ed.), Data Mining: Special Issue in Annals of Information Systems (pp. 299-313). Springer Science+Business Media B.V.
Open this publication in new window or tab >>Genetically Evolved kNN Ensembles
2009 (English)In: Data Mining: Special Issue in Annals of Information Systems / [ed] Robert Stahlbock, Sven F. Crone, Stefan Lessmann, Springer Science+Business Media B.V., 2009, 1, p. 299-313Chapter in book (Other academic)
Abstract [en]

Both theory and a wealth of empirical studies have established that ensembles are more accurate than single predictive models. For the ensemble approach to work, base classifiers must not only be accurate but also diverse, i.e., they should commit their errors on different instances. Instance-based learners are, however, very robust with respect to variations of a data set, so standard resampling methods will normally produce only limited diversity. Because of this, instance-based learners are rarely used as base classifiers in ensembles. In this chapter, we introduce a method where genetic programming is used to generate kNN base classifiers with optimized k-values and feature weights. Due to the inherent inconsistency in genetic programming (i.e., different runs using identical data and parameters will still produce different solutions) a group of independently evolved base classifiers tend to be not only accurate but also diverse. In the experimentation, using 30 data sets from the UCI repository, two slightly different versions of kNN ensembles are shown to significantly outperform both the corresponding base classifiers and standard kNN with optimized k-values, with respect to accuracy and AUC.

Place, publisher, year, edition, pages
Springer Science+Business Media B.V., 2009 Edition: 1
Series
Annals of Information Systems, ISSN 1934-3221 ; 8
National Category
Computer and Information Sciences
Research subject
Technology
Identifiers
urn:nbn:se:his:diva-3839 (URN)10.1007/978-1-4419-1280-0_13 (DOI)978-1-4419-1279-4 (ISBN)978-1-4419-1280-0 (ISBN)
Available from: 2010-04-01 Created: 2010-04-01 Last updated: 2018-01-12Bibliographically approved
König, R. (2009). Predictive Techniques and Methods for Decision Support in Situations with Poor Data Quality. (Licentiate dissertation). Örebro University
Open this publication in new window or tab >>Predictive Techniques and Methods for Decision Support in Situations with Poor Data Quality
2009 (English)Licentiate thesis, comprehensive summary (Other academic)
Abstract [en]

Today, decision support systems based on predictive modeling are becoming more common, since organizations often collect more data than decision makers can handle manually. Predictive models are used to find potentially valuable patterns in the data, or to predict the outcome of some event. There are numerous predictive techniques, ranging from simple techniques such as linear regression, to complex powerful ones like artificial neural networks. Complex models usually obtain better predictive performance, but are opaque and thus cannot be used to explain predictions or discovered patterns. The design choice of which predictive technique to use becomes even harder since no technique outperforms all others over a large set of problems. It is even difficult to find the best parameter values for a specific technique, since these settings also are problem dependent. One way to simplify this vital decision is to combine several models, possibly created with different settings and techniques, into an ensemble. Ensembles are known to be more robust and powerful than individual models, and ensemble diversity can be used to estimate the uncertainty associated with each prediction.

In real-world data mining projects, data is often imprecise, contain uncertainties or is missing important values, making it impossible to create models with sufficient performance for fully automated systems. In these cases, predictions need to be manually analyzed and adjusted. Here, opaque models like ensembles have a disadvantage, since the analysis requires understandable models. To overcome this deficiency of opaque models, researchers have developed rule extraction techniques that try to extract comprehensible rules from opaque models, while retaining sufficient accuracy.

This thesis suggests a straightforward but comprehensive method for predictive modeling in situations with poor data quality. First, ensembles are used for the actual modeling, since they are powerful, robust and require few design choices. Next, ensemble uncertainty estimations pinpoint predictions that need special attention from a decision maker. Finally, rule extraction is performed to support the analysis of uncertain predictions. Using this method, ensembles can be used for predictive modeling, in spite of their opacity and sometimes insufficient global performance, while the involvement of a decision maker is minimized.

The main contributions of this thesis are three novel techniques that enhance the performance of the purposed method. The first technique deals with ensemble uncertainty estimation and is based on a successful approach often used in weather forecasting. The other two are improvements of a rule extraction technique, resulting in increased comprehensibility and more accurate uncertainty estimations.

Place, publisher, year, edition, pages
Örebro University, 2009. p. 112
Series
Studies from the School of Science and Technology at Örebro University ; 5
Keywords
Rule Extraction, Genetic Programming, Uncertainty estimation, Machine Learning, Artificial Neural Networks, Data Mining, Information Fusion
National Category
Computer and Information Sciences
Research subject
Technology
Identifiers
urn:nbn:se:his:diva-3208 (URN)
Presentation
(English)
Available from: 2009-06-26 Created: 2009-06-26 Last updated: 2018-01-13Bibliographically approved
König, R., Johansson, U. & Niklasson, L. (2008). G-REX: A Versatile Framework for Evolutionary Data Mining. In: Francesco Bonchi; Bettina Berendt; Fosca Giannotti; Dimitrios Gunopulos; Franco Turini; Carlo Zaniolo; Naren Ramakrishnan; Xindong Wu (Ed.), Proceedings - IEEE International Conference on Data Mining Workshops, ICDM Workshops 2008: . Paper presented at IEEE International Conference on Data Mining Workshops, ICDM Workshops 2008, Pisa, 15 December 2008 through 19 December 2008 (pp. 971-974). IEEE
Open this publication in new window or tab >>G-REX: A Versatile Framework for Evolutionary Data Mining
2008 (English)In: Proceedings - IEEE International Conference on Data Mining Workshops, ICDM Workshops 2008 / [ed] Francesco Bonchi; Bettina Berendt; Fosca Giannotti; Dimitrios Gunopulos; Franco Turini; Carlo Zaniolo; Naren Ramakrishnan; Xindong Wu, IEEE, 2008, p. 971-974Conference paper, Published paper (Refereed)
Abstract [en]

This paper presents G-REX, a versatile data mining framework based on Genetic Programming. What differs G-REX from other GP frameworks is that it doesn’t strive to be a general purpose framework. This allows G-REX to include more functionality specific to data mining like preprocessing, evaluation- and optimization methods, but also a multitude of predefined classification and regression models. Examples of predefined models are decision trees, decision lists, k-NN with attribute weights, hybrid kNN-rules, fuzzy-rules and several different regression models. The main strength is, however, the flexibility, making it easy to modify, extend and combine all of the predefined functionality. G-REX is, in addition, available in a special Weka package adding useful evolutionary functionality to the standard data mining tool Weka.

 

 

Place, publisher, year, edition, pages
IEEE, 2008
Series
IEEE International Conference on Data Mining Workshops, ICDMW, ISSN 2375-9232, E-ISSN 2375-9259
National Category
Computer Sciences
Research subject
Technology
Identifiers
urn:nbn:se:his:diva-2829 (URN)10.1109/ICDMW.2008.117 (DOI)2-s2.0-62449143933 (Scopus ID)978-0-7695-3503-6 (ISBN)
Conference
IEEE International Conference on Data Mining Workshops, ICDM Workshops 2008, Pisa, 15 December 2008 through 19 December 2008
Available from: 2009-03-04 Created: 2009-03-04 Last updated: 2021-04-22Bibliographically approved
König, R., Johansson, U. & Niklasson, L. (2007). Genetic Programming: a Tool for Flexible Rule Extraction. In: 2007 IEEE Congress on Evolutionary Computation: . Paper presented at 2007 IEEE Congress on Evolutionary Computation, 25-28 Sept. 2007, Singapore (pp. 1304-1310). IEEE
Open this publication in new window or tab >>Genetic Programming: a Tool for Flexible Rule Extraction
2007 (Swedish)In: 2007 IEEE Congress on Evolutionary Computation, IEEE, 2007, p. 1304-1310Conference paper, Published paper (Refereed)
Abstract [en]

Although data mining is performed to support decision making, many of the most powerful techniques, like neural networks and ensembles, produce opaque models. This lack of interpretability is an obvious disadvantage, since decision makers normally require some sort of explanation before taking action. To achieve comprehensibility, accuracy is often sacrificed by the use of simpler, transparent models, such as decision trees. Another alternative is rule extraction; i.e. to transform the opaque model into a comprehensible model, keeping acceptable accuracy. We have previously suggested a rule extraction algorithm named G-REX, which is based on genetic programming. One key property of G-REX, due to the use of genetic programming, is the possibility to use different representation languages. In this study we apply G-REX to estimation tasks. More specifically, three representation languages are evaluated using eight publicly available data sets. The quality of the extracted rules is compared to two standard techniques producing comprehensible models; multiple linear regression and the decision tree algorithm C&RT. The results show that G-REX outperforms the standard techniques, but that the choice of representation language is important.

Place, publisher, year, edition, pages
IEEE, 2007
Series
IEEE Transactions on Evolutionary Computation, ISSN 1089-778X, E-ISSN 1941-0026
National Category
Computer Sciences Information Systems
Research subject
Technology
Identifiers
urn:nbn:se:his:diva-2102 (URN)10.1109/CEC.2007.4424621 (DOI)000256053700175 ()2-s2.0-62449331153 (Scopus ID)1-4244-1340-0 (ISBN)978-1-4244-1339-3 (ISBN)978-1-4244-1340-9 (ISBN)
Conference
2007 IEEE Congress on Evolutionary Computation, 25-28 Sept. 2007, Singapore
Funder
Knowledge Foundation, 2003/0104
Note

This work was supported by the Information Fusion Research Program (University of Skövde, Sweden) in partnership with the Swedish Knowledge Foundation under grant 2003/0104 (URL:http://www.infofusion.se).

Available from: 2008-05-30 Created: 2008-05-30 Last updated: 2021-04-27Bibliographically approved
Löfström, T., Johansson, U., Sönströd, C., König, R. & Niklasson, L. (Eds.). (2007). Proceedings of SAIS 2007: The 24th Annual Workshop of the Swedish Artificial Intelligence Society, Borås, May 22-23, 2007. Paper presented at SAIS 2007: The 24th Annual Workshop of the Swedish Artificial Intelligence Society, Borås, May 22-23, 2007. Borås: University College of Borås
Open this publication in new window or tab >>Proceedings of SAIS 2007: The 24th Annual Workshop of the Swedish Artificial Intelligence Society, Borås, May 22-23, 2007
Show others...
2007 (English)Conference proceedings (editor) (Other academic)
Place, publisher, year, edition, pages
Borås: University College of Borås, 2007
National Category
Computer Sciences
Research subject
Technology
Identifiers
urn:nbn:se:his:diva-3701 (URN)
Conference
SAIS 2007: The 24th Annual Workshop of the Swedish Artificial Intelligence Society, Borås, May 22-23, 2007
Available from: 2010-02-16 Created: 2010-02-16 Last updated: 2021-05-06Bibliographically approved
Johansson, U., Löfström, T., König, R. & Niklasson, L. (2006). Accurate Neural Network Ensembles Using Genetic Programming. In: Michael Minock; Patrik Eklund; Helena Lindgren (Ed.), Proceedings of SAIS 2006: The 23rd Annual Workshop of the Swedish Artificial Intelligence Society. Paper presented at The 23rd Annual Workshop of the Swedish Artificial Intelligence Society, SAIS 2006, Umeå, May 10-12, 2006 (pp. 117-126). Umeå: Swedish Artificial Intelligence Society - SAIS, Umeå University
Open this publication in new window or tab >>Accurate Neural Network Ensembles Using Genetic Programming
2006 (English)In: Proceedings of SAIS 2006: The 23rd Annual Workshop of the Swedish Artificial Intelligence Society / [ed] Michael Minock; Patrik Eklund; Helena Lindgren, Umeå: Swedish Artificial Intelligence Society - SAIS, Umeå University , 2006, p. 117-126Conference paper, Published paper (Refereed)
Abstract [en]

Abstract: In this paper we present and evaluate a novel algorithm for ensemble creation. The main idea of the algorithm is to first independently train a fixed number of neural networks (here ten) and then use genetic programming to combine these networks into an ensemble. The use of genetic programming makes it possible to not only consider ensembles of different sizes, but also to use ensembles as intermediate building blocks. The final result is therefore more correctly described as an ensemble of neural network ensembles. The experiments show that the proposed method, when evaluated on 22 publicly available data sets, obtains very high accuracy, clearly outperforming the other methods evaluated. In this study several micro techniques are used, and we believe that they all contribute to the increased performance.

One such micro technique, aimed at reducing overtraining, is the training method, called tombola training, used during genetic evolution. When using tombola training, training data is regularly resampled into new parts, called training groups. Each ensemble is then evaluated on every training group and the actual fitness is determined solely from the result on the hardest part.

Place, publisher, year, edition, pages
Umeå: Swedish Artificial Intelligence Society - SAIS, Umeå University, 2006
Series
Report / UMINF - Umeå University, Department of Computing Science, ISSN 0348-0542 ; 06.19
National Category
Computer Sciences Information Systems
Identifiers
urn:nbn:se:his:diva-1914 (URN)
Conference
The 23rd Annual Workshop of the Swedish Artificial Intelligence Society, SAIS 2006, Umeå, May 10-12, 2006
Available from: 2007-03-22 Created: 2007-03-22 Last updated: 2021-06-28Bibliographically approved
Löfström, T., König, R., Johansson, U., Niklasson, L., Strand, M. & Ziemke, T. (2006). Benefits of Relating the Retail Domain to Information Fusion. In: 9th International Conference on Information Fusion: IEEE ISIF. Paper presented at 9th International Conference on Information Fusion, ICIF '06, Florence, Italy, July 10-13, 2006. IEEE
Open this publication in new window or tab >>Benefits of Relating the Retail Domain to Information Fusion
Show others...
2006 (English)In: 9th International Conference on Information Fusion: IEEE ISIF, IEEE, 2006Conference paper, Published paper (Refereed)
Abstract [en]

In this paper a mapping between retail concepts and the JDL model is proposed. More specifically, the benefits of using solutions to military problems as inspiration to retail specific problems are discussed. The somewhat surprising conclusion is that there are several similarities between the military and retail domains, and that these similarities potentially could be exploited. A few examples of retail problems that could benefit from theories and techniques commonly used in the Information Fusion community are given. All examples are taken from recently started or planned projects within the Information Fusion research program at the University of Skövde, Sweden.

Place, publisher, year, edition, pages
IEEE, 2006
Keywords
Information Fusion, Retail
Identifiers
urn:nbn:se:his:diva-1956 (URN)10.1109/ICIF.2006.301644 (DOI)2-s2.0-50149109426 (Scopus ID)978-1-4244-0953-2 (ISBN)
Conference
9th International Conference on Information Fusion, ICIF '06, Florence, Italy, July 10-13, 2006
Note

ISBN: 0-9721844-6-5

Available from: 2008-04-11 Created: 2008-04-11 Last updated: 2023-06-20Bibliographically approved
Johansson, U., Löfström, T., König, R. & Niklasson, L. (2006). Building Neural Network Ensembles using Genetic Programming. In: The 2006 IEEE International Joint Conference on Neural Network Proceedings: . Paper presented at International Joint Conference on Neural Networks 2006, IJCNN '06, Vancouver, BC, 16 July 2006 through 21 July 2006 (pp. 1260-1265). IEEE
Open this publication in new window or tab >>Building Neural Network Ensembles using Genetic Programming
2006 (English)In: The 2006 IEEE International Joint Conference on Neural Network Proceedings, IEEE, 2006, p. 1260-1265Conference paper, Published paper (Refereed)
Abstract [en]

In this paper we present and evaluate a novel algorithm for ensemble creation. The main idea of the algorithm is to first independently train a fixed number of neural networks (here ten) and then use genetic programming to combine these networks into an ensemble. The use of genetic programming makes it possible to not only consider ensembles of different sizes, but also to use ensembles as intermediate building blocks. The final result is therefore more correctly described as an ensemble of neural network ensembles. The experiments show that the proposed method, when evaluated on 22 publicly available data sets, obtains very high accuracy, clearly outperforming the other methods evaluated. In this study several micro techniques are used, and we believe that they all contribute to the increased performance. One such micro technique, aimed at reducing overtraining, is the training method, called tombola training, used during genetic evolution. When using tombola training, training data is regularly resampled into new parts, called training groups. Each ensemble is then evaluated on every training group and the actual fitness is determined solely from the result on the hardest part.

Place, publisher, year, edition, pages
IEEE, 2006
Series
Proceedings of the International Joint Conference on Neural Networks, ISSN 2161-4393, E-ISSN 2161-4407
National Category
Computer Sciences Information Systems
Identifiers
urn:nbn:se:his:diva-1806 (URN)10.1109/IJCNN.2006.246836 (DOI)000245125902029 ()2-s2.0-38049049329 (Scopus ID)0-7803-9490-9 (ISBN)978-0-7803-9490-2 (ISBN)
Conference
International Joint Conference on Neural Networks 2006, IJCNN '06, Vancouver, BC, 16 July 2006 through 21 July 2006
Available from: 2007-10-10 Created: 2007-10-10 Last updated: 2021-04-22Bibliographically approved
Johansson, U., Löfström, T., König, R. & Niklasson, L. (2006). Genetically Evolved Trees Representing Ensembles. In: Leszek Rutkowski, Ryszard Tadeusiewicz, Lotfi A. Zadeh, Jacek M. Żurada (Ed.), Artificial Intelligence and Soft Computing – ICAISC 2006: 8th International Conference, Zakopane, Poland, June 25-29, 2006. Proceedings. Paper presented at Artificial Intelligence and Soft Computing – ICAISC 2006, 8th International Conference, Zakopane, Poland, June 25-29, 2006 (pp. 613-622). Springer
Open this publication in new window or tab >>Genetically Evolved Trees Representing Ensembles
2006 (English)In: Artificial Intelligence and Soft Computing – ICAISC 2006: 8th International Conference, Zakopane, Poland, June 25-29, 2006. Proceedings / [ed] Leszek Rutkowski, Ryszard Tadeusiewicz, Lotfi A. Zadeh, Jacek M. Żurada, Springer, 2006, p. 613-622Conference paper, Published paper (Refereed)
Abstract [en]

We have recently proposed a novel algorithm for ensemble creation called GEMS (Genetic Ensemble Member Selection). GEMS first trains a fixed number of neural networks (here twenty) and then uses genetic programming to combine these networks into an ensemble. The use of genetic programming makes it possible for GEMS to not only consider ensembles of different sizes, but also to use ensembles as intermediate building blocks. In this paper, which is the first extensive study of GEMS, the representation language is extended to include tests partitioning the data, further increasing flexibility. In addition, several micro techniques are applied to reduce overfitting, which appears to be the main problem for this powerful algorithm. The experiments show that GEMS, when evaluated on 15 publicly available data sets, obtains very high accuracy, clearly outperforming both straightforward ensemble designs and standard decision tree algorithms.

Place, publisher, year, edition, pages
Springer, 2006
Series
Lecture Notes in Computer Science, ISSN 0302-9743, E-ISSN 1611-3349 ; 4029
National Category
Computer Sciences Information Systems
Research subject
Technology
Identifiers
urn:nbn:se:his:diva-1587 (URN)10.1007/11785231_64 (DOI)000239600000064 ()2-s2.0-33746239343 (Scopus ID)978-3-540-35748-3 (ISBN)978-3-540-35750-6 (ISBN)3-540-35748-3 (ISBN)
Conference
Artificial Intelligence and Soft Computing – ICAISC 2006, 8th International Conference, Zakopane, Poland, June 25-29, 2006
Note

Also part of the Lecture Notes in Artificial Intelligence book sub series (LNAI, volume 4029)

Available from: 2008-02-08 Created: 2008-02-08 Last updated: 2021-04-22Bibliographically approved
König, R., Johansson, U. & Niklasson, L. (2006). Increasing rule extraction comprehensibility. International Journal of Information Technology and Intelligent Computing, 1(2), 303-314
Open this publication in new window or tab >>Increasing rule extraction comprehensibility
2006 (English)In: International Journal of Information Technology and Intelligent Computing, ISSN 1895-8648, Vol. 1, no 2, p. 303-314Article in journal (Refereed) Published
Place, publisher, year, edition, pages
Łódź: Academy of Humanities and Economics (WSHE), 2006
Identifiers
urn:nbn:se:his:diva-7190 (URN)
Available from: 2013-02-11 Created: 2013-02-11 Last updated: 2017-11-27Bibliographically approved
Organisations

Search in DiVA

Show all publications