Högskolan i Skövde

his.sePublications
Change search
Refine search result
1 - 19 of 19
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • apa-cv
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Rows per page
  • 5
  • 10
  • 20
  • 50
  • 100
  • 250
Sort
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
  • Standard (Relevance)
  • Author A-Ö
  • Author Ö-A
  • Title A-Ö
  • Title Ö-A
  • Publication type A-Ö
  • Publication type Ö-A
  • Issued (Oldest first)
  • Issued (Newest first)
  • Created (Oldest first)
  • Created (Newest first)
  • Last updated (Oldest first)
  • Last updated (Newest first)
  • Disputation date (earliest first)
  • Disputation date (latest first)
Select
The maximal number of hits you can export is 250. When you want to export more records please use the Create feeds function.
  • 1.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Calibrating Random Forests2008In: Proceedings of the Seventh International Conference on Machine Learning and Applications, IEEE, 2008, p. 121-126Conference paper (Refereed)
    Abstract [en]

    When using the output of classifiers to calculate the expected utility of different alternatives in decision situations, the correctness of predicted class probabilities may be of crucial importance. However, even very accurate classifiers may output class probabilities of rather poor quality. One way of overcoming this problem is by means of calibration, i.e., mapping the original class probabilities to more accurate ones. Previous studies have however indicated that random forests are difficult to calibrate by standard calibration methods. In this work, a novel calibration method is introduced, which is based on a recent finding that probabilities predicted by forests of classification trees have a lower squared error compared to those predicted by forests of probability estimation trees (PETs). The novel calibration method is compared to the two standard methods, Platt scaling and isotonic regression, on 34 datasets from the UCI repository. The experiment shows that random forests of PETs calibrated by the novel method significantly outperform uncalibrated random forests of both PETs and classification trees, as well as random forests calibrated with the two standard methods, with respect to the squared error of predicted class probabilities.

  • 2.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Estimating class probabilities in random forests2007In: ICMLA 2007. Sixth International Conference onMachine Learning and Applications, 2007., IEEE Computer Society, 2007, p. 211-216Conference paper (Refereed)
    Abstract [en]

    For both single probability estimation trees (PETs) and ensembles of such trees, commonly employed class probability estimates correct the observed relative class frequencies in each leaf to avoid anomalies caused by small sample sizes. The effect of such corrections in random forests of PETs is investigated, and the use of the relative class frequency is compared to using two corrected estimates, the Laplace estimate and the m-estimate. An experiment with 34 datasets from the UCI repository shows that estimating class probabilities using relative class frequency clearly outperforms both using the Laplace estimate and the m-estimate with respect to accuracy, area under the ROC curve (AUC) and Brier score. Hence, in contrast to what is commonly employed for PETs and ensembles of PETs, these results strongly suggest that a non-corrected probability estimate should be used in random forests of PETs. The experiment further shows that learning random forests of PETs using relative class frequency significantly outperforms learning random forests of classification trees (i.e., trees for which only an unweighted vote on the most probable class is counted) with respect to both accuracy and AUC, but that the latter is clearly ahead of the former with respect to Brier score.

  • 3.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Feature vs. Classifier Fusion for Predictive Data Mining: a Case Study in Pesticide Classification2007In: 10th International Conference on Information Fusion, 2007, IEEE, 2007, p. 1-7Conference paper (Refereed)
    Abstract [en]

    Two strategies for fusing information from multiple sources when generating predictive models in the domain of pesticide classification are investigated: i) fusing different sets of features (molecular descriptors) before building a model and ii) fusing the classifiers built from the individual descriptor sets. An empirical investigation demonstrates that the choice of strategy can have a significant impact on the predictive performance. Furthermore, the experiment shows that the best strategy is dependent on the type of predictive model considered. When generating a decision tree for pesticide classification, a statistically significant difference in accuracy is observed in favor of combining predictions from the individual models compared to generating a single model from the fused set of molecular descriptors. On the other hand, when the model consists of an ensemble of decision trees, a statistically significant difference in accuracy is observed in favor of building the model from the fused set of descriptors compared to fusing ensemble models built from the individual sources.

  • 4.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Maximizing the Area under the ROC Curve with Decision Lists and Rule Sets2007In: Proceedings of the 7th SIAM International Conference on Data Mining / [ed] C. Apte, B. Liu, S. Parthasarathy, D. Skillicorn, Society for Industrial and Applied Mathematics , 2007, p. 27-34Conference paper (Refereed)
    Abstract [en]

    Decision lists (or ordered rule sets) have two attractive properties compared to unordered rule sets: they require a simpler classi¯cation procedure and they allow for a more compact representation. However, it is an open question what effect these properties have on the area under the ROC curve (AUC). Two ways of forming decision lists are considered in this study: by generating a sequence of rules, with a default rule for one of the classes, and by imposing an order upon rules that have been generated for all classes. An empirical investigation shows that the latter method gives a significantly higher AUC than the former, demonstrating that the compactness obtained by using one of the classes as a default is indeed associated with a cost. Furthermore, by using all applicable rules rather than the first in an ordered set, an even further significant improvement in AUC is obtained, demonstrating that the simple classification procedure is also associated with a cost. The observed gains in AUC for unordered rule sets compared to decision lists can be explained by that learning rules for all classes as well as combining multiple rules allow for examples to be ranked according to a more fine-grained scale compared to when applying rules in a fixed order and providing a default rule for one of the classes.

  • 5.
    Boström, Henrik
    et al.
    University of Skövde, School of Humanities and Informatics.
    Andler, Sten F.
    University of Skövde, School of Humanities and Informatics.
    Brohede, Marcus
    University of Skövde, School of Humanities and Informatics.
    Johansson, Ronnie
    University of Skövde, School of Humanities and Informatics.
    Karlsson, Alexander
    University of Skövde, School of Humanities and Informatics.
    van Laere, Joeri
    University of Skövde, School of Humanities and Informatics.
    Niklasson, Lars
    University of Skövde, School of Humanities and Informatics.
    Nilsson, Marie
    University of Skövde, School of Humanities and Informatics.
    Persson, Anne
    University of Skövde, School of Humanities and Informatics.
    Ziemke, Tom
    University of Skövde, School of Humanities and Informatics.
    On the Definition of Information Fusion as a Field of Research2007Report (Other academic)
    Abstract [en]

    A more precise definition of the field of information fusion can be of benefit to researchers within the field, who may use uch a definition when motivating their own work and evaluating the contribution of others. Moreover, it can enable researchers and practitioners outside the field to more easily relate their own work to the field and more easily understand the scope of the techniques and methods developed in the field. Previous definitions of information fusion are reviewed from that perspective, including definitions of data and sensor fusion, and their appropriateness as definitions for the entire research field are discussed. Based on strengths and weaknesses of existing definitions, a novel definition is proposed, which is argued to effectively fulfill the requirements that can be put on a definition of information fusion as a field of research.

    Download full text (pdf)
    FULLTEXT01
  • 6.
    Boström, Henrik
    et al.
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Johansson, Ronnie
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Karlsson, Alexander
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    On Evidential Combination Rules for Ensemble Classifiers2008In: Proceedings of the 11th International Conference on Information Fusion: Cologne, Germany, June 30 - July 03, 2008, IEEE, 2008, p. 553-560Conference paper (Refereed)
    Abstract [en]

    Ensemble classifiers are known to generally perform better than each individual classifier of which they consist. One approach to classifier fusion is to apply Shafer’s theory of evidence. While most approaches have adopted Dempster’s rule of combination, a multitude of combination rules have been proposed. A number of combination rules as well as two voting rules are compared when used in conjunction with a specific kind of ensemble classifier, known as random forests, w.r.t. accuracy, area under ROC curve and Brier score on 27 datasets. The empirical evaluation shows that the choice of combination rule can have a significant impact on the performance for a single dataset, but in general the evidential combination rules do not perform better than the voting rules for this particular ensemble design. Furthermore, among the evidential rules, the associative ones appear to have better performance than the non-associative.

  • 7.
    Boström, Henrik
    et al.
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre. Stockholm University, Sweden.
    Norinder, Ulf
    AstraZeneca R&D, Södertälje, Sweden.
    Utilizing Information on Uncertainty for In Silico Modeling using Random Forests2009In: Proceedings of the 3rd Skövde Workshop on Information Fusion Topics (SWIFT 2009), Skövde: University of Skövde , 2009, p. 59-62Conference paper (Refereed)
    Abstract [en]

    Information on uncertainty of measurements or estimates of molecular properties are rarely utilized by in silico predictive models. In this study, different approaches to handling uncertain numerical features are explored when using the stateof- the-art random forest algorithm for generating predictive models. Two main approaches are considered: i) sampling from probability distributions prior to tree generation, which does not require any change to the underlying tree learning algorithm, and ii) adjusting the algorithm to allow for handling probability distributions, similar to how missing values typically are handled, i.e., partitions may include fractions of examples. An experiment with six datasets concerning the prediction of various chemical properties is presented, where 95% confidence intervals are included for one of the 92 numerical features. In total, five approaches to handling uncertain numeric features are compared: ignoring the uncertainty, sampling from distributions that are assumed to be uniform and normal respectively, and adjusting tree learning to handle probability distributions that are assumed to be uniform and normal respectively. The experimental results show that all approaches that utilize information on uncertainty indeed outperform the single approach ignoring this, both with respect to accuracy and area under ROC curve. A decomposition of the squared error of the constituent classification trees shows that the highest variance is obtained by ignoring the information on uncertainty, but that this also results in the highest mean squared error of the constituent trees.

  • 8.
    Deegalla, Sampath
    et al.
    Dept. of Computer and Systems Sciences, Stockholm University and Royal Institute of Technology, Sweden.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Classification of Microarrays with kNN: Comparison of Dimensionality Reduction Methods2007In: Intelligent Data Engineering and Automated Learning – IDEAL 2007: 8th International Conference Birmingham, UK, December 16-19, 2007 Proceedings / [ed] Hujun Yin, Peter Tino, Emilio Corchado, Will Byrne, Xin Yao, Springer, 2007, p. 800-809Conference paper (Refereed)
    Abstract [en]

    Dimensionality reduction can often improve the performance of the k-nearest neighbor classifier (kNN) for high-dimensional data sets, such as microarrays. The effect of the choice of dimensionality reduction method on the predictive performance of kNN for classifying microarray data is an open issue, and four common dimensionality reduction methods, Principal Component Analysis (PCA), Random Projection (RP), Partial Least Squares (PLS) and Information Gain(IG), are compared on eight microarray data sets. It is observed that all dimensionality reduction methods result in more accurate classifiers than what is obtained from using the raw attributes. Furthermore, it is observed that both PCA and PLS reach their best accuracies with fewer components than the other two methods, and that RP needs far more components than the others to outperform kNN on the non-reduced dataset. None of the dimensionality reduction methods can be concluded to generally outperform the others, although PLS is shown to be superior on all four binary classification tasks, but the main conclusion from the study is that the choice of dimensionality reduction method can be of major importance when classifying microarrays using kNN.

  • 9.
    Deegalla, Sampath
    et al.
    Dept. of Computer and Systems Sciences, Stockholm University, Sweden.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Fusion of Dimensionality Reduction Methods: a Case Study in Microarray Classification2009In: Proceedings of the 12th International Conference on Information Fusion, ISIF , 2009, p. 460-465Conference paper (Refereed)
    Abstract [en]

    Dimensionality reduction has been demonstrated to improve the performance of the k-nearest neighbor (kNN) classifier for high-dimensional data sets, such as microarrays. However, the effectiveness of different dimensionality reduction methods varies, and it has been shown that no single method constantly outperforms the others. In contrast to using a single method, two approaches to fusing the result of applying dimensionality reduction methods are investigated: feature fusion and classifier fusion. It is shown that by fusing the output of multiple dimensionality reduction techniques, either by fusing the reduced features or by fusing the output of the resulting classifiers, both higher accuracy and higher robustness towards the choice of number of dimensions is obtained.

  • 10.
    Dudas, Catarina
    et al.
    University of Skövde, School of Technology and Society. University of Skövde, The Virtual Systems Research Centre.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Using Uncertain Chemical and Thermal Data to Predict Product Quality in a Casting Process2009In: U '09: Proceedings of the 1st ACM SIGKDD Workshop on Knowledge Discovery from Uncertain Data / [ed] Jian Pei, Lise Getoor, Ander De Keijzer, ACM Digital Library, 2009, p. 57-61Conference paper (Refereed)
    Abstract [en]

    Process and casting data from different sources have been collected and merged for the purpose of predicting, and determining what factors affect, the quality of cast products in a foundry. One problem is that the measurements cannot be directly aligned, since they are collected at different points in time, and instead they have to be approximated for specific time points, hence introducing uncertainty. An approach for addressing this problem is investigated, where uncertain numeric features values are represented by intervals and random forests are extended to handle such intervals. A preliminary experiment shows that the suggested way of forming the intervals, together with the extension of random forests, results in higher predictive performance compared to using single (expected) values for the uncertain features together with standard random forests.

  • 11.
    Dudas, Catarina
    et al.
    University of Skövde, The Virtual Systems Research Centre. University of Skövde, School of Technology and Society.
    Ng, Amos
    University of Skövde, The Virtual Systems Research Centre. University of Skövde, School of Technology and Society.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Information Extraction from Solution Set of Simulation-based Multi-objective Optimisation using Data Mining2009In: Proceedings of Industrial Simulation Conference 2009 / [ed] D. B. Das, V. Nassehi & L. Deka, EUROSIS-ETI , 2009, p. 65-69Conference paper (Refereed)
    Abstract [en]

    In this work, we investigate ways of extracting information from simulations, in particular from simulation-based multi-objective optimisation, in order to acquire information that can support human decision makers that aim for optimising manufacturing processes. Applying data mining for analyzing data generated using simulation is a fairly unexplored area. With the observation that the obtained solutions from a simulation-based multi-objective optimisation are all optimal (or close to the optimal Pareto front) so that they are bound to follow and exhibit certain relationships among variables vis-à-vis objectives, it is argued that using data mining to discover these relationships could be a promising procedure. The aim of this paper is to provide the empirical results from two simulation case studies to support such a hypothesis.

  • 12.
    Dudas, Catarina
    et al.
    University of Skövde, School of Technology and Society.
    Ng, Amos
    University of Skövde, School of Technology and Society.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics.
    Knowledge Extraction in Manufacturing using Data Mining Techniques2008In: Proceedings of the Swedish Production Symposium 2008, Stockholm, Sweden, November 18-20, 2008, 2008, p. 8 sidor-Conference paper (Refereed)
    Abstract [en]

    Nowadays many production companies collect and store production and process data in large databases. Unfortunately the data is rarely used in the most value generating way, i.e.,  finding  patterns  of  inconsistencies  and  relationships  between  process  settings  and quality  outcome.  This  paper  addresses  the  benefits  of  using  data  mining  techniques  in manufacturing  applications.  Two  different  applications  are  being  laid  out  but  the  used technique  and  software  is  the  same  in  both  cases.  The  first  case  deals  with  how  data mining  can  be  used  to  discover  the  affect  of  process  timing  and  settings  on  the  quality outcome in the casting industry. The result of a multi objective optimization of a camshaft process  is  being  used  as  the  second  case.  This  study  focuses  on  finding  the  most appropriate dispatching rule settings in the buffers on the line.  The  use  of  data  mining  techniques  in  these  two  cases  generated  previously  unknown knowledge. For example, in order to maximize throughput in the camshaft production, let the dispatching rule for the most severe bottleneck be of type Shortest Processing Time (SPT) and for the second bottleneck use any but Most Work Remaining (MWKR).

  • 13.
    Johansson, Ronnie
    et al.
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Karlsson, Alexander
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    A Study on Class-Specifically Discounted Belief for Ensemble Classifiers2008In: 2008 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems, IEEE, 2008, p. 614-619, article id 4648012Conference paper (Refereed)
    Abstract [en]

    Ensemble classifiers are known to generally perform better than their constituent classifiers. Whereas a lot of work has been focusing on the generation of classifiers for ensembles, much less attention has been given to the fusion of individual classifier outputs. One approach to fuse the outputs is to apply Shafer’s theory of evidence, which provides a flexible framework for expressing and fusing beliefs. However, representing and fusing beliefs is non-trivial since it can be performed in a multitude of ways within the evidential framework. In a previous article, we compared different evidential combination rules for ensemble fusion. The study involved a single belief representation which involved discounting (i.e., weighting) the classifier outputs with classifier reliability. The classifier reliability was interpreted as the classifier’s estimated accuracy, i.e., the percentage of correctly classified examples. However, classifiers may have different performance for different classes and in this work we assign the reliability of a classifier output depending on the classspecific reliability of the classifier. Using 27 UCI datasets, we compare the two different ways of expressing beliefs and some evidential combination rules. The result of the study indicates that there is indeed an advantage of utilizing class-specific reliability compared to accuracy in an evidential framework for combining classifiers in the ensemble design considered.

  • 14.
    Johansson, Ulf
    et al.
    School of Business and Informatics, University of Borås, Sweden.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    König, Rikard
    School of Business and Informatics, University of Borås, Sweden.
    Extending Nearest Neighbor Classification with Spheres of Confidence2008In: Proceedings of the Twenty-First International FLAIRS Conference (FLAIRS 2008), AAAI Press, 2008, p. 282-287Conference paper (Refereed)
    Abstract [en]

    The standard kNN algorithm suffers from two major drawbacks: sensitivity to the parameter value k, i.e., the number of neighbors, and the use of k as a global constant that is independent of the particular region in which theexample to be classified falls. Methods using weighted voting schemes only partly alleviate these problems, since they still involve choosing a fixed k. In this paper, a novel instance-based learner is introduced that does not require kas a parameter, but instead employs a flexible strategy for determining the number of neighbors to consider for the specific example to be classified, hence using a local instead of global k. A number of variants of the algorithm are evaluated on 18 datasets from the UCI repository. The novel algorithm in its basic form is shown to significantly outperform standard kNN with respect to accuracy, and an adapted version of the algorithm is shown to be clearlyahead with respect to the area under ROC curve. Similar to standard kNN, the novel algorithm still allows for various extensions, such as weighted voting and axes scaling.

  • 15.
    Johansson, Ulf
    et al.
    School of Business and Informatics, University of Borås, Sweden.
    Löfström, Tuve
    School of Business and Informatics, University of Borås, Sweden.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    The Problem with Ranking Ensembles Based on Training or Validation Performance2008In: 2008 IEEE International Joint Conference on Neural Networks (IEEE World Congress on Computational Intelligence), IEEE, 2008, p. 3222-3228Conference paper (Refereed)
    Abstract [en]

    The main purpose of this study was to determine whether it is possible to somehow use results on training or validation data to estimate ensemble performance on novel data. With the specific setup evaluated; i.e. using ensembles built from a pool of independently trained neural networks and targeting diversity only implicitly, the answer is a resounding no. Experimentation, using 13 UCI datasets, shows that there is in general nothing to gain in performance on novel data by choosing an ensemble based on any of the training measures evaluated here. This is despite the fact that the measures evaluated include all the most frequently used; i.e. ensemble training and validation accuracy, base classifier training and validation accuracy, ensemble training and validation AUC and two diversity measures. The main reason is that all ensembles tend to have quite similar performance, unless we deliberately lower the accuracy of the base classifiers. The key consequence is, of course, that a data miner can do no better than picking an ensemble at random. In addition, the results indicate that it is futile to look for an algorithm aimed at optimizing ensemble performance by somehow selecting a subset of available base classifiers.

     

  • 16.
    Johansson, Ulf
    et al.
    University of Borås, School of Business and Informatics, Borås, Sweden.
    Sönströd, Cecilia
    University of Borås, School of Business and Informatics, Borås, Sweden.
    Löfström, Tuve
    University of Skövde, School of Humanities and Informatics.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Chipper - A Novel Algorithm for Concept Description2008In: Proceedings of the 2008 conference on Tenth Scandinavian Conference on Artificial Intelligence: SCAI 2008 / [ed] Anders Holst, Per Kreuger, Peter Funk, IOS Press, 2008, p. 133-140Conference paper (Refereed)
    Abstract [en]

    In this paper, several demands placed on concept description algorithms are identified and discussed. The most important criterion is the ability to produce compact rule sets that, in a natural and accurate way, describe the most important relationships in the underlying domain. An algorithm based on the identified criteria is presented and evaluated. The algorithm, named Chipper, produces decision lists, where each rule covers a maximum number of remaining instances while meeting requested accuracy requirements. In the experiments, Chipper is evaluated on nine UCI data sets. The main result is that Chipper produces compact and understandable rule sets, clearly fulfilling the overall goal of concept description. In the experiments, Chipper’s accuracy is similar to standard decision tree and rule induction algorithms, while rule sets have superior comprehensibility.

  • 17.
    Löfström, Tuve
    et al.
    School of Business and Informatics, University o f Borås, Borås, Sweden.
    Johansson, Ulf
    School of Business and Informatics, University o f Borås, Borås, Sweden.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Ensemble Member Selection Using Multi-Objective Optimization2009In: Proceedings of IEEE Symposium on Computational Intelligence and Data Mining (CIDM), IEEE conference proceedings, 2009, p. 245-251Conference paper (Refereed)
    Abstract [en]

    Both theory and a wealth of empirical studies have established that ensembles are more accurate than single predictive models. Unfortunately, the problem of how to maximize ensemble accuracy is, especially for classification, far from solved. In essence, the key problem is to find a suitable criterion, typically based on training or selection set performance, highly correlated with ensemble accuracy on novel data. Several studies have, however, shown that it is difficult to come up with a single measure, such as ensemble or base classifier selection set accuracy, or some measure based on diversity, that is a good general predictor for ensemble test accuracy. This paper presents a novel technique that for each learning task searches for the most effective combination of given atomic measures, by means of a genetic algorithm. Ensembles built from either neural networks or random forests were empirically evaluated on 30 UCI datasets. The experimental results show that when using the generated combined optimization criteria to rank candidate ensembles, a higher test set accuracy for the top ranked ensemble was achieved, compared to using ensemble accuracy on selection data alone. Furthermore, when creating ensembles from a pool of neural networks, the use of the generated combined criteria was shown to generally outperform the use of estimated ensemble accuracy as the single optimization criterion.

  • 18.
    Löfström, Tuve
    et al.
    School of Business and Informatics, University of Borås.
    Johansson, Ulf
    School of Business and Informatics, University of Borås.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    On the Use of Accuracy and Diversity Measures for Evaluating and Selecting Ensembles of Classifiers2008In: Proceedings of the Seventh International Conference on Machine Learning and Applications, IEEE, 2008, p. 127-132Conference paper (Refereed)
    Abstract [en]

    The test set accuracy for ensembles of classifiers selected based on single measures of accuracy and diversity as well as combinations of such measures is investigated. It is found that by combining measures, a higher test set accuracy may be obtained than by using any single accuracy or diversity measure. It is further investigated whether a multi-criteria search for an ensemble that maximizes both accuracy and diversity leads to more accurate ensembles than by optimizing a single criterion. The results indicate that it might be more beneficial to search for ensembles that are both accurate and diverse. Furthermore, the results show that diversity measures could compete with accuracy measures as selection criterion.

  • 19.
    Sönströd, Cecilia
    et al.
    School of Business and Informatics, University of Borås, Sweden.
    Johansson, Ulf
    School of Business and Informatics, University of Borås, Sweden.
    Norinder, Ulf
    AstraZeneca R&D, Södertälje, Sweden.
    Boström, Henrik
    University of Skövde, School of Humanities and Informatics. University of Skövde, The Informatics Research Centre.
    Comprehensible Models for Predicting Molecular Interaction with Heart-Regulating Genes2008In: Proceedings of the Seventh International Conference on Machine Learning and Applications, IEEE, 2008, p. 559-564Conference paper (Refereed)
    Abstract [en]

    When using machine learning for in silico modeling, the goal is normally to obtain highly accurate predictive models. Often, however, models should also bring insights into interesting relationships in the domain. It is then desirable that machine learning techniques have the ability to obtain small and transparent models, where the user can control the tradeoff between accuracy, comprehensibility and coverage. In this study, three different decision list algorithms are evaluated on a data set concerning the interaction of molecules with a human gene that regulates heart functioning (hERG). The results show that decision list algorithms can obtain predictive performance not far from the state-of-the-art method random forests, but also that algorithms focusing on accuracy alone may produce complex decision lists that are very hard to interpret. The experiments also show that by sacrificing accuracy only to a limited degree, comprehensibility (measured as both model size and classification complexity) can be improved remarkably.

1 - 19 of 19
CiteExportLink to result list
Permanent link
Cite
Citation style
  • apa
  • apa-cv
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf