SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "hsv:(NATURVETENSKAP) hsv:(Data och informationsvetenskap) ;pers:(Boström Henrik)"

Sökning: hsv:(NATURVETENSKAP) hsv:(Data och informationsvetenskap) > Boström Henrik

  • Resultat 1-10 av 173
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1.
  • Linnusson, Henrik, et al. (författare)
  • Efficient conformal predictor ensembles
  • 2020
  • Ingår i: Neurocomputing. - : Elsevier BV. - 0925-2312 .- 1872-8286. ; 397, s. 266-278
  • Tidskriftsartikel (refereegranskat)abstract
    • In this paper, we study a generalization of a recently developed strategy for generating conformal predictor ensembles: out-of-bag calibration. The ensemble strategy is evaluated, both theoretically and empirically, against a commonly used alternative ensemble strategy, bootstrap conformal prediction, as well as common non-ensemble strategies. A thorough analysis is provided of out-of-bag calibration, with respect to theoretical validity, empirical validity (error rate), efficiency (prediction region size) and p-value stability (the degree of variance observed over multiple predictions for the same object). Empirical results show that out-of-bag calibration displays favorable characteristics with regard to these criteria, and we propose that out-of-bag calibration be adopted as a standard method for constructing conformal predictor ensembles.
  •  
2.
  • Boström, Henrik (författare)
  • Maximizing the Area under the ROC Curve with Decision Lists and Rule Sets
  • 2007
  • Ingår i: Proceedings of the 7th SIAM International Conference on Data Mining. - : Society for Industrial and Applied Mathematics. - 9780898716306 ; , s. 27-34
  • Konferensbidrag (refereegranskat)abstract
    • Decision lists (or ordered rule sets) have two attractive properties compared to unordered rule sets: they require a simpler classi¯cation procedure and they allow for a more compact representation. However, it is an open question what effect these properties have on the area under the ROC curve (AUC). Two ways of forming decision lists are considered in this study: by generating a sequence of rules, with a default rule for one of the classes, and by imposing an order upon rules that have been generated for all classes. An empirical investigation shows that the latter method gives a significantly higher AUC than the former, demonstrating that the compactness obtained by using one of the classes as a default is indeed associated with a cost. Furthermore, by using all applicable rules rather than the first in an ordered set, an even further significant improvement in AUC is obtained, demonstrating that the simple classification procedure is also associated with a cost. The observed gains in AUC for unordered rule sets compared to decision lists can be explained by that learning rules for all classes as well as combining multiple rules allow for examples to be ranked according to a more fine-grained scale compared to when applying rules in a fixed order and providing a default rule for one of the classes.
  •  
3.
  • Johansson, Ronnie, et al. (författare)
  • A Study on Class-Specifically Discounted Belief for Ensemble Classifiers
  • 2008
  • Ingår i: 2008 IEEE International Conference on Multisensor Fusion and Integration for Intelligent Systems. - : IEEE. - 9781424421442 - 9781424421435 ; , s. 614-619
  • Konferensbidrag (refereegranskat)abstract
    • Ensemble classifiers are known to generally perform better than their constituent classifiers. Whereas a lot of work has been focusing on the generation of classifiers for ensembles, much less attention has been given to the fusion of individual classifier outputs. One approach to fuse the outputs is to apply Shafer’s theory of evidence, which provides a flexible framework for expressing and fusing beliefs. However, representing and fusing beliefs is non-trivial since it can be performed in a multitude of ways within the evidential framework. In a previous article, we compared different evidential combination rules for ensemble fusion. The study involved a single belief representation which involved discounting (i.e., weighting) the classifier outputs with classifier reliability. The classifier reliability was interpreted as the classifier’s estimated accuracy, i.e., the percentage of correctly classified examples. However, classifiers may have different performance for different classes and in this work we assign the reliability of a classifier output depending on the classspecific reliability of the classifier. Using 27 UCI datasets, we compare the two different ways of expressing beliefs and some evidential combination rules. The result of the study indicates that there is indeed an advantage of utilizing class-specific reliability compared to accuracy in an evidential framework for combining classifiers in the ensemble design considered.
  •  
4.
  • Johansson, Ulf, et al. (författare)
  • Conformal Prediction Using Decision Trees
  • 2013
  • Ingår i: IEEE 13th International Conference on Data Mining (ICDM). - : IEEE Computer Society. - 9780769551081 ; , s. 330-339
  • Konferensbidrag (refereegranskat)abstract
    • Conformal prediction is a relatively new framework in which the predictive models output sets of predictions with a bound on the error rate, i.e., in a classification context, the probability of excluding the correct class label is lower than a predefined significance level. An investigation of the use of decision trees within the conformal prediction framework is presented, with the overall purpose to determine the effect of different algorithmic choices, including split criterion, pruning scheme and way to calculate the probability estimates. Since the error rate is bounded by the framework, the most important property of conformal predictors is efficiency, which concerns minimizing the number of elements in the output prediction sets. Results from one of the largest empirical investigations to date within the conformal prediction framework are presented, showing that in order to optimize efficiency, the decision trees should be induced using no pruning and with smoothed probability estimates. The choice of split criterion to use for the actual induction of the trees did not turn out to have any major impact on the efficiency. Finally, the experimentation also showed that when using decision trees, standard inductive conformal prediction was as efficient as the recently suggested method cross-conformal prediction. This is an encouraging results since cross-conformal prediction uses several decision trees, thus sacrificing the interpretability of a single decision tree.
  •  
5.
  • Johansson, Ulf, et al. (författare)
  • Evolved Decision Trees as Conformal Predictors
  • 2013
  • Ingår i: 2013 IEEE Congress on Evolutionary Computation (CEC). - : IEEE. - 9781479904532 ; , s. 1794-1801
  • Konferensbidrag (refereegranskat)abstract
    • In conformal prediction, predictive models output sets of predictions with a bound on the error rate. In classification, this translates to that the probability of excluding the correct class is lower than a predefined significance level, in the long run. Since the error rate is guaranteed, the most important criterion for conformal predictors is efficiency. Efficient conformal predictors minimize the number of elements in the output prediction sets, thus producing more informative predictions. This paper presents one of the first comprehensive studies where evolutionary algorithms are used to build conformal predictors. More specifically, decision trees evolved using genetic programming are evaluated as conformal predictors. In the experiments, the evolved trees are compared to decision trees induced using standard machine learning techniques on 33 publicly available benchmark data sets, with regard to predictive performance and efficiency. The results show that the evolved trees are generally more accurate, and the corresponding conformal predictors more efficient, than their induced counterparts. One important result is that the probability estimates of decision trees when used as conformal predictors should be smoothed, here using the Laplace correction. Finally, using the more discriminating Brier score instead of accuracy as the optimization criterion produced the most efficient conformal predictions.
  •  
6.
  • Johansson, Ulf, et al. (författare)
  • Extending Nearest Neighbor Classification with Spheres of Confidence
  • 2008
  • Ingår i: Proceedings of the Twenty-First International FLAIRS Conference (FLAIRS 2008). - : AAAI Press. - 9781577353652 ; , s. 282-287
  • Konferensbidrag (refereegranskat)abstract
    • The standard kNN algorithm suffers from two major drawbacks: sensitivity to the parameter value k, i.e., the number of neighbors, and the use of k as a global constant that is independent of the particular region in which theexample to be classified falls. Methods using weighted voting schemes only partly alleviate these problems, since they still involve choosing a fixed k. In this paper, a novel instance-based learner is introduced that does not require kas a parameter, but instead employs a flexible strategy for determining the number of neighbors to consider for the specific example to be classified, hence using a local instead of global k. A number of variants of the algorithm are evaluated on 18 datasets from the UCI repository. The novel algorithm in its basic form is shown to significantly outperform standard kNN with respect to accuracy, and an adapted version of the algorithm is shown to be clearlyahead with respect to the area under ROC curve. Similar to standard kNN, the novel algorithm still allows for various extensions, such as weighted voting and axes scaling.
  •  
7.
  • Johansson, Ulf, et al. (författare)
  • Overproduce-and-Select : The Grim Reality
  • 2013
  • Ingår i: 2013 IEEE Symposium on Computational Intelligence and Ensemble Learning (CIEL). - : IEEE. - 9781467358538 ; , s. 52-59
  • Konferensbidrag (refereegranskat)abstract
    • Overproduce-and-select (OPAS) is a frequently used paradigm for building ensembles. In static OPAS, a large number of base classifiers are trained, before a subset of the available models is selected to be combined into the final ensemble. In general, the selected classifiers are supposed to be accurate and diverse for the OPAS strategy to result in highly accurate ensembles, but exactly how this is enforced in the selection process is not obvious. Most often, either individual models or ensembles are evaluated, using some performance metric, on available and labeled data. Naturally, the underlying assumption is that an observed advantage for the models (or the resulting ensemble) will carry over to test data. In the experimental study, a typical static OPAS scenario, using a pool of artificial neural networks and a number of very natural and frequently used performance measures, is evaluated on 22 publicly available data sets. The discouraging result is that although a fairly large proportion of the ensembles obtained higher test set accuracies, compared to using the entire pool as the ensemble, none of the selection criteria could be used to identify these highly accurate ensembles. Despite only investigating a specific scenario, we argue that the settings used are typical for static OPAS, thus making the results general enough to question the entire paradigm.
  •  
8.
  • Johansson, Ulf, et al. (författare)
  • Random Brains
  • 2013
  • Ingår i: The 2013 International Joint Conference on Neural Networks (IJCNN). - : IEEE. - 9781467361286 ; , s. 1-8
  • Konferensbidrag (refereegranskat)abstract
    • In this paper, we introduce and evaluate a novel method, called random brains, for producing neural network ensembles. The suggested method, which is heavily inspired by the random forest technique, produces diversity implicitly by using bootstrap training and randomized architectures. More specifically, for each base classifier multilayer perceptron, a number of randomly selected links between the input layer and the hidden layer are removed prior to training, thus resulting in potentially weaker but more diverse base classifiers. The experimental results on 20 UCI data sets show that random brains obtained significantly higher accuracy and AUC, compared to standard bagging of similar neural networks not utilizing randomized architectures. The analysis shows that the main reason for the increased ensemble performance is the ability to produce effective diversity, as indicated by the increase in the difficulty diversity measure.
  •  
9.
  • Johansson, Ulf, et al. (författare)
  • Regression Trees for Streaming Data with Local Performance Guarantees
  • 2014
  • Konferensbidrag (refereegranskat)abstract
    • Online predictive modeling of streaming data is a key task for big data analytics. In this paper, a novel approach for efficient online learning of regression trees is proposed, which continuously updates, rather than retrains, the tree as more labeled data become available. A conformal predictor outputs prediction sets instead of point predictions; which for regression translates into prediction intervals. The key property of a conformal predictor is that it is always valid, i.e., the error rate, on novel data, is bounded by a preset significance level. Here, we suggest applying Mondrian conformal prediction on top of the resulting models, in order to obtain regression trees where not only the tree, but also each and every rule, corresponding to a path from the root node to a leaf, is valid. Using Mondrian conformal prediction, it becomes possible to analyze and explore the different rules separately, knowing that their accuracy, in the long run, will not be below the preset significance level. An empirical investigation, using 17 publicly available data sets, confirms that the resulting rules are independently valid, but also shows that the prediction intervals are smaller, on average, than when only the global model is required to be valid. All-in-all, the suggested method provides a data miner or a decision maker with highly informative predictive models of streaming data.
  •  
10.
  • Johansson, Ulf, et al. (författare)
  • Rule Extraction with Guaranteed Fidelity
  • 2014
  • Ingår i: Artificial Intelligence Applications and Innovations. - Cham : Springer. - 9783662447215 - 9783662447222 ; , s. 281-290
  • Konferensbidrag (refereegranskat)abstract
    • This paper extends the conformal prediction framework to rule extraction, making it possible to extract interpretable models from opaque models in a setting where either the infidelity or the error rate is bounded by a predefined significance level. Experimental results on 27 publicly available data sets show that all three setups evaluated produced valid and rather efficient conformal predictors. The implication is that augmenting rule extraction with conformal prediction allows extraction of models where test set errors or test sets infidelities are guaranteed to be lower than a chosen acceptable level. Clearly this is beneficial for both typical rule extraction scenarios, i.e., either when the purpose is to explain an existing opaque model, or when it is to build a predictive model that must be interpretable.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1-10 av 173
Typ av publikation
konferensbidrag (120)
tidskriftsartikel (34)
doktorsavhandling (10)
proceedings (redaktörskap) (2)
bokkapitel (2)
licentiatavhandling (2)
visa fler...
konstnärligt arbete (1)
rapport (1)
annan publikation (1)
patent (1)
visa färre...
Typ av innehåll
refereegranskat (151)
övrigt vetenskapligt/konstnärligt (21)
populärvet., debatt m.m. (1)
Författare/redaktör
Johansson, Ulf (52)
Löfström, Tuve (21)
Asker, Lars (21)
Papapetrou, Panagiot ... (19)
Zhao, Jing (15)
visa fler...
Linusson, Henrik (14)
Karlsson, Isak (14)
Norinder, Ulf (13)
Löfström, Tuwe, 1977 ... (13)
Sönströd, Cecilia (13)
Henriksson, Aron (8)
Deegalla, Sampath (8)
Boström, Henrik, Pro ... (8)
Karunaratne, Thashme ... (6)
Carlsson, Lars (6)
Lindgren, Tony (6)
Dudas, Catarina (6)
Ennadir, Sofiane (5)
Vazirgiannis, Michal ... (5)
Dalianis, Hercules (5)
Alkhatib, Amr (5)
Gurung, Ram B. (5)
Ahlberg, Ernst (4)
Norinder, Ulf, 1956- (4)
Sundell, Håkan (4)
Henelius, Andreas (4)
Johansson, U (3)
Ng, Amos H. C., 1970 ... (3)
Karlsson, Alexander (3)
Johansson, Ronnie (3)
König, Rikard (3)
Walgama, Keerthi (3)
Sotomane, Constantin ... (3)
Jansson, Karl (3)
Vasiloudis, Theodore (3)
Puolamäki, Kai (3)
Abbahaddou, Yassine (2)
Lutzeyer, Johannes F ... (2)
Karunaratne, Thashme ... (2)
Ng, A (2)
Ng, Amos (2)
Löfström, Tuwe (2)
Vesterberg, Anders (2)
Dudas, C. (2)
Lidén, Per (2)
Sönströd, C. (2)
Löfström, T. (2)
Löfström, Tuve, 1977 ... (2)
Linnusson, Henrik (2)
visa färre...
Lärosäte
Kungliga Tekniska Högskolan (145)
Stockholms universitet (106)
Jönköping University (43)
Högskolan i Borås (30)
Högskolan i Skövde (21)
Örebro universitet (4)
visa fler...
Uppsala universitet (3)
Malmö universitet (1)
Mittuniversitetet (1)
Linnéuniversitetet (1)
visa färre...
Språk
Engelska (173)
Forskningsämne (UKÄ/SCB)
Naturvetenskap (173)
Teknik (3)
Samhällsvetenskap (3)
Medicin och hälsovetenskap (2)

År

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy