SwePub
Sök i SwePub databas

  Utökad sökning

Träfflista för sökning "WAKA:kon ;lar1:(hb)"

Sökning: WAKA:kon > Högskolan i Borås

  • Resultat 1651-1660 av 3464
Sortera/gruppera träfflistan
   
NumreringReferensOmslagsbildHitta
1651.
  • Johansson, Ulf, et al. (författare)
  • Oracle Coached Decision Trees and Lists
  • 2010
  • Konferensbidrag (refereegranskat)abstract
    • This paper introduces a novel method for obtaining increased predictive performance from transparent models in situations where production input vectors are available when building the model. First, labeled training data is used to build a powerful opaque model, called an oracle. Second, the oracle is applied to production instances, generating predicted target values, which are used as labels. Finally, these newly labeled instances are utilized, in different combinations with normal training data, when inducing a transparent model. Experimental results, on 26 UCI data sets, show that the use of oracle coaches significantly improves predictive performance, compared to standard model induction. Most importantly, both accuracy and AUC results are robust over all combinations of opaque and transparent models evaluated. This study thus implies that the straightforward procedure of using a coaching oracle, which can be used with arbitrary classifiers, yields significantly better predictive performance at a low computational cost.
  •  
1652.
  • Johansson, Ulf, et al. (författare)
  • Overproduce-and-Select : The Grim Reality
  • 2013
  • Ingår i: 2013 IEEE Symposium on Computational Intelligence and Ensemble Learning (CIEL). - : IEEE. - 9781467358538 ; , s. 52-59
  • Konferensbidrag (refereegranskat)abstract
    • Overproduce-and-select (OPAS) is a frequently used paradigm for building ensembles. In static OPAS, a large number of base classifiers are trained, before a subset of the available models is selected to be combined into the final ensemble. In general, the selected classifiers are supposed to be accurate and diverse for the OPAS strategy to result in highly accurate ensembles, but exactly how this is enforced in the selection process is not obvious. Most often, either individual models or ensembles are evaluated, using some performance metric, on available and labeled data. Naturally, the underlying assumption is that an observed advantage for the models (or the resulting ensemble) will carry over to test data. In the experimental study, a typical static OPAS scenario, using a pool of artificial neural networks and a number of very natural and frequently used performance measures, is evaluated on 22 publicly available data sets. The discouraging result is that although a fairly large proportion of the ensembles obtained higher test set accuracies, compared to using the entire pool as the ensemble, none of the selection criteria could be used to identify these highly accurate ensembles. Despite only investigating a specific scenario, we argue that the settings used are typical for static OPAS, thus making the results general enough to question the entire paradigm.
  •  
1653.
  • Johansson, Ulf, et al. (författare)
  • Producing Implicit Diversity in ANN Ensembles
  • 2012
  • Ingår i: The 2012 International Joint Conference on Neural Networks (IJCNN). - : IEEE. - 9781467314886 - 9781467314893 ; , s. 1-8
  • Konferensbidrag (refereegranskat)abstract
    • Combining several ANNs into ensembles normally results in a very accurate and robust predictive models. Many ANN ensemble techniques are, however, quite complicated and often explicitly optimize some diversity metric. Unfortunately, the lack of solid validation of the explicit algorithms, at least for classification, makes the use of diversity measures as part of an optimization function questionable. The merits of implicit methods, most notably bagging, are on the other hand experimentally established and well-known. This paper evaluates a number of straightforward techniques for introducing implicit diversity in ANN ensembles, including a novel technique producing diversity by using ANNs with different and slightly randomized link structures. The experimental results, comparing altogether 54 setups and two different ensemble sizes on 30 UCI data sets, show that all methods succeeded in producing implicit diversity, but that the effect on ensemble accuracy varied. Still, most setups evaluated did result in more accurate ensembles, compared to the baseline setup, especially for the larger ensemble size. As a matter of fact, several setups even obtained significantly higher ensemble accuracy than bagging. The analysis also identified that diversity was, relatively speaking, more important for the larger ensembles. Looking specifically at the methods used to increase the implicit diversity, setups using the technique that utilizes the randomized link structures generally produced the most accurate ensembles.
  •  
1654.
  • Johansson, Ulf, et al. (författare)
  • Random Brains
  • 2013
  • Ingår i: The 2013 International Joint Conference on Neural Networks (IJCNN). - : IEEE. - 9781467361286 ; , s. 1-8
  • Konferensbidrag (refereegranskat)abstract
    • In this paper, we introduce and evaluate a novel method, called random brains, for producing neural network ensembles. The suggested method, which is heavily inspired by the random forest technique, produces diversity implicitly by using bootstrap training and randomized architectures. More specifically, for each base classifier multilayer perceptron, a number of randomly selected links between the input layer and the hidden layer are removed prior to training, thus resulting in potentially weaker but more diverse base classifiers. The experimental results on 20 UCI data sets show that random brains obtained significantly higher accuracy and AUC, compared to standard bagging of similar neural networks not utilizing randomized architectures. The analysis shows that the main reason for the increased ensemble performance is the ability to produce effective diversity, as indicated by the increase in the difficulty diversity measure.
  •  
1655.
  • Johansson, Ulf, et al. (författare)
  • Regression Trees for Streaming Data with Local Performance Guarantees
  • 2014
  • Konferensbidrag (refereegranskat)abstract
    • Online predictive modeling of streaming data is a key task for big data analytics. In this paper, a novel approach for efficient online learning of regression trees is proposed, which continuously updates, rather than retrains, the tree as more labeled data become available. A conformal predictor outputs prediction sets instead of point predictions; which for regression translates into prediction intervals. The key property of a conformal predictor is that it is always valid, i.e., the error rate, on novel data, is bounded by a preset significance level. Here, we suggest applying Mondrian conformal prediction on top of the resulting models, in order to obtain regression trees where not only the tree, but also each and every rule, corresponding to a path from the root node to a leaf, is valid. Using Mondrian conformal prediction, it becomes possible to analyze and explore the different rules separately, knowing that their accuracy, in the long run, will not be below the preset significance level. An empirical investigation, using 17 publicly available data sets, confirms that the resulting rules are independently valid, but also shows that the prediction intervals are smaller, on average, than when only the global model is required to be valid. All-in-all, the suggested method provides a data miner or a decision maker with highly informative predictive models of streaming data.
  •  
1656.
  • Johansson, Ulf, et al. (författare)
  • Rule Extraction with Guaranteed Fidelity
  • 2014
  • Ingår i: Artificial Intelligence Applications and Innovations. - Cham : Springer. - 9783662447215 - 9783662447222 ; , s. 281-290
  • Konferensbidrag (refereegranskat)abstract
    • This paper extends the conformal prediction framework to rule extraction, making it possible to extract interpretable models from opaque models in a setting where either the infidelity or the error rate is bounded by a predefined significance level. Experimental results on 27 publicly available data sets show that all three setups evaluated produced valid and rather efficient conformal predictors. The implication is that augmenting rule extraction with conformal prediction allows extraction of models where test set errors or test sets infidelities are guaranteed to be lower than a chosen acceptable level. Clearly this is beneficial for both typical rule extraction scenarios, i.e., either when the purpose is to explain an existing opaque model, or when it is to build a predictive model that must be interpretable.
  •  
1657.
  • Johansson, Ulf, et al. (författare)
  • The Importance of Diversity in Neural Network Ensembles : An Empirical Investigation
  • 2007
  • Ingår i: IJCNN 2007 Conference Proceedings. - : IEEE. - 9781424413799 - 9781424413805 - 142441380X - 142441380X ; , s. 661-666
  • Konferensbidrag (refereegranskat)abstract
    • When designing ensembles, it is almost an axiom that the base classifiers must be diverse in order for the ensemble to generalize well. Unfortunately, there is no clear definition of the key term diversity, leading to several diversity measures and many, more or less ad hoc, methods for diversity creation in ensembles. In addition, no specific diversity measure has shown to have a high correlation with test set accuracy. The purpose of this paper is to empirically evaluate ten different diversity measures, using neural network ensembles and 11 publicly available data sets. The main result is that all diversity measures evaluated, in this study too, show low or very low correlation with test set accuracy. Having said that, two measures; double fault and difficulty show slightly higher correlations compared to the other measures. The study furthermore shows that the correlation between accuracy measured on training or validation data and test set accuracy also is rather low. These results challenge ensemble design techniques where diversity is explicitly maximized or where ensemble accuracy on a hold-out set is used for optimization.
  •  
1658.
  • Johansson, Ulf, et al. (författare)
  • Using Feature Selection with Bagging and Rule Extraction in Drug Discovery
  • 2010
  • Konferensbidrag (refereegranskat)abstract
    • This paper investigates different ways of combining feature selection with bagging and rule extraction in predictive modeling. Experiments on a large number of data sets from the medicinal chemistry domain, using standard algorithms implemented in theWeka data mining workbench, show that feature selection can lead to significantly improved predictive performance.When combining feature selection with bagging, employing the feature selection on each bootstrap obtains the best result.When using decision trees for rule extraction, the effect of feature selection can actually be detrimental, unless the transductive approach oracle coaching is also used. However, employing oracle coaching will lead to significantly improved performance, and the best results are obtainedwhen performing feature selection before training the opaque model. The overall conclusion is that it can make a substantial difference for the predictive performance exactly how feature selection is used in conjunction with other techniques.
  •  
1659.
  • Johansson, Ulf, et al. (författare)
  • Using Genetic Programming to Obtain Implicit Diversity
  • 2009
  • Konferensbidrag (refereegranskat)abstract
    • When performing predictive data mining, the use of ensembles is known to increase prediction accuracy, compared to single models. To obtain this higher accuracy, ensembles should be built from base classifiers that are both accurate and diverse. The question of how to balance these two properties in order to maximize ensemble accuracy is, however, far from solved and many different techniques for obtaining ensemble diversity exist. One such technique is bagging, where implicit diversity is introduced by training base classifiers on different subsets of available data instances, thus resulting in less accurate, but diverse base classifiers. In this paper, genetic programming is used as an alternative method to obtain implicit diversity in ensembles by evolving accurate, but different base classifiers in the form of decision trees, thus exploiting the inherent inconsistency of genetic programming. The experiments show that the GP approach outperforms standard bagging of decision trees, obtaining significantly higher ensemble accuracy over 25 UCI datasets. This superior performance stems from base classifiers having both higher average accuracy and more diversity. Implicitly introducing diversity using GP thus works very well, since evolved base classifiers tend to be highly accurate and diverse.
  •  
1660.
  • Johansson, Ulf, et al. (författare)
  • Using Imaginary Ensembles to Select GP Classifiers
  • 2010
  • Ingår i: Genetic Programming: 13th European Conference, EuroGP 2010, Istanbul, Turkey, April 7-9, 2010, Proceedings. - Berlin, Heidelberg : Springer. - 9783642121470 - 9783642121487 - 3642121470 ; , s. 278-288
  • Konferensbidrag (refereegranskat)abstract
    • When predictive modeling requires comprehensible models, most data miners will use specialized techniques producing rule sets or decision trees. This study, however, shows that genetically evolved decision trees may very well outperform the more specialized techniques. The proposed approach evolves a number of decision trees and then uses one of several suggested selection strategies to pick one specific tree from that pool. The inherent inconsistency of evolution makes it possible to evolve each tree using all data, and still obtain somewhat different models. The main idea is to use these quite accurate and slightly diverse trees to form an imaginary ensemble, which is then used as a guide when selecting one specific tree. Simply put, the tree classifying the largest number of instances identically to the ensemble is chosen. In the experimentation, using 25 UCI data sets, two selection strategies obtained significantly higher accuracy than the standard rule inducer J48.
  •  
Skapa referenser, mejla, bekava och länka
  • Resultat 1651-1660 av 3464
Typ av publikation
konferensbidrag (3464)
konstnärligt arbete (67)
Typ av innehåll
refereegranskat (2101)
övrigt vetenskapligt/konstnärligt (1335)
populärvet., debatt m.m. (28)
Författare/redaktör
Skrifvars, Mikael (104)
Wolmesjö, Maria, 196 ... (67)
Johansson, Ulf (66)
Cronholm, Stefan (63)
Taherzadeh, Mohammad ... (49)
Beach, Dennis (41)
visa fler...
Nierstrasz, Vincent (38)
Pal, Rudrajeet (37)
Nierstrasz, Vincent, ... (36)
Darányi, Sándor (36)
Taherzadeh, Mohammad (35)
Salomonson, Nicklas (35)
Wittek, Peter (34)
Persson, Nils-Kriste ... (32)
Seoane, Fernando, 19 ... (31)
Francke, Helena (31)
Cho, Sung-Woo (30)
Löfström, Tuve (30)
Sandsjö, Leif (30)
Holm, Ann-Sofie (29)
Ekebergh, Margaretha (27)
Lundberg, Lars (26)
Haider, Jutta (26)
Lind, Mikael (26)
Åkesson, Dan (25)
Ekwall, Daniel (25)
Guo, Li (25)
König, Rikard (25)
Nelhans, Gustaf, 197 ... (24)
Boström, Henrik (24)
Sundler, Annelie Joh ... (24)
Herlitz, Johan (23)
Martinovski, Bilyana (23)
Berglin, Lena (23)
Lindecrantz, Kaj (22)
Bolton, Kim (22)
Beach, Dennis, 1956 (22)
Evertsson, Magnus, 1 ... (22)
Bengtsson, Magnus, 1 ... (22)
Lindberg, Elisabeth (22)
Mauléon, Christina, ... (22)
Karimi, Keikhosro (21)
Bremer, Anders, 1957 ... (21)
Player-Koro, Catarin ... (21)
Ericsson, Dag (21)
Göbel, Hannes, 1976- (21)
Maurin Söderholm, Ha ... (20)
Langelotz, Lill, 196 ... (20)
Maceviciute, Elena (20)
Dahlström, Mats (20)
visa färre...
Lärosäte
Jönköping University (137)
Göteborgs universitet (129)
Chalmers tekniska högskola (100)
Linköpings universitet (77)
Kungliga Tekniska Högskolan (76)
visa fler...
Högskolan i Skövde (54)
Linnéuniversitetet (35)
Lunds universitet (32)
Stockholms universitet (29)
Karlstads universitet (24)
RISE (21)
Malmö universitet (18)
Örebro universitet (16)
Umeå universitet (15)
Mittuniversitetet (14)
Högskolan i Halmstad (12)
Karolinska Institutet (11)
Uppsala universitet (6)
Mälardalens universitet (6)
Luleå tekniska universitet (5)
Högskolan i Gävle (5)
Sveriges Lantbruksuniversitet (3)
VTI - Statens väg- och transportforskningsinstitut (3)
Högskolan Kristianstad (2)
Högskolan Väst (2)
Konstfack (1)
Högskolan Dalarna (1)
visa färre...
Språk
Engelska (2754)
Svenska (686)
Tyska (11)
Norska (4)
Danska (2)
Spanska (2)
visa fler...
Portugisiska (2)
Finska (1)
Tjeckiska (1)
Ungerska (1)
visa färre...
Forskningsämne (UKÄ/SCB)
Samhällsvetenskap (1695)
Teknik (583)
Naturvetenskap (525)
Medicin och hälsovetenskap (335)
Humaniora (290)
Lantbruksvetenskap (4)

År

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Stäng

Kopiera och spara länken för att återkomma till aktuell vy