نتایج جستجو برای: bagging model

تعداد نتایج: 2105681  

2010
Lei Zhang Guiquan Liu Xuechen Zhang Song Jiang Enhong Chen

Storage device performance prediction is a key element of self-managed storage systems and application planning tasks, such as data assignment and configuration. Based on bagging ensemble, we proposed an algorithm named selective bagging classification and regression tree (SBCART) to model storage device performance. In addition, we consider the caching effect as a feature in workload character...

2014
Daniel Gianola Kent A. Weigel Nicole Krämer Alessandra Stella Chris-Carolin Schön

We examined whether or not the predictive ability of genomic best linear unbiased prediction (GBLUP) could be improved via a resampling method used in machine learning: bootstrap aggregating sampling ("bagging"). In theory, bagging can be useful when the predictor has large variance or when the number of markers is much larger than sample size, preventing effective regularization. After present...

2005
Sotiris B. Kotsiantis George E. Tsekouras Panayiotis E. Pintelas

Structurally, a model tree is a regression method that takes the form of a decision tree with linear regression functions instead of terminal class values at its leaves. In this study, model trees are coupled with bagging for solving classification problems. In order to apply this regression technique to classification problems, we consider the conditional class probability function and seek a ...

2011
Guohua Liang Xingquan Zhu Chengqi Zhang

Bagging is a simple, yet effective design which combines multiple base learners to form an ensemble for prediction. Despite its popular usage in many real-world applications, existing research is mainly concerned with studying unstable learners as the key to ensure the performance gain of a bagging predictor, with many key factors remaining unclear. For example, it is not clear when a bagging p...

Journal: :Neural Computation 1997
Michiaki Taniguchi Volker Tresp

We compare the performance of averaged regularized estimators. We show that the improvement in performance which can be achieved by averaging depends critically on the degree of regularization which is used in training the individual estimators. We compare four different averaging approaches: simple averaging, bagging, variance-based weighting and variance-based bagging. In any of the averaging...

2012
Sotiris B. Kotsiantis

Bagging and boosting are among the most popular resampling ensemble methods that generate and combine a diversity of regression models using the same learning algorithm as base-learner. Boosting algorithms are considered stronger than bagging on noisefree data. However, there are strong empirical indications that bagging is much more robust than boosting in noisy settings. For this reason, in t...

Journal: :Computational Statistics & Data Analysis 2007

Journal: :The Annals of Statistics 2002

2005
Alípio Mário Jorge Paulo J. Azevedo

In this paper we study a new technique we call post-bagging, which consists in resampling parts of a classification model rather then the data. We do this with a particular kind of model: large sets of classification association rules, and in combination with ordinary best rule and weighted voting approaches. We empirically evaluate the effects of the technique in terms of classification accura...

Journal: :Pattern Recognition Letters 2007
Gonzalo Martínez-Muñoz Alberto Suárez

Boosting is used to determine the order in which classifiers are aggregated in a bagging ensemble. Early stopping in the aggregation of the classifiers in the ordered bagging ensemble allows the identification of subensembles that require less memory for storage, have a faster classification speed and can perform better than the original bagging ensemble. Furthermore, ensemble pruning does not ...

نمودار تعداد نتایج جستجو در هر سال

با کلیک روی نمودار نتایج را به سال انتشار فیلتر کنید