Weighted Bagging in Decision Trees: Data Mining
Open Access
- 1 October 2020
- journal article
- Published by Yayasan Ahmar Cendekia Indonesia in JINAV: Journal of Information and Visualization
- Vol. 1 (1), 1-14
- https://doi.org/10.35877/454ri.jinav149
Abstract
The main focus of this paper is on the use of resampling techniques to construct predictive models from data and the goal is to identify the best possible model which can produce better predications. Bagging or Bootstrap aggregating is a general method for improving the performance of given learning algorithm by using a majority vote to combine multiple classifier outputs derived from a single classifier on a bootstrap resample version of a training set. A bootstrap sample is generated by a random sample with replacement from the original training set. Inspired by the idea of bagging, we present an improved method based on a distance function in decision trees, called modified bagging (or weighted Bagging) in this study. The experimental results show that modified bagging is superior to the usual majority vote. These results are confirmed by both real data and artificial data sets with random noise. The Modified bagged classifier performs significantly better than usual bagging on various tree levels for all sample sizes. An interesting observation is that the weighted bagging performs somewhat better than usual bagging with sumps.Keywords
This publication has 5 references indexed in Scilit:
- Bootstrap Methods and their ApplicationPublished by Cambridge University Press (CUP) ,1997
- Bagging predictorsMachine Learning, 1996
- An Introduction to the BootstrapPublished by Springer Science and Business Media LLC ,1993
- The strength of weak learnabilityMachine Learning, 1990
- THE USE OF MULTIPLE MEASUREMENTS IN TAXONOMIC PROBLEMSAnnals of Eugenics, 1936