Learn Before
Concept

Random Forests vs. Bagging with Decision Trees

Random forests is quite similar to bagging with decision trees except that there is an improvement on spliting trees. They both use bootstrapped subset of training oberservations for a tree and repeated that process several times for the average/most common result.

The difference lies in that, random forest only use randomly selected subsets of all predictors for spliting trees. This improvement decorrelates the trees and solves the problem of bagging that, if there is a very strong predictor, most bagged trees might look similar and averaging could not reduce much variance for the highly correlated quantities. However, this can be addressed by decorrelating the trees thus reducing the variance and making the trees more reliable.

0

5

Updated 2020-03-07

Tags

Data Science