2W questions
Bagging
Boosting
Random Forest
AdaBoost / Gradiant Boosting
Conclusion
Henry Lin
替代役
2W questions
Bagging
Boosting
Random Forest
AdaBoost / Gradiant Boosting
Conclusion
Bagging exploits that idea to address the
overfitting issuein a more fundamental manner. It was invented by Leo Breiman, who called it "bootstrap aggregating" or simply "bagging" (reference: "Bagging predictors," Machine Learning, 24:123-140, 1996, cited by 7466).
high variance and Bagging(Boostrap Aggregation) is a procedure to reduce variance.\[\hat{f}_{bag}(x) = \frac{1}{B}\sum_{B}^{b=1}\hat{f}^{*b}(x)\]

The motivation for boosting was a procedure that combines the outputs of many “weak” classifiers to produce a powerful “committee.”

Bagging reduce the variance
Boosting learned from the mistakes.
Trees are ideal candidates for bagging, since they can capture complex interaction structures in the data, and if grown sufficiently deep, have relatively low bias. Since trees are
notoriously noisy, they benefit greatly from theaveraging.

Boosting tree can be considered as a human learning process. Trees are just like human decision-making process and boosting is a way to learn by weak learner, which means boosting + tree is learned by mistakes.



Generally, Boosting has better performance than Bagging. However, Sometimes Random Forests have better performance.
The advantage of Random Forests is more simple to train and tune than boosting trees.