Deep Learning - Ensemble

Ensemble

Posted by Rico's Nerd Cluster on March 3, 2022

Ensemble

An ensemble is a group of models (a.k.a base learners, weak learners) that are trained and combined to have better prediction, increased stability, and improved generalization compared to individual models.

Methods

Bagging (Bootstrap Aggregating)

Train multiple models independently using random subsets of data. Example: Random Forest. Advantage: reduces variance and helps prevent overfitting

Boosting

Train models sequentially, each tries to minimize error from the previous one. Example: Adaboost, Gradient Boosting machines like XGBoost. Advantages:?

Stacking

Stacked generatlization creates a combined “meta-model” to learn how to best combine their predicitons?

Voting and Average

For classification, use majority voting. for regression, average the predictions