Ensemble methods are classifiers that have classifiers as input (and often have higher accuracy than regular input classifiers). This is also called “meta-learning” and it only works if the individual classifiers make independent errors
See also: emergent behaviour
Boosting/Stacking
Improves training error of classifiers with high
Models that use the boosting ensemble method:
- XGBoost (regularized regression trees)
Averaging/Voting
Improves approximation error of classifiers with high
Models that uses the averaging ensemble method:
Methods
- Voting: take the mode of the predictions across the classifiers
- Stacking: fit another classifier that uses the predictions as features