Classification Ensembles
A classification ensemble is a predictive model composed of a weighted combination of multiple classification models. In general, combining multiple classification models increases predictive performance.
To explore classification ensembles interactively, use the Classification
      Learner app. For greater flexibility, use fitcensemble in the
     command-line interface to boost or bag classification trees, or to grow a random
     forest.
     For details on all supported ensembles, see Ensemble Algorithms. To reduce a multiclass problem into an ensemble of binary
     classification problems, train an error-correcting output codes (ECOC) model. For details, see
      fitcecoc.
To boost regression trees using LSBoost, or to grow a random forest of regression trees, see Regression Ensembles.
Apps
| Classification Learner | Train models to classify data using supervised machine learning | 
Blocks
| ClassificationEnsemble Predict | Classify observations using ensemble of decision trees (Since R2021a) | 
| ClassificationECOC Predict | Classify observations using error-correcting output codes (ECOC) classification model (Since R2023a) | 
| IncrementalClassificationECOC Predict | Classify observations using incremental ECOC classification model (Since R2024a) | 
Functions
Objects
Classes
Topics
- Framework for Ensemble LearningObtain highly accurate predictions by using many weak learners. 
- Ensemble AlgorithmsLearn about different algorithms for ensemble learning. 
- Train Classification EnsembleTrain a simple classification ensemble. 
- Test Ensemble QualityLearn methods to evaluate the predictive quality of an ensemble. 
- Handle Imbalanced Data or Unequal Misclassification Costs in Classification EnsemblesLearn how to set prior class probabilities and misclassification costs. 
- Classification with Imbalanced DataUse the RUSBoost algorithm for classification when one or more classes are over-represented in your data. 
- LPBoost and TotalBoost for Small EnsemblesCreate small ensembles by using the LPBoost and TotalBoost algorithms. 
- Tune RobustBoostTune RobustBoost parameters for better predictive accuracy. 
- Surrogate SplitsGain better predictions when you have missing data by using surrogate splits. 
- Train Classification Ensemble in ParallelTrain a bagged ensemble in parallel reproducibly. 
- Bootstrap Aggregation (Bagging) of Classification Trees Using TreeBaggerCreate a TreeBaggerensemble for classification.
- Credit Rating by Bagging Decision TreesThis example shows how to build an automated credit rating tool. 
- Random Subspace ClassificationIncrease the accuracy of classification by using a random subspace ensemble. 
- Predict Class Labels Using ClassificationEnsemble Predict BlockTrain a classification ensemble model with optimal hyperparameters, and then use the ClassificationEnsemble Predict block for label prediction. 
- Predict Class Labels Using ClassificationECOC Predict BlockTrain an ECOC classification model, and then use the ClassificationECOC Predict block for label prediction. (Since R2023a)