Hi,
i have been wondering for some time now how random forests (Bagging, or AdaBoost, doesn't matter) are built when using cross-validation.
Let's see we're using 5-fold cross validation to train random forests on 5 different training sets and therefore test on 5 different test sets.
How does the 'final' random forest look like when we are basically building 5 random forests (one for each fold of the cross validation). How are these forests combined into a final model?
I have never understood this step and I really hope someone can help me with this!
thanks in advance,
Steven

 채택된 답변

Ayush Aniket
Ayush Aniket 2025년 6월 10일

0 개 추천

The models used in cross validation (5, as you mentioned) are not directly combined into a single final model.
The goal of cross-validation is to evaluate the model's performance across different data splits, ensuring it generalizes well. After cross-validation, you discard the five individual models.You train a new random forest on the entire dataset using the best hyperparameters found during cross-validation. This final model is used for predictions.
Refer the following example to read the workflow of training a final model after evaluating and finding hyperparameters through cross-validation: https://www.mathworks.com/help/stats/classificationsvm.crossval.html#mw_e9fd437d-4125-4b9e-b87c-f370bc439a3e

추가 답변 (0개)

카테고리

도움말 센터File Exchange에서 Statistics and Machine Learning Toolbox에 대해 자세히 알아보기

질문:

2020년 7월 25일

답변:

2025년 6월 10일

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by