From 0d9240020083471d182629ca7ddd49401d6a0fbe Mon Sep 17 00:00:00 2001 From: smty2018 <74114936+smty2018@users.noreply.github.com> Date: Wed, 23 Oct 2024 14:18:45 +0530 Subject: [PATCH] Included that retrain_full = True does not include the user provided validation data in the docs. #1228 (#1245) * Update Task-Oriented-AutoML.md * Update Task-Oriented-AutoML.md * Update marker * Fix format --------- Co-authored-by: Li Jiang --- website/docs/Use-Cases/Task-Oriented-AutoML.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/website/docs/Use-Cases/Task-Oriented-AutoML.md b/website/docs/Use-Cases/Task-Oriented-AutoML.md index 97d64010d..c51c6a3c1 100644 --- a/website/docs/Use-Cases/Task-Oriented-AutoML.md +++ b/website/docs/Use-Cases/Task-Oriented-AutoML.md @@ -393,7 +393,7 @@ For holdout, you can also set: - `split_ratio`: the fraction for validation data, 0.1 by default. - `X_val`, `y_val`: a separate validation dataset. When they are passed, the validation metrics will be computed against this given validation dataset. If they are not passed, then a validation dataset will be split from the training data and held out from training during the model search. After the model search, flaml will retrain the model with best configuration on the full training data. - You can set`retrain_full` to be `False` to skip the final retraining or "budget" to ask flaml to do its best to retrain within the time budget. + You can set`retrain_full` to be `False` to skip the final retraining or "budget" to ask flaml to do its best to retrain within the time budget. When `retrain_full` is set to `True`, the user-provided validation data is not used in the final retraining of the model. For cross validation, you can also set `n_splits` of the number of folds. By default it is 5.