EcstaticStruggle
EcstaticStruggle t1_jc1jts4 wrote
Reply to [D] Simple Questions Thread by AutoModerator
How do you combine hyper parameter optimization with early stopping in cross-validation for LightGBM?
Do you:
- Use the same validation set for hyperparameter performance estimation as well as early stopping evaluation (e.g., 80% training, 20% early stopping + validation set)
- Create a separate fold within cross-validation for early stopping evaluation. (e.g. 80%, 10%, 10% training, early stopping, validation set)
- Set aside a different dataset altogether (like a test set) which is constantly used for early stopping across different cross-validation folds for early stopping evaluation.
In the case of 1) and 2), how would you use early stopping once you identified optimal hyperparameters? Normally, you would re-fit on the entire dataset with the best hyperparameters, but this removes the early stopping data.
EcstaticStruggle t1_jcthdzz wrote
Reply to comment by josejo9423 in [D] Simple Questions Thread by AutoModerator
Thanks. This was something I tried earlier. I noticed that using the maximum number of estimators almost always lead to the highest cross validation score. I was worried there would be some overfitting as a result.