Thanks to visit codestin.com
Credit goes to github.com

Skip to content

Should we turn on early stopping in HistGradientBoosting by default? #14303

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
GaelVaroquaux opened this issue Jul 10, 2019 · 8 comments
Closed

Comments

@GaelVaroquaux
Copy link
Member

Using n_iter_no_change=5 (or 10) in HistGradientBoosting makes a huge difference in terms of speed for me, and it seems to be harmless (at a cursory looking).

While these models are still experimental, should we make this change?

@amueller
Copy link
Member

I thought it was on by default. Is it not in lightgbm?
@NicolasHug can you maybe check in with the lightgbm people why they didn't?

@NicolasHug
Copy link
Member

hmm I can ask them. But I'd be OK too with it being active by default.

@ogrisel
Copy link
Member

ogrisel commented Jul 11, 2019

I don't think it is in lightgbm.

My concern is more about the inconsistent default behavior in scikit-learn. Why early stopping would be on by default in gbrt and not in logistic regression for instance (besides the fact that is not implemented)?

But aside from that concern, +1 as well.

@GaelVaroquaux
Copy link
Member Author

GaelVaroquaux commented Jul 11, 2019 via email

@NicolasHug
Copy link
Member

I opened microsoft/LightGBM#2270

It is not enabled by default because they require the validation set to be passed to fit.

@NicolasHug
Copy link
Member

They want to give the user as much liberty as possible since the train/val split can be application specific.

We don't have this "problem": we always use train_test_split.

@GaelVaroquaux
Copy link
Member Author

GaelVaroquaux commented Jul 19, 2019 via email

@NicolasHug
Copy link
Member

closing in favor of #14503 :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants