-
Notifications
You must be signed in to change notification settings - Fork 3.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Accelerated Gradient Boosting #1257
Comments
Hmm I am not sure wether this optimization makes sense for non-standard gradient boosting algorithms (such as LightGBM and XGBoost), where you already compute the Hessian in adition to the Gradient... Theoretically, the added benefit of computing the Hessian of the loss function should be essentialy equivalent to a method such as Nesterov's accelerated gradient (since it is just a way to provide further information to the optimization technique, in order to make de descent more efficient)... It looks like it is really an improvement over the original Friedman's gradient boosting, but with "Hessian/Newtonian boosting" that just might not be the case. It would be really cool to try it out and see if it makes learning faster, but a lot should have to be changed in LightGBM's source code in order to implement this. |
Closed in favor of being in #2302. We decided to keep all feature requests in one place. Welcome to contribute this feature! Please re-open this issue (or post a comment if you are not a topic starter) if you are actively working on implementing this feature. |
See also Lu et. al. (2020) and specifically their discussion in section 6, which seems cognizant of some of the observations here by @julioasotodv. |
I just read a cool paper on arxiv and thought it might be of interest to the LightGBM team: https://arxiv.org/pdf/1803.02042.pdf
The text was updated successfully, but these errors were encountered: