-
Notifications
You must be signed in to change notification settings - Fork 3.8k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Quantized training fails when a model is too complex #5982
Comments
Can confirm this bug with quantized gradient True and 1000 boosting rounds. |
@jameslamb , @nigimitama |
@mglowacki100 Just pushed a fix. Please check #6092. |
* fix leaf splits update after split in quantized training * fix preparation ordered gradients for quantized training * remove force_row_wise in distributed test for quantized training * Update src/treelearner/leaf_splits.hpp --------- Co-authored-by: James Lamb <[email protected]>
@shiyu1994 @jameslamb |
@mglowacki100 |
@AlexGrunt , yeah, it is non-deterministic issue, I've luck and it worked on the first time, but if you re-run the cell a couple times it'll will randomly fail/succeed. |
@mglowacki100 So the only way to try gradient quantization is to wait for next python-package release? |
@shiyu1994 sorry for bothering you with noob question, is there a way to test it on google colab?
but in the end I get:
|
@AlexGrunt We will release more frequently in the future. The next release should come soon perhaps within 1 month. @mglowacki100 It seems that the path |
@shiyu1994 Thanks for tip! I've managed to build and install lightgbm on google colab. The problem was that google collab doesn't behave like terminal and
should be 4.1.0.99 Here is google colab, btw. I've made crude test with repeating 100 times training with different seeds and there is no problem: |
I don't feel we should take on documentation in the Colab is running IPython notebooks with Jupyter, and that syntax is well covered in the IPython documentation: https://ipython.org/ipython-doc/3/interactive/shell.html The LightGBM-specific sequence of commands are already covered by the docs at https://lightgbm.readthedocs.io/en/latest/Installation-Guide.html. |
Excited for the fix to be available, definitely want to try this out for AutoGluon v1.0 release! |
…5994) (microsoft#6092) * fix leaf splits update after split in quantized training * fix preparation ordered gradients for quantized training * remove force_row_wise in distributed test for quantized training * Update src/treelearner/leaf_splits.hpp --------- Co-authored-by: James Lamb <[email protected]>
Description
When a model is too complex and
use_quantized_grad
isTrue
, training fails with this error:Reproducible example
Environment info
LightGBM version or commit hash: 4.0.0
Command(s) you used to install LightGBM
I used this dockerfile
(
main.py
is the python script above)Additional Comments
When a model is not too complex (e.g.
num_boost_round=10
), training successfully finished even ifuse_quantized_grad
isTrue
.The text was updated successfully, but these errors were encountered: