-
Notifications
You must be signed in to change notification settings - Fork 23
Issues: mit-han-lab/deepcompressor
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
Rationale behind converting proj_out of FluxSingleTransformerBlock to ConcatLinear
question
Further information is requested
svdquant
#43
opened Jan 29, 2025 by
vinovo
'NoneType' object has no attribute 'name'
bug
Something isn't working
svdquant
#34
opened Dec 6, 2024 by
jhss
OOM when use deepcompressor quantize llama2 w4a8 per-group with H100 80G
#33
opened Dec 3, 2024 by
Andy0422
torch.OutOfMemoryError: CUDA out of memory
bug
Something isn't working
svdquant
#30
opened Nov 26, 2024 by
Lenan22
AttributeError: 'tuple' object has no attribute 'shape'
bug
Something isn't working
svdquant
#29
opened Nov 26, 2024 by
Lenan22
Quantized custom flux model was still bfloat16
enhancement
New feature or request
svdquant
#27
opened Nov 20, 2024 by
samedii
How to apply SVDQuant for SD3 model?
enhancement
New feature or request
svdquant
#26
opened Nov 14, 2024 by
wxsms
will you support quantize the embedding layer and lm_head layer?
#18
opened Sep 6, 2024 by
geqian-9192
[Bug] RuntimeError: Boolean value of Tensor with more than one value is ambiguous
#17
opened Aug 10, 2024 by
ChenMnZ
ProTip!
Add no:assignee to see everything that’s not assigned.