-
Notifications
You must be signed in to change notification settings - Fork 39
Fix: Disable Sparse Decompression for Dense Compressors #237
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
mgoin
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Makes sense, thanks!
src/compressed_tensors/compressors/model_compressors/model_compressor.py
Outdated
Show resolved
Hide resolved
kylesayrs
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
src/compressed_tensors/compressors/model_compressors/model_compressor.py
Outdated
Show resolved
Hide resolved
ff2b26a to
51cee4e
Compare
dsikka
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
51cee4e to
cc4f78e
Compare
~~Contingent on merge of huggingface/transformers#34719 ~~ ^ has been merged not yet released ~~ ^ has been released Blocked on vllm-project/compressed-tensors#237 SUMMARY: * In multiple optimization tests, automatically decompress model if provided as optimized model * Fix recipe stage length * Revive old code * When running multiple optimizations (ex. oneshot then finetune, oneshot and oneshot), the recipes needs to be added to the session using `initialize_recipe`. Example here https://github.com/vllm-project/llm-compressor/pull/971/files#diff-c9ae8b3ad24d13abeea5b649a5fd6d0b0925f5c9cc40220cbfbe21ae81242f8dR63-R65 TEST PLAN: ran the test using transformers main Must pass tests/llmcompressor/transformers/obcq/test_consecutive_runs.py --------- Co-authored-by: Dipika Sikka <[email protected]> Co-authored-by: Rahul Tuli <[email protected]>
~~Contingent on merge of huggingface/transformers#34719 ~~ ^ has been merged not yet released ~~ ^ has been released Blocked on vllm-project/compressed-tensors#237 SUMMARY: * In multiple optimization tests, automatically decompress model if provided as optimized model * Fix recipe stage length * Revive old code * When running multiple optimizations (ex. oneshot then finetune, oneshot and oneshot), the recipes needs to be added to the session using `initialize_recipe`. Example here https://github.com/vllm-project/llm-compressor/pull/971/files#diff-c9ae8b3ad24d13abeea5b649a5fd6d0b0925f5c9cc40220cbfbe21ae81242f8dR63-R65 TEST PLAN: ran the test using transformers main Must pass tests/llmcompressor/transformers/obcq/test_consecutive_runs.py --------- Co-authored-by: Dipika Sikka <[email protected]> Co-authored-by: Rahul Tuli <[email protected]> Signed-off-by: Rahul Tuli <[email protected]>
…#237) * Turn off sparse decompression when sparse compressor is dense * Update: Condiition to use enum instead of raw string * Remove unnecesarry style change
…#237) * Turn off sparse decompression when sparse compressor is dense * Update: Condiition to use enum instead of raw string * Remove unnecesarry style change
Problem
When the sparse compressor is set to
"dense", sparse decompression is incorrectly triggered, causing uninitialized weights and downstream errors.Example CI failure: [GitHub Actions Log](https://github.com/vllm-project/llm-compressor/actions/runs/12659596814/job/35326229412).
Solution
Added a condition to skip sparse decompression when the sparsity configuration format is
"dense".Testing
llm-compressormain commit:03e21770.