Skip to content

Remove CompressedLinear support for compressed-tensors > 0.13#43747

Open
colldata79 wants to merge 2 commits intohuggingface:mainfrom
colldata79:ct-remove-compressed-linear-support
Open

Remove CompressedLinear support for compressed-tensors > 0.13#43747
colldata79 wants to merge 2 commits intohuggingface:mainfrom
colldata79:ct-remove-compressed-linear-support

Conversation

@colldata79
Copy link
Copy Markdown

@colldata79 colldata79 commented Feb 4, 2026

Title: Remove CompressedLinear support for compressed-tensors > 0.13

Body:

What does this PR do?

Prepares transformers for the removal of CompressedLinear from compressed-tensors (v0.14+). Users should now call model.dequantize() after loading to decompress quantized models.

Changes:

  • Stop passing run_compressed to apply_quantization_config
  • Always decompress models after loading for compressed-tensors > 0.13
  • Add _dequantize method to CompressedTensorsHfQuantizer
  • Remove tests that reference deleted CompressedLinear class (test_default_run_compressed__True, test_default_run_compressed__False)

Backwards compatibility is maintained for compressed-tensors ≤ 0.13.

Related to vllm-project/llm-compressor#2279
Companion PR: vllm-project/compressed-tensors#564

Before submitting

Who can review?

@SunMarc @MekkCyber (quantization)

- Stop passing run_compressed to apply_quantization_config
- Always decompress models after loading for CT > 0.13
- Add _dequantize method to CompressedTensorsHfQuantizer
- Remove tests that reference deleted CompressedLinear class

Related to vllm-project/llm-compressor#2279
@colldata79
Copy link
Copy Markdown
Author

CI failure in test_processing_layoutlmv2.py is unrelated to this PR - no tokenizer/processor code was modified

can a maintainer please verify

@SunMarc
Copy link
Copy Markdown
Member

SunMarc commented Feb 5, 2026

cc @dsikka

Copy link
Copy Markdown
Contributor

@kylesayrs kylesayrs left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

A couple changes are needed, but overall this is the correct approach, thanks for the changes!

@colldata79 Please let me know when all the changes are made, and I'll test them e2e, then we can ask for approval to merge this in.

Comment thread src/transformers/quantizers/quantizer_compressed_tensors.py Outdated
Comment thread src/transformers/quantizers/quantizer_compressed_tensors.py Outdated
@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented Feb 5, 2026

[For maintainers] Suggested jobs to run (before merge)

run-slow: compressed_tensors_integration

…store and add tests

Signed-off-by: Your Name <your.email@example.com>
@colldata79 colldata79 force-pushed the ct-remove-compressed-linear-support branch from c561daf to 5d76c36 Compare February 5, 2026 18:19
@colldata79
Copy link
Copy Markdown
Author

colldata79 commented Feb 5, 2026

@kylesayrs ,Fix is in place . please go ahead with the end to end test

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants