Add compressed-tensors format export support for W4A16 and W8A16 #1669
Merged
Add compressed-tensors format export support for W4A16 and W8A16 #1669
Conversation
# Please enter a commit message to explain why this merge is necessary, # especially if it merges an updated upstream into a topic branch. # # Lines starting with '#' will be ignored, and an empty message aborts # the commit.
for more information, see https://pre-commit.ci
Contributor
There was a problem hiding this comment.
Pull request overview
Adds llm_compressor (compressed-tensors) export support for INT weight-only schemes (W4A16, W8A16), and updates docs/tests accordingly.
Changes:
- Extend
llm_compressorformat to accept W4A16/W8A16 and route them through a new backend path. - Update compressed-tensors scheme construction to omit activation quantization for weight-only exports.
- Add/adjust CPU export tests and document the newly supported schemes (EN + CN).
Reviewed changes
Copilot reviewed 6 out of 6 changed files in this pull request and generated 4 comments.
Show a summary per file
| File | Description |
|---|---|
auto_round/formats.py |
Adds W4A16/W8A16 to llm_compressor support and introduces a WOQ backend selector (wint_a16). |
auto_round/export/export_to_llmcompressor/export.py |
Treats W*A16 as weight-only in compressed-tensors scheme creation; tightens dependency expectations around compress_module. |
auto_round/compressors/utils.py |
Adds helper to detect integer weight-only quantization (WOQ). |
test/test_cpu/export/test_export.py |
Refactors INT8_W8A8 export test and adds new W4A16/W8A16 llm_compressor export assertions. |
README.md |
Documents llm_compressor support for FP8_BLOCK, INT8_W8A8, W4A16, W8A16. |
README_CN.md |
Mirrors the README support-matrix update in Chinese. |
wenhuach21
approved these changes
Apr 9, 2026
yiliu30
approved these changes
Apr 9, 2026
…into thuang6/int4-ct
for more information, see https://pre-commit.ci
…into thuang6/int4-ct
Co-authored-by: Yi Liu <yi4.liu@intel.com>
Contributor
Author
|
/azp run Unit-Test-CUDA-AutoRound |
|
Azure Pipelines successfully started running 1 pipeline(s). |
Contributor
Author
|
/azp run Unit-Test-CUDA-AutoRound |
|
Azure Pipelines successfully started running 1 pipeline(s). |
xin3he
approved these changes
Apr 13, 2026
Contributor
Author
|
/azp run Unit-Test-CUDA-AutoRound |
|
Azure Pipelines successfully started running 1 pipeline(s). |
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
Description
Added compressed-tensors format export support for W4A16 and W8A16,
Replaced previous INT W8A8 support from internal NaiveQuantizationCompressor interface with new compress_module interface (require >=0.15.0)
updated PR to use BaseCompressor class method to be compatiable with old version
Type of Change
Related Issues
Fixes or relates to #1567
Checklist Before Submitting