Skip to content

Support new model Qwen/Qwen3.6-35B-A3B#1705

Merged
lvliang-intel merged 2 commits intomainfrom
lvl/support_qwen3.6
Apr 22, 2026
Merged

Support new model Qwen/Qwen3.6-35B-A3B#1705
lvliang-intel merged 2 commits intomainfrom
lvl/support_qwen3.6

Conversation

@lvliang-intel
Copy link
Copy Markdown
Contributor

Description

Support new model Qwen/Qwen3.6-35B-A3B.
Qwen/Qwen3.6-35B-A3B still use the same model type with Qwen/Qwen3.5-35B-A3B.
This PR adds compatibility for Qwen/Qwen3.6-35B-A3B in the AutoRound HF conversion path.

Type of Change

  • Bug fix
  • New feature
  • Documentation update
  • Performance improvement
  • Code refactoring
  • Other (please specify):

Related Issues

Fixes or relates to #

Checklist Before Submitting

  • My code has been tested locally.
  • Documentation has been updated as needed.
  • New or updated tests are included where applicable.

Signed-off-by: lvliang-intel <liang1.lv@intel.com>
Copilot AI review requested due to automatic review settings April 19, 2026 03:30
Copy link
Copy Markdown
Contributor

Copilot AI left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Pull request overview

Adds compatibility for the new HuggingFace model Qwen/Qwen3.6-35B-A3B by improving AutoRound’s HF conversion logic when a composite checkpoint is loaded as a text-only submodel and module path prefixes differ.

Changes:

  • Added _remap_paths_for_text_model() to remap quantization block paths using Transformers checkpoint conversion mappings.
  • Updated get_layer_config() to retry layer discovery with remapped paths when the initial block-path match finds zero layers.

Comment thread auto_round/inference/convert_model.py
Comment thread auto_round/inference/convert_model.py
@wenhuach21
Copy link
Copy Markdown
Contributor

Shall we add version control if it's related to Transformer bugs, so we can handle them cleanly if they’re fixed in future releases

@lvliang-intel
Copy link
Copy Markdown
Contributor Author

lvliang-intel commented Apr 19, 2026

Shall we add version control if it's related to Transformer bugs, so we can handle them cleanly if they’re fixed in future releases

This issue does not occur when using Qwen3_5MoeForConditionalGeneration.from_pretrained; it only appears with AutoModelForCausalLM.from_pretrained. So it's not a real issue? If users want to use AutoModelForCausalLM.from_pretrained to load Qwen/Qwen3.6-35B-A3B models, they need to this PR and transformers issue fix huggingface/transformers#45494.

@lvliang-intel
Copy link
Copy Markdown
Contributor Author

/azp run Unit-Test-CUDA-AutoRound

@azure-pipelines
Copy link
Copy Markdown

Azure Pipelines successfully started running 1 pipeline(s).

@lvliang-intel lvliang-intel merged commit 2dda035 into main Apr 22, 2026
42 checks passed
@lvliang-intel lvliang-intel deleted the lvl/support_qwen3.6 branch April 22, 2026 12:21
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants