Skip to content

Fix Qwen2IntegrationTest#45268

Merged
ydshieh merged 1 commit intomainfrom
fix_qwen2
Apr 6, 2026
Merged

Fix Qwen2IntegrationTest#45268
ydshieh merged 1 commit intomainfrom
fix_qwen2

Conversation

@ydshieh
Copy link
Copy Markdown
Collaborator

@ydshieh ydshieh commented Apr 6, 2026

What does this PR do?

Fix Qwen2IntegrationTest:

@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented Apr 6, 2026

[For maintainers] Suggested jobs to run (before merge)

run-slow: qwen2

@ydshieh
Copy link
Copy Markdown
Collaborator Author

ydshieh commented Apr 6, 2026

run-slow: qwen2

@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented Apr 6, 2026

Workflow Run ⚙️

This comment contains run-slow, running the specified jobs:

models: ["models/qwen2"]
quantizations: []

@HuggingFaceDocBuilderDev
Copy link
Copy Markdown

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

@github-actions
Copy link
Copy Markdown
Contributor

github-actions Bot commented Apr 6, 2026

CI Results

Workflow Run ⚙️

Commit Info

Context Commit Description
RUN 523a90be workflow commit (merge commit)
PR b831d8b4 branch commit (from PR)
main 5182cf38 base commit (on main)

✅ No failing test specific to this PR 🎉 👏 !

# Expected mean on dim = -1
EXPECTED_MEAN = Expectations({
(None, None): torch.tensor([[-2.2121, -1.6335, -1.4816, -1.5035, -1.9110, -1.8979, -1.9682, -2.1980]]),
(None, None): torch.tensor([[-2.2463, -1.6463, -1.4748, -1.4913, -1.9213, -1.9016, -1.9969, -2.1761]]),
Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@Abdennacer-Badaoui Maybe this will fail on AMD, if you want to double check. Although maybe I should use ("cuda", 8) and ("rocm", 9). But let's have double check anyway.

# slicing logits[0, 0, 0:30]
EXPECTED_SLICE = Expectations({
(None, None): torch.tensor([2.7344, 4.2812, 4.1562, 2.3906, 1.1875, 2.1562, 3.1719, 3.1406, 1.2891, 3.6094, 3.3125, 1.8203, 2.9219, 3.2344, 1.5938, 6.2500, 7.4062, 7.2188, 6.5938, 6.0312, 6.1562, 5.3750, 5.9688, 5.5938, 6.1250, 1.2656, 1.6016, 3.4062, 1.7891, 3.6406]),
(None, None): torch.tensor([2.6406, 4.3125, 3.9531, 2.2656, 1.0000, 2.0312, 3.2344, 2.9219, 1.1953, 3.3750, 3.1875, 2.0156, 2.8281, 3.2656, 1.4453, 6.0625, 7.2188, 7.0312, 6.4375, 5.8750, 6.0312, 5.2500, 5.8438, 5.5000, 5.9688, 1.2734, 1.5312, 3.2344, 1.6406, 3.4375]),
Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@Abdennacer-Badaoui Maybe this will fail on AMD, if you want to double check. Although maybe I should use ("cuda", 8) and ("rocm", 9). But let's have double check anyway.

@ydshieh ydshieh merged commit 02c38f7 into main Apr 6, 2026
23 of 25 checks passed
@ydshieh ydshieh deleted the fix_qwen2 branch April 6, 2026 19:18
louzongzhi pushed a commit to louzongzhi/transformers that referenced this pull request Apr 7, 2026
fix

Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
@Sai-Suraj-27
Copy link
Copy Markdown
Contributor

Fix Qwen2IntegrationTest:

Thanks for pinging @ydshieh. I don't have a fixed GPU, i use available cloud gpu's, I remember using A10 only that day, but I don't recall for sure. I just tested with A10 and it's giving the expected output that's fixed here. So, this should be the correct expected text.

@tarekziade
Copy link
Copy Markdown
Collaborator

Thanks for the fix! for #43794 I did run the GPU tests but it looks like the result was a false positive

sirzechs66 pushed a commit to sirzechs66/transformers that referenced this pull request Apr 18, 2026
fix

Co-authored-by: ydshieh <ydshieh@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

4 participants