Skip to content

add gemma4#2663

Merged
Qubitium merged 4 commits intomainfrom
gemma4
Apr 2, 2026
Merged

add gemma4#2663
Qubitium merged 4 commits intomainfrom
gemma4

Conversation

@Qubitium
Copy link
Copy Markdown
Collaborator

@Qubitium Qubitium commented Apr 2, 2026

No description provided.

@Qubitium Qubitium merged commit 25dd8ea into main Apr 2, 2026
5 checks passed
@Qubitium Qubitium deleted the gemma4 branch April 2, 2026 22:53
# The local env does not ship Marlin runtime kernels, so validation reloads must stay on Torch.
LOAD_BACKEND = BACKEND.TORCH
# Gemma 4 full-attention layers expand to 512-dim heads, which FlashAttention cannot execute.
USE_FLASH_ATTN = False
Copy link
Copy Markdown
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Gemma4 does not work with FA2. Is this a HF model upload issue due to the Gemma4 variants or Transformers just made up some bad unit tests?

huggingface/transformers#45202

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant