Skip to content

llama : fix compatibility with old 2 expert models#6735

Merged
ggerganov merged 1 commit intomasterfrom
sl/moe-extra-tensors-fix
Apr 18, 2024
Merged

llama : fix compatibility with old 2 expert models#6735
ggerganov merged 1 commit intomasterfrom
sl/moe-extra-tensors-fix

Conversation

@slaren
Copy link
Copy Markdown
Member

@slaren slaren commented Apr 18, 2024

The correct number of extra tensors needed is 3 per layer, not n_expert, so it wouldn't allocate enough tensors for models with 2 experts.

Fixes the second issue reported in #6379

@github-actions

This comment has been minimized.

@ggerganov ggerganov merged commit c71bfd7 into master Apr 18, 2024
@slaren slaren deleted the sl/moe-extra-tensors-fix branch April 18, 2024 11:08
Seunghhon pushed a commit to Seunghhon/llama.cpp that referenced this pull request Apr 26, 2026
phuongncn pushed a commit to phuongncn/llama.cpp-gx10-dgx-sparks-deepseekv4 that referenced this pull request Apr 28, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants