Skip to content

llama : avoid double token-to-piece cache#7654

Merged
ggerganov merged 1 commit intomasterfrom
gg/cache-no-special
Jun 3, 2024
Merged

llama : avoid double token-to-piece cache#7654
ggerganov merged 1 commit intomasterfrom
gg/cache-no-special

Conversation

@ggerganov
Copy link
Copy Markdown
Member

@mofosyne mofosyne added the Review Complexity : Medium Generally require more time to grok but manageable by beginner to medium expertise level label May 31, 2024
@ggerganov ggerganov merged commit 549279d into master Jun 3, 2024
@ggerganov ggerganov deleted the gg/cache-no-special branch June 3, 2024 05:34
Seunghhon pushed a commit to Seunghhon/llama.cpp that referenced this pull request Apr 26, 2026
phuongncn pushed a commit to phuongncn/llama.cpp-gx10-dgx-sparks-deepseekv4 that referenced this pull request Apr 28, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

Review Complexity : Medium Generally require more time to grok but manageable by beginner to medium expertise level

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants