Skip to content

spec : fix vocab compat checks in spec example#22426

Merged
ggerganov merged 1 commit intoggml-org:masterfrom
petersid2022:example-spec-fix-vocab-compat
Apr 30, 2026
Merged

spec : fix vocab compat checks in spec example#22426
ggerganov merged 1 commit intoggml-org:masterfrom
petersid2022:example-spec-fix-vocab-compat

Conversation

@petersid2022
Copy link
Copy Markdown
Contributor

Overview

  • port spec : fix vocab compat checks #22358 PR to examples/speculative/speculative.cpp
  • use vocab_[tgt,dft] instead of ctx_[tgt,dft] when logging on draft model / target model vocabulary mismatch

Additional information

Requirements

@petersid2022 petersid2022 requested a review from ggerganov as a code owner April 27, 2026 08:26
@ggml-gh-bot
Copy link
Copy Markdown

ggml-gh-bot Bot commented Apr 27, 2026

Hi @petersid2022, thanks for your contribution!

Per our contribution guidelines, the automated PR checker found the following issue(s) that need your attention:

  • Multiple open PRs from a new contributor: We limit new contributors (those without a previously merged PR) to 1 open PR at a time. You currently have 2 open PRs.

Please note that maintainers reserve the right to make final decisions on PRs. If you believe there is a mistake, please comment below.

* port ggml-org#22358 PR to examples/speculative/speculative.cpp
* use vocab_[tgt,dft] instead of ctx_[tgt,dft] when logging on draft
  model / target model vocabulary mismatch
@petersid2022 petersid2022 force-pushed the example-spec-fix-vocab-compat branch from 36caed3 to 62c5df8 Compare April 29, 2026 10:41
@petersid2022
Copy link
Copy Markdown
Contributor Author

@ggerganov kind reminder for a code review of this PR, that ports the vocab compatibility check to the examples/speculative.cpp module. thanks in advance
BR,
Peter Sideris

@ggerganov ggerganov added the merge ready A maintainer can use this label to indicate that they consider the changes final and ready to merge. label Apr 29, 2026
@ggerganov ggerganov merged commit b42c7fa into ggml-org:master Apr 30, 2026
46 checks passed
tekintian added a commit to tekintian/llama.cpp that referenced this pull request May 1, 2026
* 'master' of github.com:tekintian/llama.cpp: (659 commits)
  ggml-webgpu: Improve performance of mat-vec and mat-mat for MUL_MAT_ID (ggml-org#22464)
  Update llama-mmap to use ftello/fseeko (ggml-org#22497)
  common : check for null getpwuid in hf-cache (ggml-org#22550)
  vulkan: add get/set tensor 2d functions (ggml-org#22514)
  spec: fix argument typo (ggml-org#22552)
  ci : bump ty to 0.0.33 (ggml-org#22535)
  vendor : update cpp-httplib to 0.43.2 (ggml-org#22548)
  CUDA: fix tile FA kernel on Pascal (ggml-org#22541)
  scripts : add wc2wt.sh - create worktree from current HEAD (ggml-org#22513)
  add fast matmul iquants (ggml-org#22504)
  spec : fix draft model checkpoints (ggml-org#22521)
  spec : fix vocab compat checks in spec example (ggml-org#22426)
  common : do not pass prompt tokens to reasoning budget sampler (ggml-org#22488)
  hexagon: make vmem and buffer-size configurable (ggml-org#22487)
  CUDA: fuse SSM_CONV + ADD(bias) + SILU (ggml-org#22478)
  spec : disacard last drafted token with low prob (ggml-org#22506)
  sync : ggml
  ggml : bump version to 0.10.1 (ggml/1469)
  webui: fix slow mic stop and WAV encode (ggml-org#22480)
  ggml-cpu : disable tiled matmul on AIX to fix page boundary segfault (ggml-org#22293)
  ...

# Conflicts:
#	.gitignore
rsenthilkumar6 pushed a commit to rsenthilkumar6/llama.cpp that referenced this pull request May 1, 2026
* port ggml-org#22358 PR to examples/speculative/speculative.cpp
* use vocab_[tgt,dft] instead of ctx_[tgt,dft] when logging on draft
  model / target model vocabulary mismatch

Co-authored-by: Petros Sideris <petros.sideris@nokia.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

examples merge ready A maintainer can use this label to indicate that they consider the changes final and ready to merge.

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants