Skip to content

ggml-cpu : add check for ARM MATMUL_INT8/i8mm support#15922

Merged
ericcurtin merged 1 commit intoggml-org:masterfrom
danbev:ggml-cpu-arm-i8mm-feature-issue
Sep 11, 2025
Merged

ggml-cpu : add check for ARM MATMUL_INT8/i8mm support#15922
ericcurtin merged 1 commit intoggml-org:masterfrom
danbev:ggml-cpu-arm-i8mm-feature-issue

Conversation

@danbev
Copy link
Copy Markdown
Member

@danbev danbev commented Sep 10, 2025

This commit adds a check for GGML_MACHINE_SUPPORTS_i8mm when enabling MATMUL_INT8 features, ensuring that i8mm intrinsics are only used when the target hardware actually supports them.

The motivation for this is to fix ggml CI build failures where the feature detection correctly identifies that i8mm is not supported, adding the +noi8mm flag, but MATMUL_INT8 preprocessor definitions are still enabled, causing the compiler to attempt to use vmmlaq_s32 intrinsics without i8mm support.

Refs: https://github.com/ggml-org/ggml/actions/runs/17525174120/job/49909199499

This commit adds a check for GGML_MACHINE_SUPPORTS_i8mm when enabling
MATMUL_INT8 features, ensuring that i8mm intrinsics are only used when
the target hardware actually supports them.

The motivation for this is to fix ggml CI build failures where the
feature detection correctly identifies that i8mm is not supported,
adding the +noi8mm flag, but MATMUL_INT8 preprocessor definitions are
still enabled, causing the compiler to attempt to use vmmlaq_s32
intrinsics without i8mm support.

Refs: https://github.com/ggml-org/ggml/actions/runs/17525174120/job/49909199499
@github-actions github-actions Bot added the ggml changes relating to the ggml tensor library for machine learning label Sep 10, 2025
@danbev
Copy link
Copy Markdown
Member Author

danbev commented Sep 10, 2025

For some more details, this section contains some debugging information from the github runner machine.

@ericcurtin ericcurtin merged commit 24a6734 into ggml-org:master Sep 11, 2025
48 checks passed
@ggerganov
Copy link
Copy Markdown
Member

@ericcurtin Do not merge PRs of other collaborators. We've updated the contributing guidelines recently to emphasize this: #15881

@danbev danbev deleted the ggml-cpu-arm-i8mm-feature-issue branch September 24, 2025 06:18
blime4 referenced this pull request in blime4/llama.cpp Feb 5, 2026
This commit adds a check for GGML_MACHINE_SUPPORTS_i8mm when enabling
MATMUL_INT8 features, ensuring that i8mm intrinsics are only used when
the target hardware actually supports them.

The motivation for this is to fix ggml CI build failures where the
feature detection correctly identifies that i8mm is not supported,
adding the +noi8mm flag, but MATMUL_INT8 preprocessor definitions are
still enabled, causing the compiler to attempt to use vmmlaq_s32
intrinsics without i8mm support.

Refs: https://github.com/ggml-org/ggml/actions/runs/17525174120/job/49909199499
Seunghhon pushed a commit to Seunghhon/llama.cpp that referenced this pull request Apr 26, 2026
This commit adds a check for GGML_MACHINE_SUPPORTS_i8mm when enabling
MATMUL_INT8 features, ensuring that i8mm intrinsics are only used when
the target hardware actually supports them.

The motivation for this is to fix ggml CI build failures where the
feature detection correctly identifies that i8mm is not supported,
adding the +noi8mm flag, but MATMUL_INT8 preprocessor definitions are
still enabled, causing the compiler to attempt to use vmmlaq_s32
intrinsics without i8mm support.

Refs: https://github.com/ggml-org/ggml/actions/runs/17525174120/job/49909199499
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

ggml changes relating to the ggml tensor library for machine learning

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants