Skip to content

llama : add comments about experimental flags#7544

Merged
ggerganov merged 1 commit intomasterfrom
gg/fattn-warn
May 27, 2024
Merged

llama : add comments about experimental flags#7544
ggerganov merged 1 commit intomasterfrom
gg/fattn-warn

Conversation

@ggerganov
Copy link
Copy Markdown
Member

@ggerganov ggerganov commented May 26, 2024

Certain combinations of [EXPERIMENTAL] llama_context_params are not always supported:

    struct llama_context_params {
        ...

        enum ggml_type type_k; // data type for K cache [EXPERIMENTAL]
        enum ggml_type type_v; // data type for V cache [EXPERIMENTAL]

        bool flash_attn;  // whether to use flash attention [EXPERIMENTAL]

        ...
    };

Here is a list of known incompatibilities (we can try to update it in the future):

@ggerganov ggerganov merged commit eaf6e03 into master May 27, 2024
@ggerganov ggerganov deleted the gg/fattn-warn branch May 27, 2024 06:24
Seunghhon pushed a commit to Seunghhon/llama.cpp that referenced this pull request Apr 26, 2026
phuongncn pushed a commit to phuongncn/llama.cpp-gx10-dgx-sparks-deepseekv4 that referenced this pull request Apr 28, 2026
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant