convert : add Llama4ForCausalLM#16042
Conversation
|
Usually there's a |
It's the case for all llama 4 models, but traditionally this the KQ norm is something arch-specific so I didn't add GGUF metadata in the first place. but yes I think we should add it now. |
|
Hmm but as this may affect existing GGUF for the larger llama 4 MoE, I think it will be a bit messy to add it as GGUF metadata. In anyway, the rule is currently as follow: all llama 4 models uses KQ norm except for the biggest 17B_128E So I think we can keep it as-is |
Absolutely, just wondering if it's worth adding it for the future. |
* convert : add Llama4ForCausalLM * handle swa * half working version * fix use_kq_norm * fix use_kq_norm
* convert : add Llama4ForCausalLM * handle swa * half working version * fix use_kq_norm * fix use_kq_norm
Fix #16021
Tested with:
Very important note from model card:
That is the reason why model cannot response to simple "hi"