Draft: feat: Support Llama 3 model (#478)#479
Draft: feat: Support Llama 3 model (#478)#479carlrobertoh merged 10 commits intocarlrobertoh:masterfrom
Conversation
0ec7c86 to
e91a278
Compare
aa92aab to
ecb270e
Compare
|
Is there any more work to be expected here, or can we perhaps remove the "Draft" prefix? |
|
As far as I can see it's finished (can you double-check the InfillPrompt?), I'm just waiting for the llama.cpp server support 😅 |
|
Nice! It looks like the model doesn't support infilling, or at least I couldn't find anything. Maybe we can just remove the new infill template for now. |
Really? All |
|
Hmm, their model card doesn't specify anything related to that matter - https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3 |
|
All I can see is that there is a lot of confusion about the model, some people have many problems, some have none... I guess it depends on how you use it, and some HF models also seem to contain problems. |
|
No, but we were using it for a while, I think. However, I can't remember the reason why we switched back. |
* feat: Support Llama 3 model (#478) * Use new InfillPrompt * Switch to lmstudio-community * Use new Prompt * llama.cpp removed the BOS token ggml-org/llama.cpp@a55d8a9 * Add tests * I would prefer a stream based solution * Add 70B models * Add tests for skipping blank system prompt * Remove InfillPrompt for now
Wow, fresh of the press, seems to be one of the best models now!! 🚀
Waiting for llama.cpp support... 😅
ggml-org/llama.cpp#6747
ggml-org/llama.cpp#6751