feat: add changes to handle jina v2 base code#7596
feat: add changes to handle jina v2 base code#7596ggerganov merged 10 commits intoggml-org:masterfrom
Conversation
dd42a71 to
21936dd
Compare
17a5e9f to
9a65c7a
Compare
…t-jina-v2-base-code
4117c40 to
96a6f55
Compare
|
It's how the tokens are handled in llama.cpp. I'm in the middle of figuring out how Aside: I have no idea how the CD/CI is setup here. I have some experience with Jenkins, but all of this is outside of the scope of what I'm focused on. Also, I'm just a contributor. I just chime in when I think I might have something of value to add. |
Hey @teleprint-me , To be honest, I found it quite hard to work with tokenizer logic here, but I do not quite understand what you aim to achieve in #7379. If you want we can jump in a call to discuss and make this process more agile. |
Rebase on latest |
I will, thanks |
Hm, not sure why this happens. We don't escape strings in the vocab - only in the prompt input: |
I will try to investigate this |
|
I am also trying to see if I can add support for |
|
I believe the most recent model that we added and also supports Chinese is https://huggingface.co/deepseek-ai/DeepSeek-V2. See if @fairydreaming's PR could be of any help: #7519 |
|
Hey @ggerganov , I am starting to think that it is not a problem of the tokenizer. Here is my observation. I am tryng to run this code to check how the embedding behaves: gdb --args ../build/bin/embedding -m ./jina-embeddings-v2-base-code.gguf --threads 1 --verbose-prompt -p "for idx, x in enumerate(xs):\n print(idx, x)"and this is what Look at all the I am not sure if it is a problem of how the standard input is encoded or something? Do you happen to have any clue about this? if I hardcode this sentence and avoid the split lines. I get the same behavior as in Python |
|
I see, does adding ../build/bin/embedding -m ./jina-embeddings-v2-base-code.gguf --threads 1 --verbose-prompt -e -p "for idx, x in enumerate(xs):\n print(idx, x)" |
Oh, it does! |
…t-jina-v2-base-code
how then can we be sure this behavior is available in the server? I see this escape option only available in the example itself. |
|
I believe |
0481e5f to
3b44f8f
Compare
404daca to
05659d3
Compare
|
Hey @ggerganov, Is there something from my code that may have caused this CI to fail? |
|
Probably just a fluke, will restart the workflows now |
…t-jina-v2-base-code
|
@ggerganov I tested the behavior in server and works, I consider this is ready to be reviewed. |
* feat: add changes to handle jina v2 base code * fix: do not complicate things * fix: fix the usage of the code model * fix: fix comments * fix: fix linting issues * fix: remove ollama patches * style : minor --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
* feat: add changes to handle jina v2 base code * fix: do not complicate things * fix: fix the usage of the code model * fix: fix comments * fix: fix linting issues * fix: remove ollama patches * style : minor --------- Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>




PR to allow using
jinaai/jina-embeddings-v2-base-codewithllama.cpp. It has an extra normalization layer compared to other models of theJinaV2family and this is why it is considered independently.