feat: convert ollama provider to an openai configuration#34
feat: convert ollama provider to an openai configuration#34baxen merged 6 commits intosquare:mainfrom
Conversation
|
tests work with ollama and mistral-nemo, but the vision test isn't passing with llava:7b. trying something else for that |
|
hrm - may need to re-arrange things so that CI tests are run before ruff? |
9d5d28a to
b1c4070
Compare
|
ok this is merged with #23, though llama3.1:8b-instruct-q4_0 isn't happy with the password question: basically, most of the time, it will call the function, such as the output below, but not return the value. mistral-nemo responds as expected. So, we need to figure out what to do about vision, still. |
Signed-off-by: Adrian Cole <adrian.cole@elastic.co>
Signed-off-by: Adrian Cole <adrian.cole@elastic.co>
Signed-off-by: Adrian Cole <adrian.cole@elastic.co>
|
I have checked this with anthropic, ollama and openai and seems to make things work nicely across those. |
|
using mistral-nemo is the only choice until #39 when we can start playing with llama based models like llama3-groq-tool-use |
baxen
left a comment
There was a problem hiding this comment.
Looks great! This is also how we manage the databricks provider - through their openai chat compatibility. Works well for me locally too
* main: feat: Rework error handling (#48) chore(release): release version 0.9.0 (#45) chore: add just command for releases and update pyproject for changelog (#43) feat: convert ollama provider to an openai configuration (#34) fix: Bedrock Provider request (#29) test: Update truncate and summarize tests to check for sytem prompt t… (#42) chore: update test_tools to read a file instead of get a password (#38) fix: Use placeholder message to check tokens (#41) feat: rewind to user message (#30) chore: Update LICENSE (#40) fix: shouldn't hardcode truncate to gpt4o mini (#35)
This converts the ollama provider to an openai configuration, which reduces the amount of code we need to maintain.
Later we may choose to resurrect native ollama, but early on I think it is better to focus, since ollama supports the OpenAI API.
Specifically, this uses the "mistral-nemo" model until we find something else better. Currently, this configuration does not support vision, so screenshots won't work.
After this is in, we can revisit #23 as CI setup for this provider.