-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
feat: convert ollama provider to an openai configuration #34
Conversation
tests work with ollama and mistral-nemo, but the vision test isn't passing with llava:7b. trying something else for that |
hrm - may need to re-arrange things so that CI tests are run before ruff? |
9d5d28a
to
b1c4070
Compare
ok this is merged with #23, though llama3.1:8b-instruct-q4_0 isn't happy with the password question: basically, most of the time, it will call the function, such as the output below, but not return the value.
mistral-nemo responds as expected. So, we need to figure out what to do about vision, still. |
Signed-off-by: Adrian Cole <adrian.cole@elastic.co>
Signed-off-by: Adrian Cole <adrian.cole@elastic.co>
Signed-off-by: Adrian Cole <adrian.cole@elastic.co>
I have checked this with anthropic, ollama and openai and seems to make things work nicely across those. |
using mistral-nemo is the only choice until #39 when we can start playing with llama based models like llama3-groq-tool-use |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Looks great! This is also how we manage the databricks provider - through their openai chat compatibility. Works well for me locally too
* main: feat: Rework error handling (#48) chore(release): release version 0.9.0 (#45) chore: add just command for releases and update pyproject for changelog (#43) feat: convert ollama provider to an openai configuration (#34) fix: Bedrock Provider request (#29) test: Update truncate and summarize tests to check for sytem prompt t… (#42) chore: update test_tools to read a file instead of get a password (#38) fix: Use placeholder message to check tokens (#41) feat: rewind to user message (#30) chore: Update LICENSE (#40) fix: shouldn't hardcode truncate to gpt4o mini (#35)
This converts the ollama provider to an openai configuration, which reduces the amount of code we need to maintain.
Later we may choose to resurrect native ollama, but early on I think it is better to focus, since ollama supports the OpenAI API.
Specifically, this uses the "mistral-nemo" model until we find something else better. Currently, this configuration does not support vision, so screenshots won't work.
After this is in, we can revisit #23 as CI setup for this provider.