Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Feature]: Ollama Llama 3.2 11b vision support #6683

Open
davidADSP opened this issue Nov 10, 2024 · 0 comments
Open

[Feature]: Ollama Llama 3.2 11b vision support #6683

davidADSP opened this issue Nov 10, 2024 · 0 comments
Labels
enhancement New feature or request

Comments

@davidADSP
Copy link

davidADSP commented Nov 10, 2024

The Feature

Currently, the following snippet that works for ollama/llava, does not work for ollama/llama3.2-vision:11b (recently added to ollama). It interprets the byte string as text.

Do you have a timeframe for when support for this will be added to litellm?

import litellm

response = litellm.completion(
  model = "ollama/llama3.2-vision:11b",
  messages=[
      {
          "role": "user",
          "content": [
                          {
                              "type": "text",
                              "text": "Whats in this image?"
                          },
                          {
                              "type": "image_url",
                              "image_url": {
                              "url": "iVBORw0KGgoAAAANSUhEUgAAAG0AAA...SuQmCC"
                              }
                          }
                      ]
      }
  ],
)
print(response)

gives this:

message=Message(content="It looks like you've provided a large, encoded string. Unfortunately, I don't have the capability to decode or interpret it directly.\n\nHowever, based on the structure and content of the string, it appears to be a serialized data format, possibly in a binary or text-based encoding scheme such as JSON or XML.\n\nIf you could provide more context about what this string represents (e.g., is it a compressed file, an image, or some other type of data?) or what you're trying to achieve with this data, I may be able to help you better. Alternatively, if you can decode the string yourself and provide a human-readable representation, I'd be happy to assist with any questions or tasks related to the decoded data!", role='assistant'

Motivation, pitch

Currently LiteLLM does not support all models on Ollama, as stated on your website without this change :D

Twitter / LinkedIn details

https://www.linkedin.com/in/davidtfoster/

@davidADSP davidADSP added the enhancement New feature or request label Nov 10, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant