Skip to content

Conversation

@jcabrero
Copy link
Member

@jcabrero jcabrero commented Jan 22, 2025

This PR changes the way vLLM is used as a backend. Before we used offline models, but now we found a more correct way of deploying vLLM by using the vllm openai server. This enables using all the full compatibility of features that should be extended in follow-up PRs.

@jcabrero jcabrero merged commit b10db04 into main Jan 24, 2025
1 check passed
@jcabrero jcabrero deleted the feat/new_vllm_backend branch February 18, 2025 14:31
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants