codelion/optillm - Automatic prompt strategy proxy #922
Labels
AI-Chatbots
Topics related to advanced chatbot platforms integrating multiple AI models
Algorithms
Sorting, Learning or Classifying. All algorithms go here.
code-generation
code generation models and tools like copilot and aider
Git-Repo
Source code repository like gitlab or gh
human-verified
<INST>NEVER PICK THIS LABEL</INST>
in-context-learning
Examples of few-shot prompts for in-context learning.
llm
Large Language Models
llm-applications
Topics related to practical applications of Large Language Models in various fields
MachineLearning
ML Models, Training and Inference
openai
OpenAI APIs, LLMs, Recipes and Evals
prompt
Collection of llm prompts and notes
prompt-engineering
Developing and optimizing prompts to efficiently use language models for various applications and re
software-engineering
Best practice for software engineering
source-code
Code snippets
System-prompt
System prompts guide an LLMs response to every user message.
optillm
optillm is an OpenAI API compatible optimizing inference proxy which implements several state-of-the-art techniques that can improve the accuracy and performance of LLMs. The current focus is on implementing techniques that improve reasoning over coding, logical and mathematical queries. It is possible to beat the frontier models using these techniques across diverse tasks by doing additional compute at inference time.
SOTA results with moa-gpt-4o-mini on Arena-Hard-Auto
Installation
Just clone the repository with
git
and usepip install
to setup the dependencies.You can then run the optillm proxy as follows.
Usage
Once the proxy is running, you can just use it as a drop in replacement for an OpenAI client by setting the
base_url
ashttp://localhost:8000/v1
.You can control the technique you use for optimization by prepending the slug to the model name
{slug}-model-name
. E.g. in the above code we are usingmoa
ormixture of agents as the optimization approach. In the proxy logs you will see the following showing the
moa
is been used with the base model asgpt-4o-mini
.Implemented techniques
mcts
bon
moa
rto
z3
self_consistency
pvg
rstar
cot_reflection
plansearch
References
Suggested labels
None
The text was updated successfully, but these errors were encountered: