-
Notifications
You must be signed in to change notification settings - Fork 276
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Contributing/Adding new model #1542
Comments
@PhantomSpike - great idea. I'd like to nominate StableVicuna and Mosic MPT-7B, for starters. Since many of the OSS models are in HuggingFace, I wonder of HELM already has HF support, and we just need the eval compute. Or perhaps, each model still needs some HELM adaption beyond HF. |
Hi,
Let me know if you have more questions and feel free to add me as a reviewer to your PR. Happy coding! |
You might also find the existing documentation handy: development setup and adding new models. As Josselin mentioned, we plan on improving the workflow and documentation soon. |
Hi @JosselinSomervilleRoberts @yifanmai - I'm helping organize a competition to eval LLMs so was very interested in easier ways of adding models
FWIW I'm OK on settling on a single client like HF Hub if they make 1 and 3 more seamless EDIT: This already does what I need https://github.com/stanford-crfm/helm/blob/main/docs/huggingface_models.md |
Hi @msaroufim,
It mostly depends on what kind of inference servers / libraries you want to support. I would be interested in adding integrations with other inference servers / libraries that work similarly to the Hugging Face one. I'm also considering moving the Hugging Face flags into a config file, because it get very unwieldy if you have lots of models. |
Thanks @yifanmai! Yes for our use cases a local eval and an eval on object stores would be really invaluable, I'd imagine many competitors would not want to publicly share their weights until the competition is complete Regarding other libraries or inference servers that part is also interesting, on one hand, if all competitors settled on HF then that makes our eval simpler but we also want to make it easy for people who may not want to use HF to also also contribute. Maybe they're using torchserve/kserve/triton or maybe they dont like the HF abstractions and would prefer to use their favorite training loop provider lightning/mosaic/something custom, I'm still undecided what to do in this case. We figured maybe a custom helm client for the competition And yes moving the HF flags to a file would be very helfpul |
I added #1673 and #1674 for these requests. I also made the "competition" label; feel free to use this to tag competition-related issues.
For common things, we can add a client e.g. I think torchserve/kserve/triton would make sense. If we want to be entirely framework agnostic, the submitter could provide a OpenAI API compatible HTTP server, or a batch script that outputs a request / response JSON file in a HELM-compatible format, and we could make HELM have better support for these. |
Another thought on private model hosting: HELM already supports private repositories |
Another thought: HELM already supports private repositories on Hugging Face Hub. The user just needs to set the Hugging Face authentication as shell environment variables before running HELM. So that could be another way of uploading and hosting a private model. |
Closing this - we now support many other ways to run private models, including Hugging Face models on disk, vLLM, OpenAI compatible servers, etc. |
Apologies if this is already described somewhere, but how can one add a new model not in the library of models already?
I am particularly interested in OSS models and would like to add many of them.
Given that the LLM area is very dynamic/fast-paced, what would be a good solution to make this a streamlined process where users can contribute easily?
The text was updated successfully, but these errors were encountered: