Release v0.2.35
What's Changed
- add dolphin by @infwinston in #2794
- Fix tiny typo by @bofenghuang in #2805
- Add instructions for evaluating on MT bench using vLLM by @iojw in #2770
- fix missing op | for py3.8 by @dumpmemory in #2800
- Add SOLAR-10.7b Instruct Model by @BabyChouSr in #2826
- Update README.md by @eltociear in #2852
- fix: 'compeletion' typo by @congchan in #2847
- Add Tunnelmole as an open source alternative to ngrok and include usage instructions by @robbie-cahill in #2846
- Add support for CatPPT by @rishiraj in #2840
- Add functionality to ping AI2 InferD endpoints for tulu 2 by @natolambert in #2832
- add download models from www.modelscope.cn by @liuyhwangyh in #2830
- Fix conv_template of chinese alpaca 2 by @zollty in #2812
- add bagel model adapter by @jondurbin in #2814
- add root_path argument to gradio web server. by @stephanbertl in #2807
- Import
accelerate
locally to avoid it as a strong dependency by @chiragjn in #2820 - Replace dict merge with unpacking for compatibility of 3.8 in vLLM worker by @rudeigerc in #2824
- Format code by @merrymercy in #2854
- Openai API migrate by @andy-yang-1 in #2765
- Add new models (Perplexity, gemini) & Separate GPT versions by @merrymercy in #2856
- Clean error messages by @merrymercy in #2857
- Update docs by @Ying1123 in #2858
- Modify doc description by @zhangsibo1129 in #2859
- Fix the problem of not using the decoding method corresponding to the base model in peft mode by @Jingsong-Yan in #2865
- update a new sota model on MT-Bench which touch an 8.8 scores. by @xiechengmude in #2864
- NPU needs to be initialized when starting a new process by @jq460494839 in #2843
- Fix the problem with "vllm + chatglm3" (#2845) by @yaofeng in #2876
- Update token spacing for mistral conversation.py by @thavens in #2872
- check if hm in models before deleting to avoid errors by @joshua-ne in #2870
- Add TinyLlama by @Gk-rohan in #2889
- Fix bug that model doesn't automatically switch peft adapter by @Jingsong-Yan in #2884
- Update web server commands by @merrymercy in #2869
- fix the tokenize process and prompt template of chatglm3 by @WHDY in #2883
- Add
Notus
support by @gabrielmbmb in #2813 - feat: support anthropic api with api_dict by @congchan in #2879
- Update model_adapter.py by @thavens in #2895
- leaderboard code update by @infwinston in #2867
- fix: change order of SEQUENCE_LENGTH_KEYS by @congchan in #2925
- fix baichuan:apply_prompt_template call args error by @Force1ess in #2921
- Fix a typo in openai_api_server.py by @jklj077 in #2905
- feat: use variables OPENAI_MODEL_LIST by @congchan in #2907
- Add TenyxChat-7B-v1 model by @sarath-shekkizhar in #2901
- add support for iei yuan2.0 (https://huggingface.co/IEITYuan) by @wangpengfei1013 in #2919
- nous-hermes-2-mixtral-dpo by @152334H in #2922
- Bump the version to 0.2.35 by @merrymercy in #2927
New Contributors
- @dumpmemory made their first contribution in #2800
- @robbie-cahill made their first contribution in #2846
- @rishiraj made their first contribution in #2840
- @natolambert made their first contribution in #2832
- @liuyhwangyh made their first contribution in #2830
- @stephanbertl made their first contribution in #2807
- @chiragjn made their first contribution in #2820
- @rudeigerc made their first contribution in #2824
- @jq460494839 made their first contribution in #2843
- @yaofeng made their first contribution in #2876
- @thavens made their first contribution in #2872
- @joshua-ne made their first contribution in #2870
- @WHDY made their first contribution in #2883
- @gabrielmbmb made their first contribution in #2813
- @jklj077 made their first contribution in #2905
- @sarath-shekkizhar made their first contribution in #2901
- @wangpengfei1013 made their first contribution in #2919
Full Changelog: v0.2.34...v0.2.35