Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

422 Unprocessable Entity #32

Open
zhangapeng opened this issue Aug 30, 2024 · 5 comments
Open

422 Unprocessable Entity #32

zhangapeng opened this issue Aug 30, 2024 · 5 comments

Comments

@zhangapeng
Copy link

I get a “422 Unprocessable Entity” when calling a local LLM service and I don't know what's causing it。
image

@s-jse
Copy link
Member

s-jse commented Aug 30, 2024

Hi,

Can you please let us know what server and model you are using? E.g. LLaMA-3 on text-generation-inference etc.
And what command you are using to run WikiChat?

@zhangapeng
Copy link
Author

zhangapeng commented Aug 31, 2024

Hi,

Can you please let us know what server and model you are using? E.g. LLaMA-3 on text-generation-inference etc. And what command you are using to run WikiChat?

I use the api deployment code from the chatglm3 repository, which is compatible with Openai-api. I use "inv demo --engine local" to run wikichat。The error message on the terminal is as follows:
image
In addition, I successfully called the local LLM service in the litellm library alone

@s-jse
Copy link
Member

s-jse commented Aug 31, 2024

One thing to check is which port you are serving chatglm3 from. By default, WikiChat expects local models to be served from port 5002. See https://github.com/stanford-oval/WikiChat/blob/main/llm_config.yaml#L99-L103 on how to change that if needed.

If that doesn't help, you can enable LiteLLM's verbose logging (https://github.com/stanford-oval/WikiChat/blob/main/llm_config.yaml#L17) and paste the full log here, to help us with troubleshooting.

@zhangapeng
Copy link
Author

One thing to check is which port you are serving chatglm3 from. By default, WikiChat expects local models to be served from port 5002. See https://github.com/stanford-oval/WikiChat/blob/main/llm_config.yaml#L99-L103 on how to change that if needed.

If that doesn't help, you can enable LiteLLM's verbose logging (https://github.com/stanford-oval/WikiChat/blob/main/llm_config.yaml#L17) and paste the full log here, to help us with troubleshooting.

I use vllm to deploy a local LLM. How should I modify the ”local: huggingface/local" field in the llm_config.yaml file? I tried to change it to the name set when vllm was deployed, but it reported an error that the model does not exist. If I don't modify it, huggingface will report an error。
image

@s-jse
Copy link
Member

s-jse commented Sep 13, 2024

I just tested, and it does not seem to work with vLLM. I will need to look into it more closely.
In the meantime, you can use https://github.com/huggingface/text-generation-inference/, which I just tested and works with this code base.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants