-
Notifications
You must be signed in to change notification settings - Fork 11.2k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Output of quantized Vicuna is so inappropriate that I can't use it #1656
Comments
I lack experience with that particular model but I do notice that you attempt a complex instruction solving translation using a 7B model. |
@JerryYao80 You didn't use the correct prompt format for Vicuna models. You also asked it to translate from "Englisth" to Chinese. 我没想批评您的英语水平,希望我的话不会让您不舒服。显然您的英语比我的中文好得多啊! Because of the way LLMs just complete text, the input makes a huge difference. Typos and grammar mistakes in the prompt, unfortunately will generally cause you to get low quality output. Also not using the prompt format the model expects. I'd also note that while Vicuna can speak a little Mandarin, that only made up a small part of its training. Even with the best possible prompting, I wouldn't expect the results for translations or generating text to be very good (especially if you're using a 7B model). |
Also this really isn't a llamacpp issue unless it's a tokenizer problem. You can confirm whether the input tokens match the vocab. |
Does the following work better?
|
This issue was closed because it has been inactive for 14 days since being marked as stale. |
Prerequisites
Please answer the following questions for yourself before submitting an issue.
Current Behavior
In my llama.cpp environment, I get the following files under the guide of README:
When I executed the command:
But I got the output:
This is one of my experiment, and the outputs of quantized Vicuna are so inappropriate that I can't do anything,
I'm not sure what's wrong with my quantized model. Does anyone meet it?
Environment and Context
The text was updated successfully, but these errors were encountered: