-
Notifications
You must be signed in to change notification settings - Fork 1.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
The output from the WizardCoder model has leading whitespace. #472
Comments
Thank you for reporting the bug. This seems to be a case that can be addressed through post-processing. Additionally, enhancing the prompt template of WizardCoder might also lead to improved output. |
With the decoding fix released in v0.2.1, this situation should have improved to some extent.
Hi @MinjeJeon, could you please verify if you have observed an improvement in your experience? |
I tested it with the curl you told me and it works well.
But when I ran another auto-completion, I noticed that there was leading whitespace as shown below. I wonder if the
Thank you for your reply. |
Describe the bug
A clear and concise description of what the bug is.
After launching a self-hosted server with the WizardCoder-3B model and using auto-completion in VSCode, there is an issue where additional whitespace is added to the generated string at the beginning.
Information about your GPU
Please provide output of
nvidia-smi
Additional context
Add any other context about the problem here.
I am using the tabby docker image 0.1.2.
Tabby extension version: 0.5.0
VSCode version: 1.82.2
The text was updated successfully, but these errors were encountered: