Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Baichuan2-7B-Chat model converted to ggml-model-q4_0.gguf, AI answer does not stop automatically when inference is made #5034

Closed
Lyzin opened this issue Jan 19, 2024 · 4 comments

Comments

@Lyzin
Copy link

Lyzin commented Jan 19, 2024

I converted the Baichuan2-7B-Chat model to ggml-model-q4_0.gguf and then used . /main inference and found that I could not stop answering

my system: macos
python version: 3.9.10

Here's my version using llama.cpp
. /main version:
image

llama.cpp git commit id:
image

Here are the steps I took to convert

python convert-hf-to-gguf.py ./models/Baichuan2-7B-Chat

./quantize ./models/Baichuan2-7B-Chat/ggml-model-f16.gguf ./models/Baichuan2-7B-Chat/ggml-model-q4_0.gguf q4_0

Here's how to reason with . /main reasoning results

./main -m ./models/Baichuan2-7B-Chat/ggml-model-q4_0.gguf -n 256 --repeat_penalty 1.0 -ngl 0 --color -i -r "User:" -f prompts/chat-with-baichuan.txt

Here is the result of the inference, which keeps repeating the output
image

Is this a problem caused by the conversion?

@Lyzin Lyzin changed the title baichuan2-7B-Chat model converted to ggml-model-q4_0.gguf, AI answer does not stop automatically when inference is made Baichuan2-7B-Chat model converted to ggml-model-q4_0.gguf, AI answer does not stop automatically when inference is made Jan 19, 2024
@hiepxanh
Copy link

@Lyzin I belive it is the same issue with this #3969
can you download Phi-2 model and confirm the bug still happen?

@Lyzin
Copy link
Author

Lyzin commented Jan 25, 2024

@Lyzin I belive it is the same issue with this #3969 can you download Phi-2 model and confirm the bug still happen?

I downloaded the phi-2 model, requantized it, and the output automatically stopped
llama.cpp version tag: b1966

./main -m ./models/phi-2/ggml-model-q4_0.gguf -n 512 --color -i -cml -ngl 0 -r "User:" -f prompts/chat-with-bob.txt

Here is the output of the AI
image

Copy link
Contributor

This issue is stale because it has been open for 30 days with no activity.

@github-actions github-actions bot added the stale label Mar 18, 2024
Copy link
Contributor

github-actions bot commented Apr 3, 2024

This issue was closed because it has been inactive for 14 days since being marked as stale.

@github-actions github-actions bot closed this as completed Apr 3, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants