You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Lyzin
changed the title
baichuan2-7B-Chat model converted to ggml-model-q4_0.gguf, AI answer does not stop automatically when inference is made
Baichuan2-7B-Chat model converted to ggml-model-q4_0.gguf, AI answer does not stop automatically when inference is made
Jan 19, 2024
I converted the Baichuan2-7B-Chat model to ggml-model-q4_0.gguf and then used . /main inference and found that I could not stop answering
my system: macos
python version: 3.9.10
Here's my version using llama.cpp
. /main version:
llama.cpp git commit id:
Here are the steps I took to convert
Here's how to reason with . /main reasoning results
Here is the result of the inference, which keeps repeating the output
Is this a problem caused by the conversion?
The text was updated successfully, but these errors were encountered: