Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

The prompt is not converted to tokens #113

Closed
zhoujian1028 opened this issue Mar 14, 2023 · 8 comments
Closed

The prompt is not converted to tokens #113

zhoujian1028 opened this issue Mar 14, 2023 · 8 comments
Labels
bug Something isn't working

Comments

@zhoujian1028
Copy link

./main -m ./models/7B/ggml-model-q4_0.bin -p "Building a website can be done in 10 simple steps:" -t 8 -n 512
image

llama.cpp Is the latest version
Can anyone help me? Thanks!

@zhoujian1028
Copy link
Author

@ggerganov Please,Thanks!

@gjmulder gjmulder added the bug Something isn't working label Mar 15, 2023
@MillionthOdin16
Copy link

I'm seeing the same issue. Model works fine in f16 version, but doesn't work after quantization

@caseylai
Copy link

Same issue here. Everything is fine on my mac laptop but after I deployed it on Linux (Centos 7), I found this issue.
Have tried to change the version of gcc, python, torch, and tried to modify the source code of 'llama_tokenize' to make the tokenizer working as expected. But none of these works. And also checked md5 sum for all files, all of the md5 sum are right.

Environment:
Mac (works fine): gcc 9.4.0, Python 3.9.6, Torch 1.9.0
Centos 7 (issue): gcc 9.3.1, Python 3.9.16, Torch 1.9.1

@MillionthOdin16
Copy link

MillionthOdin16 commented Mar 17, 2023 via email

@caseylai
Copy link

./main -m ./models/7B/ggml-model-q4_0.bin -p "Building a website can be done in 10 simple steps:" -t 8 -n 512 image

llama.cpp Is the latest version Can anyone help me? Thanks!

@zhoujian1028 I tried to deploy on another machine and it worked smoothly, for your reference: Ubuntu 22.04 + gcc11.3.0 + python 3.10.9 + torch 1.13.1

@xlichao
Copy link

xlichao commented Mar 18, 2023

./main -m ./models/7B/ggml-model-q4_0.bin -p "Building a website can be done in 10 simple steps:" -t 8 -n 512 image

llama.cpp Is the latest version Can anyone help me? Thanks!

Centos 7 even with sudo yum install devtoolset-11-gcc* can't work and output nothing meaningful, but just using an docker conatiner base on ubuntu jemmy with sudo apt-get install build-essential everything works fine in fp16.

@ar-jan
Copy link

ar-jan commented Mar 20, 2023

Give it another try now that #252 landed.

@sw
Copy link
Contributor

sw commented Apr 7, 2023

No complaints after three weeks, let's assume this is fixed, possibly by #252.

@sw sw closed this as completed Apr 7, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

7 participants