Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Model not loading #27

Open
glemley opened this issue Sep 18, 2024 · 1 comment
Open

Model not loading #27

glemley opened this issue Sep 18, 2024 · 1 comment

Comments

@glemley
Copy link

glemley commented Sep 18, 2024

I successfully compiled LlamaGPTJ and downloaded the 7b and 13b LLaMA models. It was not clear where to place them but the only 'models' folder was at \gpt4all-backend\llama.cpp\models. I placed them here and am getting the following message, followed by continuously running loading dots.

Your computer supports AVX2
LlamaGPTJ-chat: loading .\models\ggml-vicuna-13b-1.1-q4_2.bin
.........

Is there something else that needs to be done to point the program to the model?

Thanks

@kuvaus
Copy link
Owner

kuvaus commented Sep 20, 2024

Hi,

You can put the model bin files anywhere on your computer. Just use -m flag and then path to the bin file.

So like:
./chat -m "/Users/kuvaus/mynewfolderformodels/ggml-vicuna-13b-1.1-q4_2.bin"

But without the -m it looks for the models folder in the same directory as the chat executable. So you need to make a new models folder in the same directory where you have the executable. (After you compiled the chat, you can move it anywhere, just have the models folder next to it). You were right to ask the question, by default it won't find the models from the \gpt4all-backend\llama.cpp\models but just from .\models.

One more thing: it can take a while to load the model so you'll be seeing the running dots loop until the model is fully ready.

Thanks for trying the program. :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants