-
Notifications
You must be signed in to change notification settings - Fork 203
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
model is not None #71
Comments
Same here ! |
Thanks for the issue. Might be some issue from llama.cpp. Will have a look. |
need to use llama-cpp-python==0.1.78 or better yet, use the gguf models |
There so many issues in the code... but if this is you main struggle... What worked for me as workaround,, before I stumble to the next issue is: Then errors appears to be in: nn_modules\qlinear\qlinear_cuda_old.py And then I surrender :) |
It is the new gguf format that llama.cpp is designen for.
I replaced the code to download the gguf versión and it loads it well,
however the output seems to be in hexadecimal which I do not understand why.
The other thing is to stick to the old version of llama.cpp , but I think
that is not the idea.
El jue, 28 sept 2023, 17:35, AIGoldenfinger ***@***.***>
escribió:
… There so many issues in the code... but if this is you main struggle...
What worked for me as workaround,, before I stumble to the next issue is:
Replace in .env file , BACKEND_TYPE = "llama.cpp" with "gptq" it will
download the model and continued to the next error ...
Then errors appears to be in: nn_modules\qlinear\qlinear_cuda_old.py
line 83, in *init* self.autogptq_cuda = autogptq_cuda_256
NameError: name 'autogptq_cuda_256' is not defined
And then I surrender :)
—
Reply to this email directly, view it on GitHub
<#71 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AA3NNMYOJJJ3WJJ3E77QWEDX4WKK5ANCNFSM6AAAAAA5AZH254>
.
You are receiving this because you commented.Message ID:
***@***.***>
|
@jamartinh @HyperUpscale Thanks for the feedback. I believe the issue is mainly due to the new GGUF format.
to stick the old version of |
@jamartinh @HyperUpscale hi now the new release llama2-wrapper=0.1.13 will lock on llama-cpp-python = "0.1.77" to support old ggml models. |
#73 now supporting gguf models. |
Running on backend llama.cpp.
Traceback (most recent call last):
File "/root/llama2-webui/app.py", line 325, in
main()
File "/root/llama2-webui/app.py", line 56, in main
llama2_wrapper = LLAMA2_WRAPPER(
^^^^^^^^^^^^^^^
File "/root/llama2-webui/llama2_wrapper/model.py", line 99, in init
self.init_model()
File "/root/llama2-webui/llama2_wrapper/model.py", line 103, in init_model
self.model = LLAMA2_WRAPPER.create_llama2_model(
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/root/llama2-webui/llama2_wrapper/model.py", line 125, in create_llama2_model
model = Llama(
^^^^^^
File "/root/llama2-webui/new1/lib/python3.11/site-packages/llama_cpp/llama.py", line 323, in init
assert self.model is not None
^^^^^^^^^^^^^^^^^^^^^^
AssertionError
The text was updated successfully, but these errors were encountered: