GPU question #217
Replies: 14 comments 19 replies
-
Unfortunately not. The current implementation would work with CPU only. I am trying to make this work on GPU too. So far, the first few steps I can provide are:
Update: I have successfully ran the model on my GPU. Planning to drop commits. |
Beta Was this translation helpful? Give feedback.
-
@maozdemir please ping me at jakub.zboina@comtegra.pl when you drop new commit with GPU support. Thanks :) |
Beta Was this translation helpful? Give feedback.
-
I had been working on CUDA support yesterday with no luck. Glad to hear you had some success. Waiting for the commit, as well. |
Beta Was this translation helpful? Give feedback.
-
https://github.com/maozdemir/privateGPT-colab/blob/main/privateGPT-colab.ipynb Set n_gpu_layers=500 for colab in LlamaCpp and LlamaCppEmbeddings functions, also don't use GPT4All, it won't run on GPU. |
Beta Was this translation helpful? Give feedback.
-
Hi all, on Windows here but I finally got inference with GPU working! (These tips assume you already have a working version of this project, but just want to start using GPU instead of CPU for inference).
This issue was quite helpful to me if you aren't able to get it working with tips above: abetlen/llama-cpp-python#250 |
Beta Was this translation helpful? Give feedback.
-
In case you guys are curious, our h2oGPT is focused more on GPU, has fully UI, and otherwise is like privateGPT: https://github.com/h2oai/h2ogpt |
Beta Was this translation helpful? Give feedback.
-
check #425 |
Beta Was this translation helpful? Give feedback.
-
what about using MPS for training and the Neural Engine for inference on Apple Silicon? |
Beta Was this translation helpful? Give feedback.
-
If someone get stuck here (like me) with this error: |
Beta Was this translation helpful? Give feedback.
-
With a 3060Ti it's for some reason much slower than on my old i5-7400, single core takes like 5 min, and quad core takes nearly 4 minutes... But on the 3060Ti it takes 10 mins+. Anyone knows why it's happening ? |
Beta Was this translation helpful? Give feedback.
-
For some reason, when i made these changes, im getting this error and i cant fix it
|
Beta Was this translation helpful? Give feedback.
-
Hello, @maozdemir ! 1st of all, congratulations for effort to providing GPU support to 2nd, I'm starting to use CUDA, and I've just downloaded the CUDA framework for my old fashioned GTX 750 Ti. Currently NVIDIA provides the version 12.2 for its framework, and no longer 11.8 However, your install demands only the 11.8 version for CUDA. Do you think it can run in 12.2 too ? Thanks for any help you can offer. |
Beta Was this translation helpful? Give feedback.
-
You can get it here: |
Beta Was this translation helpful? Give feedback.
-
What about support for running on a Havana Labs Gaudi base box? |
Beta Was this translation helpful? Give feedback.
-
I'm curious to setup this model myself. I have two 3090's and 128 gigs of ram on an i9 all liquid cooled. Would the GPU play any relevance in this or is that only used for training models?
Beta Was this translation helpful? Give feedback.
All reactions