How to run Llama3-8b instruct model on multiple GPUs? #7086
aitechguy0105
started this conversation in
General
Replies: 3 comments 2 replies
-
Why would you run it on multiple GPUs? |
Beta Was this translation helpful? Give feedback.
0 replies
-
to test benchmark. other models seems to be supported on multiple GPUs. |
Beta Was this translation helpful? Give feedback.
1 reply
-
One reason for trying to run on multiple GPUs, as in my case is, it does not fit on one. I have 2 Nividia 2080 tI, each has around 11GB of memory. By default it choses GPU:0. I was wondering if I can parallelize between the two GPUs to run it. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Deos llama.cpp support llama3-8b to run on multiple GPUs?
Beta Was this translation helpful? Give feedback.
All reactions