How to run Llama3-8b instruct model on multiple GPUs? #7086
aitechguy0105
started this conversation in
General
Replies: 3 comments 2 replies
-
Why would you run it on multiple GPUs? |
Beta Was this translation helpful? Give feedback.
0 replies
-
to test benchmark. other models seems to be supported on multiple GPUs. |
Beta Was this translation helpful? Give feedback.
1 reply
-
One reason for trying to run on multiple GPUs, as in my case is, it does not fit on one. I have 2 Nividia 2080 tI, each has around 11GB of memory. By default it choses GPU:0. I was wondering if I can parallelize between the two GPUs to run it. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Deos llama.cpp support llama3-8b to run on multiple GPUs?
Beta Was this translation helpful? Give feedback.
All reactions