65B working on multi-gpu #39
ortegaalfredo
started this conversation in
General
Replies: 3 comments 6 replies
-
. |
Beta Was this translation helpful? Give feedback.
0 replies
-
Moving this here instead of just closing the issue. I'm very happy to hear that it's working out for people. Especially that you can usable performance from multiple GPUs. Could I ask what CPU you're using? |
Beta Was this translation helpful? Give feedback.
3 replies
-
I tried to get llama-65b to work on g5.12x (4 GPUs with 24G vram each), but it gave me oom error. Any clue how to get it to work?
|
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
This is not a issue, just reporting that it works great with Guanaco-65B-GPTQ-4bit.act-order.safetensors from TheBloke using 2x3090. Speed is great, about 15t/s.
Beta Was this translation helpful? Give feedback.
All reactions