65B working on multi-gpu #39
ortegaalfredo
started this conversation in
General
Replies: 3 comments 6 replies
-
Moving this here instead of just closing the issue. I'm very happy to hear that it's working out for people. Especially that you can usable performance from multiple GPUs. Could I ask what CPU you're using? |
Beta Was this translation helpful? Give feedback.
3 replies
-
I tried to get llama-65b to work on g5.12x (4 GPUs with 24G vram each), but it gave me oom error. Any clue how to get it to work?
|
Beta Was this translation helpful? Give feedback.
3 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
This is not a issue, just reporting that it works great with Guanaco-65B-GPTQ-4bit.act-order.safetensors from TheBloke using 2x3090. Speed is great, about 15t/s.
Beta Was this translation helpful? Give feedback.
All reactions