Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Running Hunyuan Video on 4 x 3090 Graphics Cards #96

Closed
Smiden3 opened this issue Dec 9, 2024 · 16 comments
Closed

Running Hunyuan Video on 4 x 3090 Graphics Cards #96

Smiden3 opened this issue Dec 9, 2024 · 16 comments

Comments

@Smiden3
Copy link

Smiden3 commented Dec 9, 2024

Hello! I was lucky enough to have access to a Windows-based system with a 4 graphics card and a 3090. I installed and started the video generator using ComfyUI. But right now I'm only using 24GB of GPUs out of 96. Could you talk about using XDIT for parallel use? I will be glad to have any information on this topic, thank you
Screenshot_20241209_183209_Browser
Screenshot_20241209_183233_Browser

@Lissanro
Copy link

Lissanro commented Dec 10, 2024

My configuration is similar, except I am on Linux. I have four 3090, so 96GB of VRAM in total, but how to utilize all the cards?

EDIT: I know there are some command-line examples for xDiT, but just like the person who reported this bug, it is unclear to me how to get this working in ComfyUI, or it is not supported yet?

@feifeibear
Copy link
Contributor

At present, the xDiT has not yet been tested within ComfyUI, and it appears that four 3090 GPUs may not suffice to meet the memory demands of HunayunVideo. However, we are considering exploring ComfyUI in the near future.

@Smiden3
Copy link
Author

Smiden3 commented Dec 10, 2024

In principle, I don't mind if with XDiT I can use a 4 graphics card via the command line. But I've never worked with the command line in this format, so I'd like to know if there is a detailed and step-by-step guide (text, video) on how to run video generation?

@feifeibear
Copy link
Contributor

You can follow the instructions in the README and run it from the command line. However, I need to remind you that running HunyuanVideo on four 3090 GPUs is not feasible due to the substantial memory requirements of the model.

https://github.com/xdit-project/HunyuanVideo-xDiT?tab=readme-ov-file#parallel-inference-on-multiple-gpus-by-xdit

@inspire-boy
Copy link

You can follow the instructions in the README and run it from the command line. However, I need to remind you that running HunyuanVideo on four 3090 GPUs is not feasible due to the substantial memory requirements of the model.

https://github.com/xdit-project/HunyuanVideo-xDiT?tab=readme-ov-file#parallel-inference-on-multiple-gpus-by-xdit

why It‘s Gone.

@Smiden3
Copy link
Author

Smiden3 commented Dec 10, 2024

I've seen the quick startup, this should really help me figure out how to start properly, however, the examples don't include running Hunyuan Video.
Screenshot_20241210_095310_Browser
Screenshot_20241210_095143_Browser

@feifeibear
Copy link
Contributor

@Smiden3
Copy link
Author

Smiden3 commented Dec 10, 2024

You can follow the instructions in the README and run it from the command line. However, I need to remind you that running HunyuanVideo on four 3090 GPUs is not feasible due to the substantial memory requirements of the model.

https://github.com/xdit-project/HunyuanVideo-xDiT?tab=readme-ov-file#parallel-inference-on-multiple-gpus-by-xdit

I also wonder why you say that there is not enough computing power. I can run the model even on a single 3090 graphics card, but of course I can't use a resolution of 1280/720 at 50 steps, but I can use a lower resolution.

@Smiden3
Copy link
Author

Smiden3 commented Dec 10, 2024

https://github.com/xdit-project/HunyuanVideo-xDiT?tab=readme-ov-file#параллельное выполнение на нескольких графических процессорах с помощью xdit

Пожалуйста, следуйте инструкциям в HunyuanVideo readme!

Screenshot_20241210_100351_Samsung Internet
This is what I see when I click on the link

@feifeibear
Copy link
Contributor

@Smiden3 sorry! its all my fault. The link should be:

https://github.com/Tencent/HunyuanVideo?tab=readme-ov-file#-parallel-inference-on-multiple-gpus-by-xdit

Sorry, I pointed you to a private repo.

@Smiden3
Copy link
Author

Smiden3 commented Dec 10, 2024

thank you @feifeibear

@doogyhatts
Copy link

At present, the xDiT has not yet been tested within ComfyUI, and it appears that four 3090 GPUs may not suffice to meet the memory demands of HunayunVideo. However, we are considering exploring ComfyUI in the near future.

Just curious on the memory allocation for the model under xDiT conditions.
Does each GPU loads a copy of the model or the model is still loaded once?
I can understand why using 4x3090 will not work if the 25gb model has to be loaded on each GPU.

@feifeibear
Copy link
Contributor

At present, the xDiT has not yet been tested within ComfyUI, and it appears that four 3090 GPUs may not suffice to meet the memory demands of HunayunVideo. However, we are considering exploring ComfyUI in the near future.

Just curious on the memory allocation for the model under xDiT conditions. Does each GPU loads a copy of the model or the model is still loaded once? I can understand why using 4x3090 will not work if the 25gb model has to be loaded on each GPU.

The model is replicated on each GPUs when using sequence parallel. But we have other parallel approaches with lower memory requirements. If you are interested, you can find them in our xDiT paper.

https://arxiv.org/abs/2411.01738

@doogyhatts
Copy link

doogyhatts commented Dec 10, 2024

The model is replicated on each GPUs when using sequence parallel. But we have other parallel approaches with lower memory requirements. If you are interested, you can find them in our xDiT paper.

https://arxiv.org/abs/2411.01738

Thanks for the clarification!

@rzgarespo
Copy link

My configuration is similar, except I am on Linux. I have four 3090, so 96GB of VRAM in total, but how to utilize all the cards?

EDIT: I know there are some command-line examples for xDiT, but just like the person who reported this bug, it is unclear to me how to get this working in ComfyUI, or it is not supported yet?

I've RTX3080 and 96GB DDR4 RAM
https://www.youtube.com/watch?v=ylfeJ7Cv8AE

@lamarquenet
Copy link

We really need a solution for running this on 4 3090 cards

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

8 participants