Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

out-of-memory #60

Open
Z-Times opened this issue Dec 24, 2024 · 2 comments
Open

out-of-memory #60

Z-Times opened this issue Dec 24, 2024 · 2 comments

Comments

@Z-Times
Copy link

Z-Times commented Dec 24, 2024

Hi, when i use the instruction python synthesize_fuse.py -S data/may -M output/may_talkingface --use_train --audio data/may/aud.npy, the error comes: slurmstepd: error: Detected 1 oom-kill event(s) in StepId=3618241.batch cgroup. Some of your processes may have been killed by the cgroup out-of-memory handler.
So i am wondering how much memory should i have to work it. And how large the batch_size is? Thank you.
@Fictionarry
Copy link
Owner

Hi, it should take memory less than the training process. If you successfully run the training, you can run the final inference as well with the same device. To fully load the data into the memory, it takes about 60GB data for the sample of May. However, most of these data are not in use in synthesize_fuse.py. You can manually code the dataloader to prune them.

@Z-Times
Copy link
Author

Z-Times commented Dec 24, 2024

Thank you, I tried several times and finally it worked.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants