Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Mismatch in Mesh Video Output after Training TALKSHOW and used in trained make your anchor model #28

Open
FD01076 opened this issue Jan 28, 2025 · 1 comment

Comments

@FD01076
Copy link

FD01076 commented Jan 28, 2025

I used both SHOW and TALKSHOW.
I had a video that was split into 9-second segments, and I used SHOW to generate mesh videos for each segment. For each split, I used the final_all.pkl files and new audio (WAV) files in TALKSHOW, and trained three models: body pixel, face, and body-vq. However, the mesh video generated after training doesn't match what I got from SHOW, and it doesn't accurately reflect how the person is positioned in the original video provided to SHOW and because of this the anchor output videos are distorted. What can be the issue, how to train TALKSHOW for exact same mesh videos and mesh position as we gor from SHOW, on which the unet and controlnet are trained.

@vibhu-raturi
Copy link

Hello, how can i get mesh videos from TALKSHOW which resembles to the pose and human which is used to train unet and controlnet in this repo. The mesh from talkshow is used for inference we get a distorted video output

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants