Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Any way to apply Lora result of training script to Diffusers model? #95

Open
congdm opened this issue Jun 16, 2024 · 1 comment
Open

Comments

@congdm
Copy link

congdm commented Jun 16, 2024

Hi, at the moment I noticed that the transformer model in huggingface-cli Tencent-Hunyuan/HunyuanDiT has different structure than the transformer model in Diffusers pipeline. This makes me cannot apply lora such as links to Diffusers pipeline.

For example, the transformer in Tencent-Hunyuan/HunyuanDiT has fused qkv layer while the one in diffusers is unfused. They also different modules naming too.
In long run, having to train lora separately for both types of transformer is very impractical.

@xljh0520
Copy link

Thank you for your attention to our work.
Due to certain compatibility adjustments and modifications made by the diffuser, there may be issues with the model state_dict keys not matching. We have provided a code example in lora README that demonstrates how to merge LoRA weights through keyword matching.
We hope this can help you.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants