Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Strange Results From W2A8 Model #7

Open
cantbebetter2 opened this issue Jun 4, 2024 · 6 comments
Open

Strange Results From W2A8 Model #7

cantbebetter2 opened this issue Jun 4, 2024 · 6 comments

Comments

@cantbebetter2
Copy link

cantbebetter2 commented Jun 4, 2024

Hi,

I executed the W2A8 ImageNet finetuing using the script by directly use n_bits_w = 2, n_bits_a = 8.
But it produces unexpected results in the W2A8 setting. Could you please advise if there are any specific hyperparameters or configurations that need adjustment in the default code to address this problem?

Here are my results.
image

@yoohyewony
Copy link

I also have the same problem.

@cantbebetter2
Copy link
Author

@ThisisBillhe Sorry to bother, but I still can not reproduce the result of W2A8 setting, is there any way to fix this?

@ThisisBillhe
Copy link
Owner

Hi, I will look into this when I am available..working on other project now
BTW, have you successfully trained w4a8 or w4a4 model with EfficientDM? Just want to make sure you setup the project correctly.

@cantbebetter2
Copy link
Author

Hi, I will look into this when I am available..working on other project now BTW, have you successfully trained w4a8 or w4a4 model with EfficientDM? Just want to make sure you setup the project correctly.

Thanks for reply so soon! Yes, I successfully trained the w4a4 model and the results look good. But for W2A8, I really dont know why the result is that strange.

@ThisisBillhe
Copy link
Owner

what about using more steps and epochs during training, e.g., 250 ddim_steps and more epochs?

@cantbebetter2
Copy link
Author

what about using more steps and epochs during training, e.g., 250 ddim_steps and more epochs?

The best result I got is by directly training 20 ddim_steps with 800 epoch, and I got FID 21 and sFID 12, still far away from the paper results. When I double the training epoch, the model training crashes at 1200 epoch. I must doing something wrong, any idea how to reproduce the paper result?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants