-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Performance looks similar #3
Comments
The findings match the paper. The performance of batch renorm is just slightly higher than batchnorm. Renorm is only faster to get to that high score than batchnorm. |
Thanks. Sorry I did not give more information. I mean the benefit in small batch size as figure 2 and figure 3. For larger batch size, I totally agree with you that they are similar. Could you measure performance in small batch size case? |
Hmm, I did try with batch size of 4 and 8, and the results were similar to that with larger batch sizes. Batch renorm quickly hits a high score whereas for batch norm, it takes more time. I think I removed all the code and images for those tests during a cleanup of the codebase though. |
I got similar results for batch normalization and re normalization on very small (1-2) batch sizes which is inconsistent with the results in the paper. |
Thanks for your implementation of batch renormalization. I saw that your performance of batch renorm and batch norm are similar. Do you check the performance with a simple network as in the paper? The paper shows a simple network with high gain between batch renorm and batch norm.
The text was updated successfully, but these errors were encountered: