Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

VicReg implementation gives "nan" loss #335

Open
ManafMukred opened this issue Jun 21, 2023 · 3 comments
Open

VicReg implementation gives "nan" loss #335

ManafMukred opened this issue Jun 21, 2023 · 3 comments

Comments

@ManafMukred
Copy link

ManafMukred commented Jun 21, 2023

I was trying to explore other algorithms like vicreg using LAMB & LARS optimizer, but in both cases the loss is "nan"

`Epoch 1/200
175/175 - 49s - loss: nan - proj_std: nan - val_loss: nan - val_proj_std: nan - binary_accuracy: 0.1000 - 49s/epoch - 281ms/step

Epoch 2/200
175/175 - 36s - loss: nan - proj_std: nan - val_loss: nan - val_proj_std: nan - binary_accuracy: 0.1000 - 36s/epoch - 203ms/step

Epoch 3/200
175/175 - 36s - loss: nan - proj_std: nan - val_loss: nan - val_proj_std: nan - binary_accuracy: 0.1000 - 36s/epoch - 204ms/step
`

any suggestions?

@owenvallis
Copy link
Collaborator

Hi @ManafMukred, can you share some more details about the model architecture, hyperparameters, and dataset?

@ManafMukred
Copy link
Author

@owenvallis I used the same notebook here , and I tried to use the LARS optimizer also with lr = 0.2 * int(BATCH_SIZE / 256) same as the paper says.
but I get the same error

@owenvallis
Copy link
Collaborator

Thanks for the details, I'll take a look and see if I can repro the issue on my side.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants