Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

I broke it #208

Open
CoderUnkn0wn opened this issue Jul 23, 2024 · 5 comments
Open

I broke it #208

CoderUnkn0wn opened this issue Jul 23, 2024 · 5 comments

Comments

@CoderUnkn0wn
Copy link

image

@Sanj-bot
Copy link

how did u break it? @CoderUnkn0wn

@CoderUnkn0wn
Copy link
Author

You can see in the image all of the settings I used. Then I just let it run and suddenly everything was NaN.

@Sanj-bot
Copy link

coooll

@TuchoTurco
Copy link

this is not an issue, it is normal behavior.

@frozenfrank
Copy link

Second Witness

"Breaking" the model by causing the weights to blow up to infinity isn't difficult to do. Setting the learning rate high at the start of a complex model causes the weights to explode to high values, evaluate to Infinity, and then the next epoch reports them all as "NaN."

Discussion

I think it is normal behavior— it's just the way the math works out. Setting the learning rate so high is bound to give inaccurate results.

This kind of playground is the perfect place to be able to play with settings like that and experience how the models behave when given diverse kinds of inputs. The researcher student will quickly learn to be more intentional about the use of a high learning rate!

Minimal Reproduction

I was able to reproduce the behavior previously reported by doing the following:

  1. Set the learning rate to the maximum
  2. Max out the number of hidden layers and hidden nodes
  3. Set the problem type to "regression"
  4. Step through the training one epoch at a time
  5. Discover the values overflowing within 14 training epochs

Walkthrough

Epoch 1: Loss stats already reporting a value with 46 zeros behind it

Screenshot 2024-09-30 at 10 56 52 PM

Epoch 7: Loss stats overflowing and reporting "Infinity"

Screenshot 2024-09-30 at 10 57 36 PM

Epoch 14: Loss stats reporting NaN

Screenshot 2024-09-30 at 10 58 01 PM

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants