-
Notifications
You must be signed in to change notification settings - Fork 84
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
NaN Loss on train #13
Comments
Please specify the model you are using. |
Hi, thank you for the response. I am using essentially the example from the read me, but with my own game. I figured out if I lower the learning rate dramatically to .001, it fixes the problem. Could there be something in the way I've designed my game that would cause the discrepancy? I'd rather use a slightly larger learning rate if possible. Thanks again |
Some sort of normalization of your reward might help. Paste your code here, I will take a look. |
I found that in logs before loss became nan Epoch 143/1000 | Loss 2214.4102 | Epsilon 0.00 | Win count 65 |
I have been getting NaN losses when I try to train my agent on a game. I tracked it back to the get_batch function in memory - Y (the output of the model's predictions) turns to all NaNs about halfway through the first epoch. I haven't been able to figure it out from there, though.
Any suggestion would be much appreciated. This package is fantastic!
The text was updated successfully, but these errors were encountered: