Skip to content

Loss suddenly increases #11

Discussion options

You must be logged in to vote

It seems that the problem lies in the Adam optimizer. ( ᐛ )
Accordding to:

https://discuss.pytorch.org/t/loss-suddenly-increases-using-adam-optimizer/11338

In anycase, the problem with Adam is that it uses moving average in the denominator term. So if the gradients get really small and the whole of denominator will be small. Since the gradients are already small, the denominator results in blowup thus pushing you very far away hence huge loss. You may have a look at https://openreview.net/forum?id=ryQu7f-RZ 560 . I think there are many recent methods which avert this problem including AMSGrad (in the earlier mentioned paper), Hyper-gradient descent (for Adam) etc. Also look for comment…

Replies: 1 comment 2 replies

Comment options

You must be logged in to vote
2 replies
@YukinoshitaLove
Comment options

YukinoshitaLove Oct 24, 2024
Collaborator Author

@YukinoshitaLove
Comment options

YukinoshitaLove Oct 24, 2024
Collaborator Author

Answer selected by YukinoshitaLove
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
1 participant