You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I noticed that lr in code is 0.01 while paper is 0.05 with adagrad, and I tried with 0.05 to train the model but train loss doesn't decrease at all, may be due to the high lr ?
Why you set lr to 0.01 ? And since lr is different, may be there is a bug in code ?
The text was updated successfully, but these errors were encountered:
I noticed that lr in code is 0.01 while paper is 0.05 with adagrad, and I tried with 0.05 to train the model but train loss doesn't decrease at all, may be due to the high lr ?
Why you set lr to 0.01 ? And since lr is different, may be there is a bug in code ?
The text was updated successfully, but these errors were encountered: