You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I have reproduced the algorithm in the paper. In your original paper, you set the inner loop learning rate to 5e-5 and outer-loop learning rate to 5e-6, which from my perspective is too low to have a good learning process. So I reset the parameters and test the MAE of my implementation. And it turns out to be better than your results.
So I wonder whether the hyperparameters in your paper isn't proper or is there any other reasons to lower inner and outer learning rate to that extent.
The text was updated successfully, but these errors were encountered:
waterhorse1
changed the title
Better results with reasonable hyperparameters
My implementation with better results based on reasonable hyperparameters
May 1, 2020
Dear authors,
I have reproduced the algorithm in the paper. In your original paper, you set the inner loop learning rate to 5e-5 and outer-loop learning rate to 5e-6, which from my perspective is too low to have a good learning process. So I reset the parameters and test the MAE of my implementation. And it turns out to be better than your results.
So I wonder whether the hyperparameters in your paper isn't proper or is there any other reasons to lower inner and outer learning rate to that extent.
My implementation is in:
https://github.com/waterhorse1/MELU_pytorch
The text was updated successfully, but these errors were encountered: