-
Notifications
You must be signed in to change notification settings - Fork 84
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
ListOps performance #15
Comments
Have you solved this problem? I have a similar issue. |
Hi @sihyun-yu, I was not able to solve it |
It seems the issue is fixed with the latest code push. Please add comment if the issue still comes up. |
I found either lowering the learning rate or increasing the batch size was useful for this task. I think their hyperparameters are for a large effective batch size because they run on TPU. |
I am still getting the same problem, my validation during training is high on the listops, but when running test_only option, I am getting very low accuracy! |
The problem is that the data is shuffled every time the code is ran, so the tokens are changed when running the test script giving a random accuracy. |
@BalloutAI Hi, I also found this issue: high training accuracy, low test accuracy; I also found if I run training process multiple times, sometimes the model cannot even converge. Could you explain your idea a little bit more? Thank you. |
On running the ListOps task as-is from the repo, I got a validation performance similar to that reported in the paper but the test performance on results.json is very low:
I saw that the code is saving the model from the last checkpoint as compared to the model with the best validation performance. Could you detail the evaluation setup used in the paper i.e. for the paper do you evaluate the model from the last checkpoint of from the best validation checkpoint?
Thank you very much! :-)
The text was updated successfully, but these errors were encountered: