-
Notifications
You must be signed in to change notification settings - Fork 54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Pretraining works fine, but rl training stays at 0 Accuracy #1
Comments
Try to reduce the learning rate plz. I‘m not sure what's going wrong, but it worth a try. |
thank you for your quick reply. I tried that over the weekend,and a lower learning rate (0.00002) indeed helped a little bit. Would it be possible for you to share a pretraining model and a set of hyperparameters that work for you? |
I have just rerun the pretraining for the NYT10 dataset with
and gets about 62 F1 on the test set. Here's the log output
Then I train the model using RL with
and the F1 score continues to rise,
The model is still training, only the logs of the first 3 epochs are quoted here. Environment:
|
Similar question.@BenjaminWinter Train the model with Dev and test set F1 are 0, training accuracy is 0 on each training epoh while the loss is continuing decline Environment: |
@misaki-sysu |
I have pre-trained first, and the learning rate was 0.00002, but the accuracy was 0 when I tested the test set. |
I meet the same problem where F1 in training process is good but test set F1 are 0. Have you solved this problem? |
What are you change code for 0.3 to 1.0.1?Can you give me you rewrite code?Thank you very much! |
@truthless11 I have encountered this problem. What is the reason and how to solve it THCudaCheck FAIL file=C:\w\1\s\tmp_conda_3.7_055306\conda\conda-bld\pytorch_1556690124416\work\torch/csrc/generic/StorageSharing.cpp line=245 error=63 : OS call failed or operation not supported on this OS |
@YiYingsheng Excuse me, Have you solved it? |
Running:
The RL training doesn't work for me for the NYT10 dataset (havent checked other yet).
I first ran pretraining for 10 epochs with:
python main.py --epochPre 10 --numprocess 8 --datapath ../data/NYT10/ --pretrain True
which gets roughly 58 F1 on the test set, and then afterwards try the RL training with:
python main.py --epochRL 10 --numprocess 8 --start checkpoints/model_HRL_10 --datapath ../data/NYT10/
I stopped RL training after 3 epochs because not only was dev and test set F1 at 0, even training accuracy is 0.
Loss started at around 30, then after only 60 batches decreases to about -20, then slowly increases again and ends up hovering around -0.00005
Checking the optimize() method all reward arrays contain either straight 0's or negative numbers.
The text was updated successfully, but these errors were encountered: