Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

have you tried to use bert to improve the performance of JMEE? #4

Open
xiaoya6666 opened this issue Dec 11, 2019 · 11 comments
Open

have you tried to use bert to improve the performance of JMEE? #4

xiaoya6666 opened this issue Dec 11, 2019 · 11 comments

Comments

@xiaoya6666
Copy link

Hi,
Thank you for sharing.
I'm interested if you tried to use bert to improve the performance of JMEE.
I try to reproduce JMEE,but I can't achieve the result of paper.

@xiaoya6666 xiaoya6666 changed the title have you try to use bert to improve the performance of JMEE? have you tried to use bert to improve the performance of JMEE? Dec 11, 2019
@bowbowbow
Copy link
Contributor

Hi @xiaoya6666

I couldn't achieve the performance of the JMEE paper too.
The following results were printed on my console:

python -m enet.run.ee.runner --train "ace-05-splits/train.json"  --test "ace-05-splits/test.json" --dev "ace-05-splits/dev.json" --earlystop 10 --restart 10 --optimizer "adadelta" --lr 1 --webd "./ace-05-splits/glove.6B.300d.txt" --batch 8 --epochs 99999 --device "cuda:0" --out "models/enet-081" --hps "{'wemb_dim': 300, 'wemb_ft': True, 'wemb_dp': 0.5, 'pemb_dim': 50, 'pemb_dp': 0.5, 'eemb_dim': 50, 'eemb_dp': 0.5, 'psemb_dim': 50, 'psemb_dp': 0.5, 'lstm_dim': 220, 'lstm_layers': 1, 'lstm_dp': 0, 'gcn_et': 3, 'gcn_use_bn': True, 'gcn_layers': 3, 'gcn_dp': 0.5, 'sa_dim': 300, 'use_highway': True, 'loss_alpha': 5}"

Epoch 40  dev loss:  3.0913915507072076 
dev ed p:  0.48264984227129337  dev ed r:  0.6375  dev ed f1:  0.5493716337522442 
dev ae p:  0.2878411910669975  dev ae r:  0.1281767955801105  dev ae f1:  0.17737003058103976
Epoch 40  test loss:  2.784576788090766 
test ed p:  0.3360323886639676  test ed r:  0.590047393364929  test ed f1:  0.4282029234737747 
test ae p:  0.20881226053639848  test ae r:  0.12219730941704036  test ae f1:  0.15417256011315417

Epoch 80  dev loss:  3.8771536317780955 
dev ed p:  0.5329949238578681  dev ed r:  0.65625  dev ed f1:  0.5882352941176472 
dev ae p:  0.24006908462867013  dev ae r:  0.15359116022099448  dev ae f1:  0.18733153638814018
Epoch 80  test loss:  3.8047063166558157 
test ed p:  0.3799705449189985  test ed r:  0.6113744075829384  test ed f1:  0.46866485013623976 
test ae p:  0.22857142857142856  test ae r:  0.18834080717488788  test ae f1:  0.20651505838967424

Epoch 120  dev loss:  4.38567394134314 
dev ed p:  0.572992700729927  dev ed r:  0.6541666666666667  dev ed f1:  0.6108949416342413 
dev ae p:  0.23627287853577372  dev ae r:  0.1569060773480663  dev ae f1:  0.18857901726427623
Epoch 120  test loss:  4.248081724495084 
test ed p:  0.40793650793650793  test ed r:  0.6090047393364929  test ed f1:  0.48859315589353614 
test ae p:  0.2297476759628154  test ae r:  0.19394618834080718  test ae f1:  0.21033434650455926

Epoch 160  dev loss:  4.3482774938757345 
dev ed p:  0.574585635359116  dev ed r:  0.65  dev ed f1:  0.6099706744868035 
dev ae p:  0.23304347826086957  dev ae r:  0.14806629834254142  dev ae f1:  0.18108108108108106
Epoch 160  test loss:  4.217268991275621 
test ed p:  0.41423948220064727  test ed r:  0.6066350710900474  test ed f1:  0.49230769230769234 
test ae p:  0.23285714285714285  test ae r:  0.1827354260089686  test ae f1:  0.20477386934673367

Epoch 199  dev loss:  4.394452537438701 
dev ed p:  0.5831775700934579  dev ed r:  0.65  dev ed f1:  0.6147783251231527 
dev ae p:  0.23861566484517305  dev ae r:  0.14475138121546963  dev ae f1:  0.1801925722145805
Epoch 199  test loss:  4.19947991046335 
test ed p:  0.4169381107491857  test ed r:  0.6066350710900474  test ed f1:  0.4942084942084942 
test ae p:  0.2422907488986784  test ae r:  0.18497757847533633  test ae f1:  0.20979020979020976

I think word embedding of JMEE can be replaced with BERT, but I haven't tried it yet.

@ll0ruc
Copy link

ll0ruc commented Dec 13, 2019

@xiaoya6666 I also try to reproduce JMEE,but can't achieve the result of paper. I try to use Bert to replace the word embedding of JMEE ,but ed F1 just achieve 0.69. It seems that GCN in JMME paper actually can't bring very obvious performance. Maybe I combine GCN and bert in a inappropriate way.

@xiaoya6666
Copy link
Author

Hi @xiaoya6666

I couldn't achieve the performance of the JMEE paper too.
The following results were printed on my console:

python -m enet.run.ee.runner --train "ace-05-splits/train.json"  --test "ace-05-splits/test.json" --dev "ace-05-splits/dev.json" --earlystop 10 --restart 10 --optimizer "adadelta" --lr 1 --webd "./ace-05-splits/glove.6B.300d.txt" --batch 8 --epochs 99999 --device "cuda:0" --out "models/enet-081" --hps "{'wemb_dim': 300, 'wemb_ft': True, 'wemb_dp': 0.5, 'pemb_dim': 50, 'pemb_dp': 0.5, 'eemb_dim': 50, 'eemb_dp': 0.5, 'psemb_dim': 50, 'psemb_dp': 0.5, 'lstm_dim': 220, 'lstm_layers': 1, 'lstm_dp': 0, 'gcn_et': 3, 'gcn_use_bn': True, 'gcn_layers': 3, 'gcn_dp': 0.5, 'sa_dim': 300, 'use_highway': True, 'loss_alpha': 5}"

Epoch 40  dev loss:  3.0913915507072076 
dev ed p:  0.48264984227129337  dev ed r:  0.6375  dev ed f1:  0.5493716337522442 
dev ae p:  0.2878411910669975  dev ae r:  0.1281767955801105  dev ae f1:  0.17737003058103976
Epoch 40  test loss:  2.784576788090766 
test ed p:  0.3360323886639676  test ed r:  0.590047393364929  test ed f1:  0.4282029234737747 
test ae p:  0.20881226053639848  test ae r:  0.12219730941704036  test ae f1:  0.15417256011315417

Epoch 80  dev loss:  3.8771536317780955 
dev ed p:  0.5329949238578681  dev ed r:  0.65625  dev ed f1:  0.5882352941176472 
dev ae p:  0.24006908462867013  dev ae r:  0.15359116022099448  dev ae f1:  0.18733153638814018
Epoch 80  test loss:  3.8047063166558157 
test ed p:  0.3799705449189985  test ed r:  0.6113744075829384  test ed f1:  0.46866485013623976 
test ae p:  0.22857142857142856  test ae r:  0.18834080717488788  test ae f1:  0.20651505838967424

Epoch 120  dev loss:  4.38567394134314 
dev ed p:  0.572992700729927  dev ed r:  0.6541666666666667  dev ed f1:  0.6108949416342413 
dev ae p:  0.23627287853577372  dev ae r:  0.1569060773480663  dev ae f1:  0.18857901726427623
Epoch 120  test loss:  4.248081724495084 
test ed p:  0.40793650793650793  test ed r:  0.6090047393364929  test ed f1:  0.48859315589353614 
test ae p:  0.2297476759628154  test ae r:  0.19394618834080718  test ae f1:  0.21033434650455926

Epoch 160  dev loss:  4.3482774938757345 
dev ed p:  0.574585635359116  dev ed r:  0.65  dev ed f1:  0.6099706744868035 
dev ae p:  0.23304347826086957  dev ae r:  0.14806629834254142  dev ae f1:  0.18108108108108106
Epoch 160  test loss:  4.217268991275621 
test ed p:  0.41423948220064727  test ed r:  0.6066350710900474  test ed f1:  0.49230769230769234 
test ae p:  0.23285714285714285  test ae r:  0.1827354260089686  test ae f1:  0.20477386934673367

Epoch 199  dev loss:  4.394452537438701 
dev ed p:  0.5831775700934579  dev ed r:  0.65  dev ed f1:  0.6147783251231527 
dev ae p:  0.23861566484517305  dev ae r:  0.14475138121546963  dev ae f1:  0.1801925722145805
Epoch 199  test loss:  4.19947991046335 
test ed p:  0.4169381107491857  test ed r:  0.6066350710900474  test ed f1:  0.4942084942084942 
test ae p:  0.2422907488986784  test ae r:  0.18497757847533633  test ae f1:  0.20979020979020976

I think word embedding of JMEE can be replaced with BERT, but I haven't tried it yet.

@bowbowbow exm, could you tell me how to set the parameters in this bert model? thank you very much!

@xiaoya6666
Copy link
Author

@xiaoya6666 I also try to reproduce JMEE,but can't achieve the result of paper. I try to use Bert to replace the word embedding of JMEE ,but ed F1 just achieve 0.69. It seems that GCN in JMME paper actually can't bring very obvious performance. Maybe I combine GCN and bert in a inappropriate way.

exm, do you have an overfitting problem when you replaced word embedding into BERT in JMEE? and What is the F1 score of 0.69 you got, argument classification or trigger classtication? I found it will have very serious overfitting problem when i add the entity type embedding into the model.

@xiaoya6666
Copy link
Author

xiaoya6666 commented Dec 31, 2019 via email

@xiaoya6666
Copy link
Author

xiaoya6666 commented Jan 2, 2020 via email

@ll0ruc
Copy link

ll0ruc commented Jan 2, 2020

@xiaoya6666 您好,你的触发词70%是bert的代码还是jmee的,能加下QQ详聊吗(2512156864)

@xiaoya6666
Copy link
Author

xiaoya6666 commented Jan 3, 2020 via email

@edchengg
Copy link

edchengg commented Mar 18, 2020

I also got F1 for trigger classification around 69 with BERT + linear classification layer. But this is way below results reported from paper (https://www.aclweb.org/anthology/P19-1522.pdf, https://www.aclweb.org/anthology/K19-1061.pdf). They got F1 ranges from 73 - 80. I don't think a CRF will help a lot.
Anyone experience the same problem?

@Wangpeiyi9979
Copy link

I also got F1 for trigger classification around 69 with BERT + linear classification layer. But this is way below results reported from paper (https://www.aclweb.org/anthology/P19-1522.pdf, https://www.aclweb.org/anthology/K19-1061.pdf). They got F1 ranges from 73 - 80. I don't think a CRF will help a lot.
Anyone experience the same problem?

I sent emails to all the authors with 80% experimental results, but no one replied to me. Moreover, there was no detail of hyparameters in the paper, so I still could not reproduce the effect. I think their experimental data is false

@znsoftm
Copy link

znsoftm commented Mar 8, 2022

Remember, most of the announcements in papers which are made by a new author always always are not able to be reproduced.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

6 participants