Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cannot reproduce the baseline score of question answering with transformers v4.19.2 #9

Open
kumapo opened this issue Feb 1, 2023 · 2 comments

Comments

@kumapo
Copy link

kumapo commented Feb 1, 2023

I tried to reproduce the baseline score with run_squad.py parameters you provided and patched transformers v4.19.2.
but the result score in eval_results.json is quite low compared to the baseline.

    "exact": 42.30076542098154,
    "f1": 42.390814948221525,

based on fune-tuning/README.md, I think you confirmed that transformers v4.19.2 worked.
How was the score then?

I'm attaching the requirements.txt and eval_results.json when I tested with transformers v4.19.2.

@tomohideshibata
Copy link
Contributor

Thank you for your report. I will check it.

Which pretrained model have you used?

@kumapo
Copy link
Author

kumapo commented Feb 1, 2023

@tomohideshibata

Thank you for quick reply.
I used cl-tohoku/bert-base-japanese-v2.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants