You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Thank you for this amazing work and making the code available online. I was able to train the tagger and parser, and could gather results for both of them. For Parser, I was able to train and evaluate the model but I could only get LAS and UAS metrics. Would it be possible to share a way to get precision, recall and F1 score for the parser? I built my own evaluation script however, I couldn't get the numbers to match with reported numbers. So I may be doing something wrong there.
I also wanted to clarify something in this file. In lines 224, 239, 272 and 292, we are writing an ID of the token to the file. However, for each recipe, the token must start from 1 according to CoNLL-U notations. But, when I look at the generated file, the IDs go from 1 to 3827 for the tagger. If we feed this generated tagger results to the Parser, it will treat entire file as a single recipe. So we may have to reset ID variable i in read_prediction.py for each recipe. Let me know if there is something wrong in my understanding.
The text was updated successfully, but these errors were encountered:
@TheresaSchmidt would it be possible to share all the folds used for training? Since the table reports numbers of Yamakata'20, is it safe to assume that 10-fold validation is performed? I would like to ensure that the 1 fold available on GitHub is not biased, resulting in performance drop. I am still assuming that I am probably doing something wrong with my evaluation. So any direction to calculate Precision, recall and F1 for the parser will be highly appreciated.
Hi team,
Thank you for this amazing work and making the code available online. I was able to train the tagger and parser, and could gather results for both of them. For Parser, I was able to train and evaluate the model but I could only get LAS and UAS metrics. Would it be possible to share a way to get precision, recall and F1 score for the parser? I built my own evaluation script however, I couldn't get the numbers to match with reported numbers. So I may be doing something wrong there.
I also wanted to clarify something in this file. In lines 224, 239, 272 and 292, we are writing an ID of the token to the file. However, for each recipe, the token must start from 1 according to CoNLL-U notations. But, when I look at the generated file, the IDs go from 1 to 3827 for the tagger. If we feed this generated tagger results to the Parser, it will treat entire file as a single recipe. So we may have to reset ID variable
i
in read_prediction.py for each recipe. Let me know if there is something wrong in my understanding.The text was updated successfully, but these errors were encountered: