-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Experiment Results' Reproduce using provided Checkpoint #3
Comments
Hi, thank for your feedback! The responses to your question are as follows:
|
I made a submission and now the results of the given checkpoint are consistent with README (EM / F0.5 : 34.10 / 45.48). |
Hello! |
Hi, it seems you have used multiple values for |
#!/bin/bash Copyright 2022 The ZJU MMF Authors (Lvxiaowei Xu, Jianwang Wu, Jiawei Peng, Jiayu Fu and Ming Cai *).Licensed under the Apache License, Version 2.0 (the "License");you may not use this file except in compliance with the License.You may obtain a copy of the License athttp://www.apache.org/licenses/LICENSE-2.0Unless required by applicable law or agreed to in writing, softwaredistributed under the License is distributed on an "AS IS" BASIS,WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.See the License for the specific language governing permissions andlimitations under the License.Train and Test for STG-JointGlobal Variable (!!! SHOULD ADAPT TO YOUR CONFIGURATION !!!)CUDA_ID=1 Roberta-base-chinese can be downloaded at https://github.com/ymcui/Chinese-BERT-wwm#PLM_PATH=/datadisk2/xlxw/Resources/pretrained_models/roberta-base-chinese/ # pretrained-model path JOINT_CHECK_DIR=1021_jointmodel_stg STEP 1 - PREPROCESS DATASET#DATA_BASE_DIR=dataset #python preprocess_data.py --mode normal --err_only True STEP 2 - TRAIN STG-Joint MODEL#python joint_stg.py --mode train STEP 3 - TRAIN STG-Joint MODELpython joint_evaluate.py --mode test --gpu_id ${CUDA_ID} --seed ${SEED} run: sh run_stg_joint.sh |
The error I can currently find is that you have commented out two parameters, |
Sorry! In evaluate_joint_config.py, I forgot to modify the parameters # Pretrained Model Params
But I encountered a new problem, Some weights of the model checkpoint at ./pretrained_models/chinese-roberta-wwm-ext/ were not used when initializing BertModel: ['cls.seq_relationship.bias', 'cls.predictions.transform. LayerNorm.bias', 'cls.predictions.transform.dense.bias', 'cls.predictions.decoder.weight', 'cls.predictions.bias', 'cls.seq_relationship.weight', 'cls.predictions.transform. dense.weight', 'cls.predictions.transform.LayerNorm.weight']
RuntimeError: Error(s) in loading state_dict for JointModel: The pre-trained language model uses https://huggingface.co/hfl/chinese-roberta-wwm-ext/ |
Thank you, I solved it. In the py file, set the Number of Tagger Classes to 5, and the Number of Max Token Generation to 5. Thank you very much |
You're welcome. |
Hello!
I downloaded the trained checkpoint in README for inferring on the test set to reproduce the results.
The results given in README are (EM / F0.5 : 34.10 / 45.48). But my results (utilizing the run_stg_joint.sh) are (EM / F0.5 : 50.5 / 37.9). This difference cannot be neglected.
Actually, I adjusted some code while inferring.
The text was updated successfully, but these errors were encountered: