Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RoFormer_pytorch继续预训练性能下降 #16

Open
hflserdaniel opened this issue Nov 2, 2021 · 3 comments
Open

RoFormer_pytorch继续预训练性能下降 #16

hflserdaniel opened this issue Nov 2, 2021 · 3 comments

Comments

@hflserdaniel
Copy link

非常感谢您在huggingface上的贡献!
在我的项目中,我尝试利用transformers/example/pytorch/language_modeling中的mlm脚本对RoFormer继续在领域文本上预训练,但是预训练后的表现出现下降的问题。作为对比,对roberta继续预训练能够得到提升。
可以请教下具体的原因嘛?是否和脚本中调用的AutoModelForMaskedLM有关?
再次感谢!

@JunnYu
Copy link
Owner

JunnYu commented Nov 2, 2021

训练的是什么模型?char级别的还是word级别的模型

@JunnYu
Copy link
Owner

JunnYu commented Nov 2, 2021

模型搭建的是没有问题,https://github.com/ZhuiyiTechnology/roformer/blob/main/train.py 可以参考原版的keras代码继续预训练,然后再把模型转成pytorch试试

@hflserdaniel
Copy link
Author

训练的是什么模型?char级别的还是word级别的模型

使用的是word级别的模型和对应的词表及分词器

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants