-
Notifications
You must be signed in to change notification settings - Fork 67
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Transformer appears to be outdated, lower BLEU #112
Comments
Hi @ableiweisss Regarding the BLEU score, there was some discussion before also whether it should be kept the same as the score reported in the original paper. My understanding is that, once the BLEU score has crossed value of 22~23, further improvement in the BLEU score is marginal with respect to the number of training epochs and can bring certain inconsistency in training time due to random seed ( see mlcommons/training#125 ). |
Hi @tremblerz |
Maybe we will take a deeper look for 1.0 .... thanks for pointing that out.
At this point, for 0.5, maybe we could just let this version be.
…On Thu, Oct 11, 2018 at 9:13 AM ableiweisss ***@***.***> wrote:
Hi @tremblerz <https://github.com/tremblerz>
The official TensorFlow repo reports a BLEU score of 27.7 for base and
28.9 for big. MLPerf reports 25, which is significantly lower. Is this due
to code differences or rather as a target simply related to minimizing the
number of epochs?
—
You are receiving this because you are subscribed to this thread.
Reply to this email directly, view it on GitHub
<mlcommons/policies#112 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AA1h-ubTh2v13HmoKGvE8p1eIbd8NnzPks5uj24YgaJpZM4XXRv4>
.
--
-Debo~
|
SWG: |
The Transfomer code in MLPerf is a bit outdated, and has a BLEU score lower than the official TensorFlow version:
https://github.com/tensorflow/models/tree/master/official/transformer
The text was updated successfully, but these errors were encountered: