This repository has been archived by the owner on Jul 7, 2023. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 3.5k
Cannot reproduce "transformer" paper's result with BPE32k #309
Comments
Have you made sure that you preprocessed the translatable text in the same way like the training data? You'll need to tokenize and apply bpe before translating. |
I believe I tried both newstest2015.en, and newstest2015.tok.clean.bpe.32000.en. |
Here is what I got. Most paper reports bleu scores 0~100, is this really 0.31 or 31.3? |
The original BLEU was defined as 0-1, but most people report it as 0-100. Note that approx_bleu is not BLEU (usually it is about 1.5 times higher). |
0.313501 / 1.5 is still a reasonable number, I suppose |
Sign up for free
to subscribe to this conversation on GitHub.
Already have an account?
Sign in.
I can train translate_ende_wmt32k and get reasonable bleu with transformer_big model. But if I train translate_ende_wmt_bpe32k, it doesn't translate anything meaningful.
I've tried several times, all with default settings. Any suggestions?
The text was updated successfully, but these errors were encountered: