Machine Translation On Wmt2014 English German

评估指标

BLEU score
Hardware Burden
Operations per network pass

评测结果

各个模型在此基准测试上的表现结果

模型名称
BLEU score
Hardware Burden
Operations per network pass
Paper TitleRepository
Transformer Big + adversarial MLE29.52Improving Neural Language Modeling via Adversarial Training-
MAT---Multi-branch Attentive Transformer-
AdvAug (aut+adv)29.57--AdvAug: Robust Adversarial Augmentation for Neural Machine Translation-
CMLM+LAT+4 iterations27.35Incorporating a Local Translation Mechanism into Non-autoregressive Translation-
FlowSeq-large (IWD n = 15)22.94FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow-
Transformer (ADMIN init)30.1--Very Deep Transformers for Neural Machine Translation-
MUSE(Parallel Multi-scale Attention)29.9MUSE: Parallel Multi-Scale Attention for Sequence to Sequence Learning-
Transformer-DRILL Base28.1Deep Residual Output Layers for Neural Language Generation-
Transformer Big with FRAGE29.11FRAGE: Frequency-Agnostic Word Representation-
GLAT25.21--Glancing Transformer for Non-Autoregressive Neural Machine Translation-
PartialFormer29.56--PartialFormer: Modeling Part Instead of Whole for Machine Translation-
Bi-SimCut30.78--Bi-SimCut: A Simple Strategy for Boosting Neural Machine Translation-
Transformer + SRU28.434GSimple Recurrent Units for Highly Parallelizable Recurrence-
PBMT20.7--
Local Joint Self-attention29.7Joint Source-Target Self Attention with Locality Constraints-
Lite Transformer26.5--Lite Transformer with Long-Short Range Attention-
Average Attention Network (w/o FFN)26.05--Accelerating Neural Transformer via an Average Attention Network-
Unsupervised NMT + Transformer17.16Phrase-Based & Neural Unsupervised Machine Translation-
KERMIT28.7KERMIT: Generative Insertion-Based Modeling for Sequences-
T2R + Pretrain28.7Finetuning Pretrained Transformers into RNNs-
0 of 91 row(s) selected.