An Iterative Knowledge Transfer NMT System for WMT20 News Translation Task

2020
This paper describes our submission to the WMT20 news translation shared task in English to Japanese direction. Our main approach is based on transferring knowledge of domain and linguistic characteristics by pre-training the encoder-decoder model with large amount of in-domain monolingual data through unsupervised and supervised prediction task. We then fine-tune the model with parallel data and in-domain synthetic data, generated with iterative back-translation. For additional gain, we generate final results with an ensemble model and re-rank them with averaged models and language models. Through these methods, we achieve +5.42 BLEU score compare to the baseline model.
    • Correction
    • Source
    • Cite
    • Save
    14
    References
    1
    Citations
    NaN
    KQI
    []
    Baidu
    map