Issue #131 – Measuring and Increasing Context Usage in Context-Aware NMT

20 May21 Issue #131 – Measuring and Increasing Context Usage in Context-Aware NMT in Model improvement, The Neural MT Weekly Author: Dr. Patrik Lambert, Senior Machine Translation Scientist @ Iconic Introduction Context-aware neural machine translation (NMT) is a topic which has often been covered in this blog, for its application to domain adaptation or document-level NMT (see issues #15, #31, #34, #39, #98, #128). However, most papers on context-aware NMT present approaches which have the ability to leverage context information, […]

Read more

Issue #98 – Unified and Multi-encoders for Context-aware Neural MT

10 Sep20 Issue #98 – Unified and Multi-encoders for Context-aware Neural MT Author: Dr. Patrik Lambert, Senior Machine Translation Scientist @ Iconic Introduction Context-aware Neural MT uses context information to perform document-level translation or domain adaptation. The context of surrounding sentences allows the model to capture discourse phenomena. The context of similar sentences can also be useful to dynamically adapt the translation to a domain. In this post, we take a look at two papers which compare uni-encoder and multi-encoder […]

Read more