Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine Translation
This repo hosts the code to accompany the camera-ready version of “Cross-Attention is All You Need: Adapting Pretrained Transformers for Machine Translation” in EMNLP 2021. Setup We provide our scripts and modifications to Fairseq. In this section, we describe how to go about running the code and, for instance, reproduce Table 2 in the paper. Data To view the data as we prepared and used it, switch to the main branch. But we recommend cloning code from this branch to […]
Read more