Issue #63 – Neuron Interaction Based Representation Composition for Neural Machine Translation

05 Dec19

Issue #63 – Neuron Interaction Based Representation Composition for Neural Machine Translation

Author: Dr. Rohit Gupta, Sr. Machine Translation Scientist @ Iconic

Transformer models are state of the art in Neural Machine Translation. In this blog post, we will take a look at a recently proposed approach by Li et al (2019) which further improves upon the transformer model by modeling more neuron interactions. Li et al (2019) claim that their approach models better encoder representation and captures semantic and syntactic properties better than the baseline transformer. 

Bilinear Pooling:

Bilinear pooling (Tenenbaum and Freeman (2000)) is defined as an outer product of two representation vectors followed by a linear projection. For the transformer model we have multi-layer and multi-head representation vectors. These vectors are concatenated to form the input to bilinear pooling. 

Bilinear pooling only encodes second-order (i.e., multiplicative) interactions among individual neurons, therefore, they proposed the extended bilinear pooling which also includes first-order representations. The figure below illustrates bilinear and extended bilinear pooling. 

bilinear and extended bilinear poolingTo finish reading, please visit source site

Leave a Reply