View on GitHub

Research Review Notes

Summaries of academic research papers

Controlling Politeness in Neural Machine Translation via Side Constraints


Idea

The authors intend to control politeness while translating one language to another by using honorifics. The politeness is not grammatically implied in the source text, but should be predicted in the target text.

The term side constraints in this paper is just another way of saying attributes that the generated text is conditioned on.

Background

The method relies on the attention model for neural translation. Each word $y_i$ is predicted based on 3 variables: the recurrent state $s_i$, the previously predict word $y_{i - 1}$ and the context vector $c_i$.

The model relies on a GRU-backed bi-directional sequence autoencoder and the hidden state from both directions are combined to form the annotation vector. The weight of each of these annotation vectors is computed via a single layer feed-forward neural network. The weights are computed through an alignment model $\alpha_{ij}$, which models the probability that $x_j$ is aligned with $y_i$.

A beam search decoder is used for the generation phase.

Method

Observations

Side constraints can be applied to other phenomena too, given that we already possess a classifier that can be used to label the training set without manual annotations.