View on GitHub

Research Review Notes

Summaries of academic research papers

Word Translation Without Parallel Data


Idea

The main idea of the paper is to separately and unsupervisedly learn word embeddings from 2 different languages using a method like Word2Vec and then align the embedding spaces of the 2 languages using adversarial training. This allows the similar words in each language to be mapped to roughly the same point on the manifold. This manifold can then be queried using one language’s words and the corresponding word in the other languages closest in the embedding space should be the transalated word. In this way, the authors propose building an unsupervised bilingual dictionary.

Method

Experiments

Observations