View on GitHub

Research Review Notes

Summaries of academic research papers

Improved Variational Autoencoders for Text Modeling using Dilated Convolutions


Idea

The paper presents an alternative architecture to LSTM based VAEs. As shown in an earlier paper, LSTM-VAEs don’t have a significant advantage over LSTM language models. The authors address this by using a dilated CNN decoder to vary the conditioning context of the decoder. The hypothesis is the the typical collapse of the loss function in favor of the KL-divergence term could be addressed by varying the contextual capacity of the decoder.

Method

Observations