Background: Paragram Embeddings

Paragram embeddings are a type of word embeddings focused on the semantics of individual words. They were created by instituting various semantic similarity constraints from the Paraphrase Database onto GloVe vectors in order to improve their semantic content.

What did you try?

After last week, we finally made a shift to multi-turn dialogues as that is what we cared about the most and what we wanna optimize on.

Instead of just having the Paragram embedding representations of the user utterance passed into the model, we wanted to go ahead and see if there were any performance improvements to including the ELMO embeddings to the utterance representation.

We focused on modifying the user utterance representation.

We focused on modifying the user utterance representation.

By adding ELMO, we hoped to take advantage of its contextual word understanding within the dialogues that the user provides. By doing so, we hoped that the accuracy of our belief state tracker would increase as the model would have a better understanding of the user's dialogue.

To do so, we had to make several modifications to our pipeline to support a trainable ELMo (softmax-normalized weights on the hidden representations from the language model and

​ the task-specific scaling factor)

Exciting Results

We ran our experiment to train ELMo vs no ELmo for 120 epochs for the multi turn and noticed some interesting results.

Issues we ran into

Engineering Efforts