Learning to Rank Question-Answer Pairs using Hierarchical Recurrent Encoder with Latent Topic Clustering
About
In this paper, we propose a novel end-to-end neural architecture for ranking candidate answers, that adapts a hierarchical recurrent neural network and a latent topic clustering module. With our proposed model, a text is encoded to a vector representation from an word-level to a chunk-level to effectively capture the entire meaning. In particular, by adapting the hierarchical structure, our model shows very small performance degradations in longer text comprehension while other state-of-the-art recurrent neural network models suffer from it. Additionally, the latent topic clustering module extracts semantic information from target samples. This clustering module is useful for any text related tasks by allowing each data sample to find its nearest topic cluster, thus helping the neural network model analyze the entire data. We evaluate our models on the Ubuntu Dialogue Corpus and consumer electronic domain question answering dataset, which is related to Samsung products. The proposed model shows state-of-the-art results for ranking question-answer pairs.
Related benchmarks
| Task | Dataset | Result | Rank | |
|---|---|---|---|---|
| Multi-turn Response Selection | Ubuntu Dialogue Corpus V1 (test) | R10@168.4 | 102 | |
| Response Selection | Ubuntu v2 (test) | -- | 20 | |
| Response Ranking | Ubuntu Dialog Corpus v1 (test) | Recall@1 (1/2)91.6 | 16 | |
| Multi-turn Response Selection | Ubuntu Dialogue Corpus V1 | Recall@1 (Pool 10)68.4 | 14 | |
| Multi-turn Response Selection | Ubuntu Dialogue Corpus V2 | Recall@1 (R2 Variant)91.5 | 13 | |
| Answer Ranking | Ubuntu v2 (test) | Recall@1 (1/2 Pool)91.5 | 11 | |
| Answer Ranking | Samsung QA (test) | Recall@1 (N=2)98.3 | 5 |