Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Attentive Pooling Networks

About

In this work, we propose Attentive Pooling (AP), a two-way attention mechanism for discriminative model training. In the context of pair-wise ranking or classification with neural networks, AP enables the pooling layer to be aware of the current input pair, in a way that information from the two input items can directly influence the computation of each other's representations. Along with such representations of the paired inputs, AP jointly learns a similarity measure over projected segments (e.g. trigrams) of the pair, and subsequently, derives the corresponding attention vector for each input to guide the pooling. Our two-way attention mechanism is a general framework independent of the underlying representation learning, and it has been applied to both convolutional neural networks (CNNs) and recurrent neural networks (RNNs) in our studies. The empirical results, from three very different benchmark tasks of question answering/answer selection, demonstrate that our proposed models outperform a variety of strong baselines and achieve state-of-the-art performance in all the benchmarks.

Cicero dos Santos, Ming Tan, Bing Xiang, Bowen Zhou• 2016

Related benchmarks

TaskDatasetResultRank
Answer SelectionWikiQA (test)
MAP0.689
149
QA Answer SelectionTREC Answer Selection (test)
MAP0.753
33
Question AnsweringYahooCQA (test)
P@156.8
27
Question AnsweringTrecQA clean (test)
MRR85.1
24
Answer Sentence SelectionTREC-QA clean-version
MAP75.3
14
Sentence RankingWikiQA
MAP68.86
13
Community Question AnsweringSemEvalCQA (test)
MAP77.1
13
Answer SelectionInsuranceQA (dev)
Accuracy68.8
6
Answer SelectionInsuranceQA (test1)
Accuracy71.7
6
Answer SelectionInsuranceQA (test2)
Accuracy66.4
6
Showing 10 of 11 rows

Other info

Code

Follow for update