Our new X account is live! Follow @wizwand_team for updates
WorkDL logo mark

Embracing data abundance: BookTest Dataset for Reading Comprehension

About

There is a practically unlimited amount of natural language data available. Still, recent work in text comprehension has focused on datasets which are small relative to current computing possibilities. This article is making a case for the community to move to larger data and as a step in that direction it is proposing the BookTest, a new dataset similar to the popular Children's Book Test (CBT), however more than 60 times larger. We show that training on the new data improves the accuracy of our Attention-Sum Reader model on the original CBT test data by a much larger margin than many recent attempts to improve the model architecture. On one version of the dataset our ensemble even exceeds the human baseline provided by Facebook. We then show in our own human study that there is still space for further improvement.

Ondrej Bajgar, Rudolf Kadlec, Jan Kleindienst• 2016

Related benchmarks

TaskDatasetResultRank
Machine ComprehensionCNN (val)
Accuracy0.739
80
Machine ComprehensionCNN (test)
Accuracy75.4
77
Machine ComprehensionCBT-CN (test)
Accuracy83.7
56
Machine ComprehensionCBT NE (test)
Accuracy78.4
56
Machine Reading ComprehensionDaily Mail (test)
Accuracy77.7
46
Machine ComprehensionCBT-NE (val)
Accuracy82.3
37
Machine ComprehensionCBT-CN (val)
Accuracy85.7
37
Machine Reading ComprehensionDaily Mail (val)
Accuracy78.7
36
Cloze-style Question AnsweringWDW Strict 1.0 (test)
Accuracy57
10
Cloze-style Question AnsweringWDW Relaxed 1.0 (test)
Accuracy59
9
Showing 10 of 10 rows

Other info

Follow for update