Share your thoughts, 1 month free Claude Pro on usSee more
WorkDL logo mark

Recurrent Neural Network for Text Classification with Multi-Task Learning

About

Neural network based methods have obtained great progress on a variety of natural language processing tasks. However, in most previous works, the models are learned based on single-task supervised objectives, which often suffer from insufficient training data. In this paper, we use the multi-task learning framework to jointly learn across multiple related tasks. Based on recurrent neural network, we propose three different mechanisms of sharing information to model text with task-specific and shared layers. The entire network is trained jointly on all these tasks. Experiments on four benchmark text classification tasks show that our proposed models can improve the performance of a task with the help of other related tasks.

Pengfei Liu, Xipeng Qiu, Xuanjing Huang• 2016

Related benchmarks

TaskDatasetResultRank
Subjectivity ClassificationSubj
Accuracy94.1
329
Question ClassificationTREC
Accuracy87.19
259
Text ClassificationMR (test)
Accuracy77.33
148
Text ClassificationSST-2
Accuracy87.9
125
Text ClassificationIMDB
Accuracy91.3
112
Text ClassificationMR
Accuracy77.33
106
Text ClassificationR8
Accuracy96.09
71
Text ClassificationR52
Accuracy90.48
56
Text ClassificationR8 (test)
Accuracy96.3
56
Document ClassificationOhsumed (test)
Accuracy51.1
54
Showing 10 of 25 rows

Other info

Follow for update