Abstract | ||
---|---|---|
Recently proposed machine comprehension (MC) application is an effort to deal with natural language understanding problem. However, the small size of machine comprehension labeled data confines the application of deep neural networks architectures that have shown advantage in semantic inference tasks. Previous methods use a lot of NLP tools to extract linguistic features but only gain little improvement over simple baseline. In this paper, we build an attention-based recurrent neural network model, train it with the help of external knowledge which is semantically relevant to machine comprehension, and achieves a new state-of-the-art result. |
Year | Venue | Field |
---|---|---|
2016 | IJCAI | Inference,Computer science,Recurrent neural network,Natural language understanding,Natural language processing,Artificial intelligence,Labeled data,Comprehension,Deep neural networks,Machine learning |
DocType | Citations | PageRank |
Conference | 6 | 0.42 |
References | Authors | |
8 | 5 |
Name | Order | Citations | PageRank |
---|---|---|---|
Bingning Wang | 1 | 49 | 4.26 |
Shangmin Guo | 2 | 9 | 1.47 |
Kang Liu | 3 | 1542 | 89.33 |
Shizhu He | 4 | 409 | 24.02 |
Jun Zhao | 5 | 2119 | 115.52 |