Abstract | ||
---|---|---|
Neural sequence-to-sequence model has achieved great success in abstractive summarization task. However, due to the limit of input length, most of previous works can only utilize lead sentences as the input to generate the abstractive summarization, which ignores crucial information of the document. To alleviate this problem, we propose a novel approach to improve neural sentence summarization by using extractive summarization, which aims at taking full advantage of the document information as much as possible. Furthermore, we present both of streamline strategy and system combination strategy to achieve the fusion of the contents in different views, which can be easily adapted to other domains. Experimental results on CNN/Daily Mail dataset demonstrate both our proposed strategies can significantly improve the performance of neural sentence summarization. |
Year | DOI | Venue |
---|---|---|
2017 | 10.1007/978-3-319-73618-1_2 | Lecture Notes in Artificial Intelligence |
DocType | Volume | ISSN |
Conference | 10619 | 0302-9743 |
Citations | PageRank | References |
0 | 0.34 | 0 |
Authors | ||
6 |
Name | Order | Citations | PageRank |
---|---|---|---|
Junnan Zhu | 1 | 4 | 3.45 |
Long Zhou | 2 | 8 | 2.28 |
Haoran Li | 3 | 16 | 7.27 |
Jiajun Zhang | 4 | 257 | 46.34 |
Yu Zhou | 5 | 34 | 6.58 |
Chengqing Zong | 6 | 1004 | 102.38 |