Multi-task learning for abstractive text summarization with key information guide network

作者: Weiran Xu , Chenliang Li , Minghao Lee , Chi Zhang

DOI: 10.1186/S13634-020-00674-7

关键词:

摘要: Neural networks based on the attentional encoder-decoder model have good capability in abstractive text summarization. However, these models are hard to be controlled process of generation, which leads a lack key information. And some information, such as time, place, and people, is indispensable for humans understand main content. In this paper, we propose information guide network summarization multi-task learning framework. The core idea automatically extract that people need most an end-to-end way use it generation process, so get more human-compliant summary. our model, document encoded into two parts: results normal encoder encoding, includes sentences keywords. A framework introduced sophisticated model. To fuse novel multi-view attention obtain dynamic representations source addition, incorporated module summary generation. We evaluate CNN/Daily Mail dataset experimental show significant improvements.

参考文章(22)
Matthew D. Zeiler, ADADELTA: An Adaptive Learning Rate Method arXiv: Learning. ,(2012)
Ilya Sutskever, Quoc V. Le, Oriol Vinyals, Sequence to Sequence Learning with Neural Networks neural information processing systems. ,vol. 27, pp. 3104- 3112 ,(2014)
Elad Hazan, Yoram Singer, John Duchi, Adaptive Subgradient Methods for Online Learning and Stochastic Optimization Journal of Machine Learning Research. ,vol. 12, pp. 2121- 2159 ,(2011)
Ramesh Nallapati, Bowen Zhou, Bing Xiang, Sequence-to-Sequence RNNs for Text Summarization arXiv: Computation and Language. ,(2016)
Sumit Chopra, Michael Auli, Alexander M. Rush, Abstractive Sentence Summarization with Attentive Recurrent Neural Networks north american chapter of the association for computational linguistics. pp. 93- 98 ,(2016) , 10.18653/V1/N16-1012
Zhanyu Ma, Jing-Hao Xue, Arne Leijon, Zheng-Hua Tan, Zhen Yang, Jun Guo, Decorrelation of Neutral Vector Variables: Theory and Applications IEEE Transactions on Neural Networks. ,vol. 29, pp. 129- 143 ,(2018) , 10.1109/TNNLS.2016.2616445
Abigail See, Peter J. Liu, Christopher D. Manning, Get To The Point: Summarization with Pointer-Generator Networks meeting of the association for computational linguistics. ,vol. 1, pp. 1073- 1083 ,(2017) , 10.18653/V1/P17-1099
Dragomir Radev, Kshitijh Meelu, Michihiro Yasunaga, Ayush Pareek, Krishnan Srinivasan, Rui Zhang, Graph-based Neural Multi-Document Summarization arXiv: Computation and Language. ,(2017)
Jiwei Tan, Xiaojun Wan, Jianguo Xiao, Abstractive Document Summarization with a Graph-Based Attentional Neural Model meeting of the association for computational linguistics. ,vol. 1, pp. 1171- 1181 ,(2017) , 10.18653/V1/P17-1108
Chenliang Li, Weiran Xu, Si Li, Sheng Gao, Guiding Generation for Abstractive Text Summarization Based on Key Information Guide Network Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers). ,vol. 2, pp. 55- 60 ,(2018) , 10.18653/V1/N18-2009