融合引导注意力的中文长文本摘要生成

Zhe Guo, Zhi Bo Zhang, Wei Jie Zhou, Yang Yu Fan, Yan Ning Zhang

科研成果: 期刊稿件文章同行评审

1 引用 (Scopus)

摘要

Current research on Chinese long text summarization based on deep learning has the following problems: (1) summarization models lack information guidance, fail to focus on keywords and sentences, leading to the problem of losing critical information under long-distance span; (2) the word lists of existing Chinese long text summarization models are often word-based and do not contain common Chinese words and punctuation, which is not conducive to extracting multi-grained semantic information. To solve the above problems, a Chinese long text summarization method with guided attention (CLSGA) is proposed in this paper. Firstly, for the long text summarization task, an extraction model is presented to extract the core words and sentences in the long text to construct the guided text, which can guide the generation model to focus on more important information in the encoding process. Secondly, the Chinese long text vocabulary is designed to changing the text structure from words statistics to phrases statistics, which is conducive to extracting richer multi-granularity features. Hierarchical location decomposition encoding is then introduced to efficiently extend location encoding of long text and accelerate network convergence. Finally, the local attention mechanism is combined with the guided attention mechanism to effectively capture the important information under the long text span and improve the accuracy of summarization. Experimental results on four public Chinese abstract datasets with different lengths, LCSTS, CNewSum, NLPCC2017 and SFZY2020, show that our proposed method has significant advantages over long text summarization and can effectively improve the value of ROUGE-1, ROUGE-2 and ROUGE-L.

投稿的翻译标题Chinese Long Text Summarization with Guided Attention
源语言繁体中文
页(从-至)3914-3930
页数17
期刊Tien Tzu Hsueh Pao/Acta Electronica Sinica
52
12
DOI
出版状态已出版 - 25 12月 2024

关键词

  • Chinese long text summarization
  • guided attention
  • hierarchical location decomposition coding
  • local attention
  • natural language processing

指纹

探究 '融合引导注意力的中文长文本摘要生成' 的科研主题。它们共同构成独一无二的指纹。

引用此