期刊论文详细信息
Computer Science and Information Systems
Variational neural decoder for abstractive text summarization
article
Huan Zhao1  Jie Cao1  Mingquan Xu1  Jian Lu1 
[1] College of Computer Science and Electronic Engineering, Hunan University
关键词: abstractive summarization;    sequence-to-sequence;    variational auto-encoder;    variation neural inferer;   
DOI  :  10.2298/CSIS200131012Z
学科分类:土木及结构工程学
来源: Computer Science and Information Systems
PDF
【 摘 要 】

In the conventional sequence-to-sequence (seq2seq) model for abstractive summarization, the internal transformation structure of recurrent neural networks (RNNs) is completely determined. Therefore, the learned semantic information is far from enough to represent all semantic details and context dependencies, resulting in a redundant summary and poor consistency. In this paper, we propose a variational neural decoder text summarization model (VND). The model introduces a series of implicit variables by combining variational RNN and variational autoencoder, which is used to capture complex semantic representation at each step of decoding. It includes a standard RNN layer and a variational RNN layer [5]. These two network layers respectively generate a deterministic hidden state and a random hidden state. We use these two RNN layers to establish the dependence between implicit variables between adjacent time steps. In this way, the model structure can better capture the complex semantics and the strong dependence between the adjacent time steps when outputting the summary, thereby improving the performance of generating the summary. The experimental results show that, on the text summary LCSTS and English Gigaword dataset, our model has a significant improvement over the baseline model.

【 授权许可】

CC BY-NC-ND   

【 预 览 】
附件列表
Files Size Format View
RO202307150003218ZK.pdf 415KB PDF download
  文献评价指标  
  下载次数:1次 浏览次数:0次