Computer Science and Information Systems | |
Variational neural decoder for abstractive text summarization | |
article | |
Huan Zhao1  Jie Cao1  Mingquan Xu1  Jian Lu1  | |
[1] College of Computer Science and Electronic Engineering, Hunan University | |
关键词: abstractive summarization; sequence-to-sequence; variational auto-encoder; variation neural inferer; | |
DOI : 10.2298/CSIS200131012Z | |
学科分类:土木及结构工程学 | |
来源: Computer Science and Information Systems | |
【 摘 要 】
In the conventional sequence-to-sequence (seq2seq) model for abstractive summarization, the internal transformation structure of recurrent neural networks (RNNs) is completely determined. Therefore, the learned semantic information is far from enough to represent all semantic details and context dependencies, resulting in a redundant summary and poor consistency. In this paper, we propose a variational neural decoder text summarization model (VND). The model introduces a series of implicit variables by combining variational RNN and variational autoencoder, which is used to capture complex semantic representation at each step of decoding. It includes a standard RNN layer and a variational RNN layer [5]. These two network layers respectively generate a deterministic hidden state and a random hidden state. We use these two RNN layers to establish the dependence between implicit variables between adjacent time steps. In this way, the model structure can better capture the complex semantics and the strong dependence between the adjacent time steps when outputting the summary, thereby improving the performance of generating the summary. The experimental results show that, on the text summary LCSTS and English Gigaword dataset, our model has a significant improvement over the baseline model.
【 授权许可】
CC BY-NC-ND
【 预 览 】
Files | Size | Format | View |
---|---|---|---|
RO202307150003218ZK.pdf | 415KB | download |