Variational Neural Decoder for Abstractive Text Summarization

Huan Zhao1, Jie Cao1, Mingquan Xu1 and Jian Lu1

  1. College of Computer Science and Electronic Engineering, Hunan University
    Changsha, Hunan, P.R. China, 410000
    {hzhao,jiecao,hb xmq,jianlu}@hnu.edu.cn

Abstract

In the conventional sequence-to-sequence (seq2seq) model for abstractive summarization, the internal transformation structure of recurrent neural networks (RNNs) is completely determined. Therefore, the learned semantic information is far from enough to represent all semantic details and context dependencies, resulting in a redundant summary and poor consistency. In this paper, we propose a variational neural decoder text summarization model (VND). The model introduces a series of implicit variables by combining variational RNN and variational autoencoder, which is used to capture complex semantic representation at each step of decoding. It includes a standard RNN layer and a variational RNN layer [5]. These two network layers respectively generate a deterministic hidden state and a random hidden state. We use these two RNN layers to establish the dependence between implicit variables between adjacent time steps. In this way, the model structure can better capture the complex semantics and the strong dependence between the adjacent time steps when outputting the summary, thereby improving the performance of generating the summary. The experimental results show that, on the text summary LCSTS and English Gigaword dataset, our model has a significant improvement over the baseline model.

Key words

abstractive summarization, sequence-to-sequence, variational auto-encoder, variation neural inferer

Digital Object Identifier (DOI)

https://doi.org/10.2298/CSIS200131012Z

Publication information

Volume 17, Issue 2 (June 2020)
Year of Publication: 2020
ISSN: 2406-1018 (Online)
Publisher: ComSIS Consortium

Full text

DownloadAvailable in PDF
Portable Document Format

How to cite

Zhao, H., Cao, J., Xu, M., Lu, J.: Variational Neural Decoder for Abstractive Text Summarization. Computer Science and Information Systems, Vol. 17, No. 2, 537–552. (2020), https://doi.org/10.2298/CSIS200131012Z