JSAI2020

Presentation information

Interactive Session

[3Rin4] Interactive 1

Thu. Jun 11, 2020 1:40 PM - 3:20 PM Room R01 (jsai2020online-2-33)

[3Rin4-90] Verification of the Effectiveness of Introducing Latent Topics into Text Summarization Model with Pretrained Encoders

〇Kana Ozaki1, Ichiro Kobayashi1 (1.Ochanomizu University)

Keywords:text summarization, pre-trained language model, topic model

In recent years, the necessity of text summarization that makes a summary of documents have increased because of having to deal with a numerous number of documents. Motivated by the remarkable development in deep learning, the studies of text summarization are developing dramatically. Among them, BERTSUM employing pretrained language model BERT for summarization achieved state-of-the-art results in both extractive and abstractive summarization. On the other hand, some summarization models outperform other models by adding the latent topic information of the target documents to summarize. In this paper, we propose a model expanding BERTSUM in terms of being able to use the latent topics extracted from source documents for BERT tokens, and verify the effectiveness of our model through the experiments for extractive summarization. Moreover, we analyze the results to verify whether the topic information works to make a better summary by measuring the coverage of topic distributions between system output and gold summary, and the coherence of topic in system output.

Authentication for paper PDF access

A password is required to view paper PDFs. If you are a registered participant, please log on the site from Participant Log In.
You could view the PDF with entering the PDF viewing password bellow.

Password