Multi-Document Summarization Using Selective Attention Span and Reinforcement Learning.
IEEE ACM Trans. Audio Speech Lang. Process.(2023)
摘要
Abstractive text summarization systems using recently improved RNN-based sequence-to-sequence architecture have shown great promise for single-document summarization. However, such neural models fail to perpetuate the performance in the multi-document summarization setting owing to the long-range dependencies within the documents, overlapping/contradicting facts and extrinsic model hallucinations. These shortcomings augment the model to generate inconsistent, repetitive and non-factual summaries. In this work, we introduce
REISA
, a sequence-to-sequence model with a novel
reinforced selective attention span
that attends over the input and recalibrates the local attention weights to focus on important segments while generating output at each time step.
REISA
utilizes a reinforcement learning-based policy gradient algorithm to reward the model and formulate attention distributions over the encoder input. We further benchmark
REISA
on two widely-used multi-document summarization corpora – Multinews and CQASumm, and observe an improvement of
$+2.91$
and
$+6.64$
ROUGE-L scores, respectively. The qualitative analyses on semantic similarity by BERTScore, faithfulness by question-answer evaluation and human evaluation show significant improvement over the baseline-generated summaries.
更多查看译文
关键词
selective attention span,selective attention,multi-document
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要