论文标题
孟加拉抽象新闻摘要(BANS):一种神经关注方法
Bengali Abstractive News Summarization(BANS): A Neural Attention Approach
论文作者
论文摘要
抽象性摘要是基于从原始文本文档中提取的信息生成新句子的过程,同时保留上下文。由于抽象摘要的基本复杂性,过去的大多数研究工作都是在提取性摘要方法上完成的。然而,随着序列到序列(SEQ2SEQ)模型的胜利,抽象摘要变得更加可行。尽管基于抽象性摘要在英语中进行了大量著名的研究,但仅对孟加拉抽象新闻摘要(BANS)进行了几项工作。在本文中,我们介绍了一个基于SEQ2SEQ的长短期内存(LSTM)网络模型,并在Encoder-Decoder上进行了关注。我们提出的系统部署了一个基于本地注意力的模型,该模型与原始文档的明显信息一起产生了一系列长顺序的单词,并带有人类般的句子。我们还准备了一个超过19k文章的数据集,并从Bangla.bdnews24.com1收集的相应的人工写的摘要是到目前为止,这是孟加拉新闻文档摘要最广泛的数据集并在Kaggle2上公开发布。我们在定性和定量上评估了模型,并将其与其他已发表的结果进行了比较。它在人类评估评分方面具有明显的改善,并采用最先进的禁令方法。
Abstractive summarization is the process of generating novel sentences based on the information extracted from the original text document while retaining the context. Due to abstractive summarization's underlying complexities, most of the past research work has been done on the extractive summarization approach. Nevertheless, with the triumph of the sequence-to-sequence (seq2seq) model, abstractive summarization becomes more viable. Although a significant number of notable research has been done in the English language based on abstractive summarization, only a couple of works have been done on Bengali abstractive news summarization (BANS). In this article, we presented a seq2seq based Long Short-Term Memory (LSTM) network model with attention at encoder-decoder. Our proposed system deploys a local attention-based model that produces a long sequence of words with lucid and human-like generated sentences with noteworthy information of the original document. We also prepared a dataset of more than 19k articles and corresponding human-written summaries collected from bangla.bdnews24.com1 which is till now the most extensive dataset for Bengali news document summarization and publicly published in Kaggle2. We evaluated our model qualitatively and quantitatively and compared it with other published results. It showed significant improvement in terms of human evaluation scores with state-of-the-art approaches for BANS.