首页 | 官方网站   微博 | 高级检索  
     

基于优势演员-评论家算法的强化自动摘要模型
引用本文:杜嘻嘻,程华,房一泉. 基于优势演员-评论家算法的强化自动摘要模型[J]. 计算机应用, 2021, 41(3): 699-705. DOI: 10.11772/j.issn.1001-9081.2020060837
作者姓名:杜嘻嘻  程华  房一泉
作者单位:华东理工大学 信息科学与工程学院, 上海 200237
基金项目:赛尔网络下一代互联网技术创新项目
摘    要:针对长文本自动摘要任务中抽取式模型摘要较为冗余,而生成式摘要模型时常有关键信息丢失、摘要不准确和生成内容重复等问题,提出一种面向长文本的基于优势演员-评论家算法的强化自动摘要模型(A2C-RLAS).首先,用基于卷积神经网络(CNN)和循环神经网络(RNN)的混合神经网络的抽取器(extractor)来提取原文关键句;...

关 键 词:自动摘要模型  抽取式摘要模型  生成式摘要模型  编码器-解码器  强化学习  优势演员-评论家算法
收稿时间:2020-06-17
修稿时间:2020-10-08

Reinforced automatic summarization model based on advantage actor-critic algorithm
DU Xixi,CHENG Hua,FANG Yiquan. Reinforced automatic summarization model based on advantage actor-critic algorithm[J]. Journal of Computer Applications, 2021, 41(3): 699-705. DOI: 10.11772/j.issn.1001-9081.2020060837
Authors:DU Xixi  CHENG Hua  FANG Yiquan
Affiliation:Institute of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China
Abstract:The extractive summary model is relatively redundant and the abstractive summary model often loses key information and has inaccurate summary and repeated generated content in long text automatic summarization task. In order to solve these problems, a Reinforced Automatic Summarization model based on Advantage Actor-Critic algorithm (A2C-RLAS) for long text was proposed. Firstly, the key sentences of the original text were extracted by the extractor based on the hybrid neural network of Convolutional Neural Network (CNN) and Recurrent Neural Network (RNN). Then, the key sentences were refined by the rewriter based on the copy mechanism and the attention mechanism. Finally, the Advantage Actor-Critic (A2C) algorithm in reinforcement learning was used to train the entire network, and the semantic similarity between the rewritten summary and the reference summary (BERTScore (Evaluating Text Generation with Bidirectional Encoder Representations from Transformers) value) was used as a reward to guide the extraction process, so as to improve the quality of sentences extracted by the extractor. The experimental results on CNN/Daily Mail dataset show that, compared with models such as Reinforcement Learning-based Extractive Summarization (Refresh) model, a Recurrent Neural Network based sequence model for extractive summarization (SummaRuNNer) and Distributional Semantics Reward (DSR) model, the A2C-RLAS has the final summary with content more accurate, language more fluent and redundant content effectively reduced, at the same time, A2C-RLAS has both the ROUGE (Recall-Oriented Understudy for Gisting Evaluation) and BERTScore indicators improved. Compared to the Refresh model and the SummaRuNNer model, the ROUGE-L value of the A2C-RLAS model is increased by 6.3% and 10.2% respectively; compared with the DSR model, the F1 value of the A2C-RLAS model is increased by 30.5%.
Keywords:automatic summary model  extractive summary model  abstractive summary model  encoder-decoder  reinforcement learning  Advantage Actor-Critic (A2C) algorithm  
本文献已被 万方数据 等数据库收录!
点击此处可从《计算机应用》浏览原始摘要信息
点击此处可从《计算机应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号