首页 | 官方网站   微博 | 高级检索  
     

基于BERT的三阶段式问答模型
引用本文:彭宇,李晓瑜,胡世杰,刘晓磊,钱伟中.基于BERT的三阶段式问答模型[J].计算机应用,2022,42(1):64-70.
作者姓名:彭宇  李晓瑜  胡世杰  刘晓磊  钱伟中
作者单位:电子科技大学 信息与软件工程学院,成都 610054
基金项目:四川省科技计划项目(重点研发项目)(19ZDYF0794)。
摘    要:预训练语言模型的发展极大地推动了机器阅读理解任务的进步.为了充分利用预训练语言模型中的浅层特征,并进一步提升问答模型预测答案的准确性,提出了一种基于BERT的三阶段式问答模型.首先,基于BERT设计了预回答、再回答及答案调整三个阶段;然后,在预回答阶段将BERT嵌入层的输入视作浅层特征来进行答案预生成;接着,在再回答阶...

关 键 词:自然语言处理  机器阅读理解  抽取式问答  BERT  深度学习
收稿时间:2021-03-08
修稿时间:2021-05-12

Three-stage question answering model based on BERT
PENG Yu,LI Xiaoyu,HU Shijie,LIU Xiaolei,QIAN Weizhong.Three-stage question answering model based on BERT[J].journal of Computer Applications,2022,42(1):64-70.
Authors:PENG Yu  LI Xiaoyu  HU Shijie  LIU Xiaolei  QIAN Weizhong
Affiliation:School of Information and Software Engineering,University of Electronic Science and Technology of China,Chengdu Sichuan 610054,China
Abstract:The development of pre-trained language models has greatly promoted the progress of machine reading comprehension tasks. In order to make full use of shallow features of the pre-trained language model and further improve the accuracy of predictive answer of question answering model, a three-stage question answering model based on Bidirectional Encoder Representation from Transformers (BERT) was proposed. Firstly, the three stages of pre-answering, re-answering and answer-adjusting were designed based on BERT. Secondly, the inputs of embedding layer of BERT were treated as shallow features to pre-generate an answer in pre-answering stage. Then, the deep features fully encoded by BERT were used to re-generate another answer in re-answering stage. Finally, the final prediction result was generated by combining the previous two answers in answer-adjusting stage. Experimental results on English dataset Stanford Question Answering Dataset 2.0 (SQuAD2.0) and Chinese dataset Chinese Machine Reading Comprehension 2018 (CMRC2018) of span-extraction question answering task show that the Exact Match (EM) and F1 score (F1) of the proposed model are improved by the average of 1 to 3 percentage points compared with those of the similar baseline models, and the model has the extracted answer fragments more accurate. By combining shallow features of BERT with deep features, this three-stage model extends the abstract representation ability of BERT, and explores the application of shallow features of BERT in question answering models, and has the characteristics of simple structure, accurate prediction, and fast speed of training and inference.
Keywords:Natural Language Processing(NLP)  machine reading comprehension  span-extraction question answering  Bidirectional Encoder Representation from Transformers(BERT)  deep learning
本文献已被 维普 万方数据 等数据库收录!
点击此处可从《计算机应用》浏览原始摘要信息
点击此处可从《计算机应用》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号