首页 | 官方网站   微博 | 高级检索  
     

融入注意力机制的越南语组块识别方法
引用本文:王闻慧,毕玉德,雷树杰.融入注意力机制的越南语组块识别方法[J].中文信息学报,2019,33(12):91-100.
作者姓名:王闻慧  毕玉德  雷树杰
作者单位:1.信息工程大学 洛阳校区,河南 洛阳 471003;
2.复旦大学 外国语言文学学院,上海 200433
摘    要:对于越南语组块识别任务,在前期对越南语组块内部词性构成模式进行统计调查的基础上,该文针对Bi-LSTM+CRF模型提出了两种融入注意力机制的方法: 一是在输入层融入注意力机制,从而使得模型能够灵活调整输入的词向量与词性特征向量各自的权重;二是在Bi-LSTM之上加入了多头注意力机制,从而使模型能够学习到Bi-LSTM输出值的权重矩阵,进而有选择地聚焦于重要信息。实验结果表明,在输入层融入注意力机制后,模型对组块识别的F值提升了3.08%,在Bi-LSTM之上加入了多头注意力机制之后,模型对组块识别的F值提升了4.56%,证明了这两种方法的有效性。

关 键 词:越南语  组块识别  Bi-LSTM+CRF模型  注意力机制  

Vietnamese Chunk Identification Incorporating Attention Mechanism
WANG Wenhui,BI Yude,LEI Shujie.Vietnamese Chunk Identification Incorporating Attention Mechanism[J].Journal of Chinese Information Processing,2019,33(12):91-100.
Authors:WANG Wenhui  BI Yude  LEI Shujie
Affiliation:1.Luoyang Division, Information Engineering University, Luoyang, Henan 471003, China;
2.College of Foreign Language and Literature, Fudan University, Shanghai 200433, China
Abstract:For the Vietnamese chunk identification task, this paper proposes two ways to integrate the attention mechanism with the Bi-LSTM+CRF model. The first is to integrate the attention mechanism at the input layer, which allows the model to flexibly adjust weights of word embeddings and POS feature embeddings. The second is to add a multi-head attention mechanism on the top of Bi-LSTM, which enables the model to learn weight matrix of the Bi-LSTM outputs and selectively focus on important information. Experimental results show that, after integrating the attention mechanism at the input layer, the F-value of Vietnamese chunk identification is increased by 3.08%; and after adding the multi-head attention mechanism on the top of Bi-LSTM, the F-value of Vietnamese chunk identification is improved by 4.56%.
Keywords:Vietnamese  chunk identification  Bi-LSTM+CRF model  attention mechanism  
点击此处可从《中文信息学报》浏览原始摘要信息
点击此处可从《中文信息学报》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号