首页 | 官方网站   微博 | 高级检索  
     

听觉注意模型的语谱图语音情感识别方法
引用本文:张昕然,查诚,宋鹏,陶华伟,赵力.听觉注意模型的语谱图语音情感识别方法[J].信号处理,2016,32(9):1117-1125.
作者姓名:张昕然  查诚  宋鹏  陶华伟  赵力
作者单位:东南大学水声信号处理教育部重点实验室
基金项目:国家自然科学基金项目(61273266,61375028);教育部博士点专项基金(20110092130004);山东省自然科学基金(ZR2014FQ016)
摘    要:在语音情感识别技术中,由于噪声环境、说话方式和说话人特质原因,会造成实验数据库特征不匹配的情况。从语音学上分析,该问题多存在于跨数据库情感识别实验。训练的声学模型和用于测试的语句样本之间的错位,会使语音情感识别性能剧烈下降。本文据此所研究的选择性注意声学模型能有效探测变化的情感特征。同时,利用时频原子对模型进行改进,使之能提取跨语音数据库中的显著性特征用于情感识别。实验结果表明,利用文章所提方法在跨库情感样本上进行特征提取,再通过典型的分类器,识别性能提高了9个百分点,从而验证了该方法对不同数据库具有更好的鲁棒性。 

关 键 词:语音情感识别    跨数据库    语谱图特征    听觉注意机制    时频原子
收稿时间:2015-12-23

Spectrogram Speech Emotion Recognition Method Based on Auditory Attention Model
Affiliation:Key Laboratory of Underwater Acoustic Signal Processing of Ministry of Education, Southeast University
Abstract:When there exists mismatch between the trained acoustic models and the test utterances due to noise conditions, speaking styles and speaker traits, unmatched features may appear in cross-corpus. The resulting is the drastic degression in the performance of speech emotion recognition. Hence, the auditory attention model is found to be very effective for variational emotion features detection in our work. Therefore, Chirplet has been adopted to obtain salient gist features which show their relation to the expected performance in cross-corpus testing. In our experimental results, the prototypical classifier with the proposed feature extraction approach can deliver a gain of up to 9.6% accuracy in cross-corpus speech recognition, which is observed insensitive to different databases. 
Keywords:
点击此处可从《信号处理》浏览原始摘要信息
点击此处可从《信号处理》下载全文
设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司    京ICP备09084417号-23

京公网安备 11010802026262号