周红标. 融合语音和脉搏的多模态情感识别研究[J]. 微电子学与计算机, 2015, 32(6): 5-9. DOI: 10.19304/j.cnki.issn1000-7180.2015.06.002
引用本文: 周红标. 融合语音和脉搏的多模态情感识别研究[J]. 微电子学与计算机, 2015, 32(6): 5-9. DOI: 10.19304/j.cnki.issn1000-7180.2015.06.002
ZHOU Hong-biao. Research of Multimodal Emotion Recognition Based on Speech and Pulse Signal[J]. Microelectronics & Computer, 2015, 32(6): 5-9. DOI: 10.19304/j.cnki.issn1000-7180.2015.06.002
Citation: ZHOU Hong-biao. Research of Multimodal Emotion Recognition Based on Speech and Pulse Signal[J]. Microelectronics & Computer, 2015, 32(6): 5-9. DOI: 10.19304/j.cnki.issn1000-7180.2015.06.002

融合语音和脉搏的多模态情感识别研究

Research of Multimodal Emotion Recognition Based on Speech and Pulse Signal

  • 摘要: 针对单独利用语音或某种生理信号进行情感识别容易误判的问题,提出融合语音和脉搏的多模态情感识别方法.首先对预处理后的语音信号提取梅尔倒谱系数特征,并用隐马尔科夫构建语音情感识别模型,然后计算脉搏信号K值和小波包系数能量值,并输入到最小二乘支持向量机识别模型中进行判别,最后对两个模型的判别结果进行决策级的融合.实验结果表明:对于哀伤和平静两种情感,语音识别率较高,融合后识别率达到100%;对于高兴和愤怒两种情感,语音识别率为75%和80%,融合后提高到95%和90%.

     

    Abstract: In order to solve the problem of individual emotion recognition misjudged easily by speech or some kind of physiological signal,a multimodal emotion recognition research method of fusing speech and pulse is proposed.The features of MEL-frequency cepstral coefficient is extracted from the speech signals, which have experienced pretreaments, and a speech emotion recognition is constructed according to the hidden Markov model. The K value of pulse signals and energy value of wavelet packet coefficient are calculated, and the results inputed into least square support vector machine are judged. Then, the two results of the two models are carried out decision-level fusion. The experiment results show that as for sadness and calculates, speech recognition reates is higher than others, which can reach to 100% after fusion, as for happiness and anger, speech recognition tates are 75% and 80%, which can be increased to 95% and 90% after fusion.

     

/

返回文章
返回