Abstract
In this paper, we propose a technique for estimating the degree or intensity of emotional expressions and speaking styles appearing in speech. The key idea is based on a style control technique for speech synthesis using a multiple regression hidden semi-Markov model (MRHSMM), and the proposed technique can be viewed as the inverse of the style control. In the proposed technique, the acoustic features of spectrum, power, fundamental frequency, and duration are simultaneously modeled using the MRHSMM. We derive an algorithm for estimating explanatory variables of the MRHSMM, each of which represents the degree or intensity of emotional expressions and speaking styles appearing in acoustic features of speech, based on a maximum likelihood criterion. We show experimental results to demonstrate the ability of the proposed technique using two types of speech data, simulated emotional speech and spontaneous speech with different speaking styles. It is found that the estimated values have correlation with human perception.
Original language | English |
---|---|
Pages (from-to) | 116-124 |
Number of pages | 9 |
Journal | IEICE Transactions on Information and Systems |
Volume | E93-D |
Issue number | 1 |
DOIs | |
Publication status | Published - 2010 |
Externally published | Yes |
Keywords
- Emotion recognition
- Emotional expression
- Hidden semi-Markov model (HSMM)
- Intensity of style
- Multiple-regression HSMM (MRHSMM)
- Speaking style
ASJC Scopus subject areas
- Software
- Hardware and Architecture
- Computer Vision and Pattern Recognition
- Electrical and Electronic Engineering
- Artificial Intelligence