TY - GEN
T1 - Construction and analysis of phonetically and prosodically balanced emotional speech database
AU - Takeishi, Emika
AU - Nose, Takashi
AU - Chiba, Yuya
AU - Ito, Akinori
N1 - Funding Information:
We appreciate Prof. Kentaro Inui for providing the Twitter data. Part of this study was supported by JSPS Grant-in-Aid for Scientific Research 15H02720.
Publisher Copyright:
© 2016 IEEE.
PY - 2017/5/3
Y1 - 2017/5/3
N2 - We designed an emotional speech database that can be used for emotion recognition as well as recognition and synthesis of speech with various emotions. The database was designed by compiling tweets acquired from Twitter and selecting emotion-dependent tweets considering phonetic and prosodic balance. We classified gathered tweets into four emotions: joy, anger, sadness and neutral, and then selected 50 sentences from sentences of each emotion based on the entropy-based algorithm. We compared the selected sentence sets with randomly selected sentence sets from aspects of phonetic and prosodic balance and sentence length, and confirmed that the sets selected by the algorithm were more balanced. Next, we recorded emotional speech based on the selected sentences. Then, we evaluated the speech from the viewpoint of emotional recognition and emotional speech recognition.
AB - We designed an emotional speech database that can be used for emotion recognition as well as recognition and synthesis of speech with various emotions. The database was designed by compiling tweets acquired from Twitter and selecting emotion-dependent tweets considering phonetic and prosodic balance. We classified gathered tweets into four emotions: joy, anger, sadness and neutral, and then selected 50 sentences from sentences of each emotion based on the entropy-based algorithm. We compared the selected sentence sets with randomly selected sentence sets from aspects of phonetic and prosodic balance and sentence length, and confirmed that the sets selected by the algorithm were more balanced. Next, we recorded emotional speech based on the selected sentences. Then, we evaluated the speech from the viewpoint of emotional recognition and emotional speech recognition.
KW - emotion recognition
KW - emotional speech database
KW - emotional speech recognition
KW - speech corpus
UR - http://www.scopus.com/inward/record.url?scp=85020215768&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85020215768&partnerID=8YFLogxK
U2 - 10.1109/ICSDA.2016.7918977
DO - 10.1109/ICSDA.2016.7918977
M3 - Conference contribution
AN - SCOPUS:85020215768
T3 - 2016 Conference of the Oriental Chapter of International Committee for Coordination and Standardization of Speech Databases and Assessment Techniques, O-COCOSDA 2016
SP - 16
EP - 21
BT - 2016 Conference of the Oriental Chapter of International Committee for Coordination and Standardization of Speech Databases and Assessment Techniques, O-COCOSDA 2016
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 19th Annual Conference of the Oriental Chapter of International Committee for Coordination and Standardization of Speech Databases and Assessment Techniques, O-COCOSDA 2016
Y2 - 26 October 2016 through 28 October 2016
ER -