Speech Emotion Recognition for Indonesian Language Using Long Short-Term Memory

Author(s):  
Jeremia Jason Lasiman ◽  
Dessi Puji Lestari
2019 ◽  
Vol 2019 ◽  
pp. 1-9 ◽  
Author(s):  
Linqin Cai ◽  
Yaxin Hu ◽  
Jiangong Dong ◽  
Sitong Zhou

With the rapid development in social media, single-modal emotion recognition is hard to satisfy the demands of the current emotional recognition system. Aiming to optimize the performance of the emotional recognition system, a multimodal emotion recognition model from speech and text was proposed in this paper. Considering the complementarity between different modes, CNN (convolutional neural network) and LSTM (long short-term memory) were combined in a form of binary channels to learn acoustic emotion features; meanwhile, an effective Bi-LSTM (bidirectional long short-term memory) network was resorted to capture the textual features. Furthermore, we applied a deep neural network to learn and classify the fusion features. The final emotional state was determined by the output of both speech and text emotion analysis. Finally, the multimodal fusion experiments were carried out to validate the proposed model on the IEMOCAP database. In comparison with the single modal, the overall recognition accuracy of text increased 6.70%, and that of speech emotion recognition soared 13.85%. Experimental results show that the recognition accuracy of our multimodal is higher than that of the single modal and outperforms other published multimodal models on the test datasets.


IEEE Access ◽  
2020 ◽  
Vol 8 ◽  
pp. 124928-124938 ◽  
Author(s):  
Simin Wang ◽  
Junhuai Li ◽  
Ting Cao ◽  
Huaijun Wang ◽  
Pengjia Tu ◽  
...  

IEEE Access ◽  
2018 ◽  
Vol 6 ◽  
pp. 49325-49338 ◽  
Author(s):  
Bahareh Nakisa ◽  
Mohammad Naim Rastgoo ◽  
Andry Rakotonirainy ◽  
Frederic Maire ◽  
Vinod Chandran

2021 ◽  
pp. 103153
Author(s):  
Tian Chen ◽  
Hongfang Yin ◽  
Xiaohui Yuan ◽  
Yu Gu ◽  
Fuji Ren ◽  
...  

Electronics ◽  
2020 ◽  
Vol 9 (5) ◽  
pp. 713 ◽  
Author(s):  
Yeonguk Yu ◽  
Yoon-Joong Kim

We propose a speech-emotion recognition (SER) model with an “attention-long Long Short-Term Memory (LSTM)-attention” component to combine IS09, a commonly used feature for SER, and mel spectrogram, and we analyze the reliability problem of the interactive emotional dyadic motion capture (IEMOCAP) database. The attention mechanism of the model focuses on emotion-related elements of the IS09 and mel spectrogram feature and the emotion-related duration from the time of the feature. Thus, the model extracts emotion information from a given speech signal. The proposed model for the baseline study achieved a weighted accuracy (WA) of 68% for the improvised dataset of IEMOCAP. However, the WA of the proposed model of the main study and modified models could not achieve more than 68% in the improvised dataset. This is because of the reliability limit of the IEMOCAP dataset. A more reliable dataset is required for a more accurate evaluation of the model’s performance. Therefore, in this study, we reconstructed a more reliable dataset based on the labeling results provided by IEMOCAP. The experimental results of the model for the more reliable dataset confirmed a WA of 73%.


Sign in / Sign up

Export Citation Format

Share Document