Speech Emotion Recognition Based on Voice Fundamental Frequency
The human voice is one of the basic means of communication, thanks to which one also can easily convey the emotional state. This paper presents experiments on emotion recognition in human speech based on the fundamental frequency. AGH Emotional Speech Corpus was used. This database consists of audio...
Saved in:
| Main Authors: | , , |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
Institute of Fundamental Technological Research Polish Academy of Sciences
2019-04-01
|
| Series: | Archives of Acoustics |
| Subjects: | |
| Online Access: | https://acoustics.ippt.pan.pl/index.php/aa/article/view/2338 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
| _version_ | 1849711479989731328 |
|---|---|
| author | Teodora DIMITROVA-GREKOW Aneta KLIS Magdalena IGRAS-CYBULSKA |
| author_facet | Teodora DIMITROVA-GREKOW Aneta KLIS Magdalena IGRAS-CYBULSKA |
| author_sort | Teodora DIMITROVA-GREKOW |
| collection | DOAJ |
| description | The human voice is one of the basic means of communication, thanks to which one also can easily convey the emotional state. This paper presents experiments on emotion recognition in human speech based on the fundamental frequency. AGH Emotional Speech Corpus was used. This database consists of audio samples of seven emotions acted by 12 different speakers (6 female and 6 male). We explored phrases of all the emotions – all together and in various combinations. Fast Fourier Transformation and magnitude spectrum analysis were applied to extract the fundamental tone out of the speech audio samples. After extraction of several statistical features of the fundamental frequency, we studied if they carry information on the emotional state of the speaker applying different AI methods. Analysis of the outcome data was conducted with classifiers: K-Nearest Neighbours with local induction, Random Forest, Bagging, JRip, and Random Subspace Method from algorithms collection for data mining WEKA. The results prove that the fundamental frequency is a prospective choice for further experiments. |
| format | Article |
| id | doaj-art-36fd89dbffcf47bdbc011cb526345411 |
| institution | DOAJ |
| issn | 0137-5075 2300-262X |
| language | English |
| publishDate | 2019-04-01 |
| publisher | Institute of Fundamental Technological Research Polish Academy of Sciences |
| record_format | Article |
| series | Archives of Acoustics |
| spelling | doaj-art-36fd89dbffcf47bdbc011cb5263454112025-08-20T03:14:36ZengInstitute of Fundamental Technological Research Polish Academy of SciencesArchives of Acoustics0137-50752300-262X2019-04-0144210.24425/aoa.2019.128491Speech Emotion Recognition Based on Voice Fundamental FrequencyTeodora DIMITROVA-GREKOW0Aneta KLIS1Magdalena IGRAS-CYBULSKA2Bialystok University of TechnologyBialystok University of TechnologyAGH University of Science and TechnologyThe human voice is one of the basic means of communication, thanks to which one also can easily convey the emotional state. This paper presents experiments on emotion recognition in human speech based on the fundamental frequency. AGH Emotional Speech Corpus was used. This database consists of audio samples of seven emotions acted by 12 different speakers (6 female and 6 male). We explored phrases of all the emotions – all together and in various combinations. Fast Fourier Transformation and magnitude spectrum analysis were applied to extract the fundamental tone out of the speech audio samples. After extraction of several statistical features of the fundamental frequency, we studied if they carry information on the emotional state of the speaker applying different AI methods. Analysis of the outcome data was conducted with classifiers: K-Nearest Neighbours with local induction, Random Forest, Bagging, JRip, and Random Subspace Method from algorithms collection for data mining WEKA. The results prove that the fundamental frequency is a prospective choice for further experiments.https://acoustics.ippt.pan.pl/index.php/aa/article/view/2338emotion recognitionspeech signal analysisvoice analysisfundamental frequencyspeech corpora |
| spellingShingle | Teodora DIMITROVA-GREKOW Aneta KLIS Magdalena IGRAS-CYBULSKA Speech Emotion Recognition Based on Voice Fundamental Frequency Archives of Acoustics emotion recognition speech signal analysis voice analysis fundamental frequency speech corpora |
| title | Speech Emotion Recognition Based on Voice Fundamental Frequency |
| title_full | Speech Emotion Recognition Based on Voice Fundamental Frequency |
| title_fullStr | Speech Emotion Recognition Based on Voice Fundamental Frequency |
| title_full_unstemmed | Speech Emotion Recognition Based on Voice Fundamental Frequency |
| title_short | Speech Emotion Recognition Based on Voice Fundamental Frequency |
| title_sort | speech emotion recognition based on voice fundamental frequency |
| topic | emotion recognition speech signal analysis voice analysis fundamental frequency speech corpora |
| url | https://acoustics.ippt.pan.pl/index.php/aa/article/view/2338 |
| work_keys_str_mv | AT teodoradimitrovagrekow speechemotionrecognitionbasedonvoicefundamentalfrequency AT anetaklis speechemotionrecognitionbasedonvoicefundamentalfrequency AT magdalenaigrascybulska speechemotionrecognitionbasedonvoicefundamentalfrequency |