Advanced Search
ZHANG Yuzhe, GUO Chuanjie, JIN Shuya, et al. Speech emotion classification based on multi-feature extraction[J]. Technical Acoustics, 2025, 44(2): 261-269. DOI: 10.16300/j.cnki.1000-3630.23103001
Citation: ZHANG Yuzhe, GUO Chuanjie, JIN Shuya, et al. Speech emotion classification based on multi-feature extraction[J]. Technical Acoustics, 2025, 44(2): 261-269. DOI: 10.16300/j.cnki.1000-3630.23103001

Speech emotion classification based on multi-feature extraction

  • Emotion recognition is a type of computer simulation for human emotion perception process, which is significant in research and applications. Traditional speech recognition systems usually employ a single feature extraction method, which sometimes loses important information from speech emotion signals. Therefore, based on the improved complete ensemble empirical mode decomposition with adaptive noise (ICEEMDAN), a combined multi-feature extraction method to classify semantically independent speech emotion signals is proposed in this paper. ICEEMDAN decomposes one-dimensional speech signals into multiple intrinsic modes, and then extracts characteristics such as energy intensity, average, variance, kurtosis, skewness, center frequency, peak amplitude, permutation entropy from each decomposed mode. Finally, four emotions such as anger, happiness, sadness, and no emotion are classified. The results show that the proposed method achieves an average recognition rate of 88.52% after training with an 8∶2 model of the support vector machine (SVM). It can provide an important reference for speech emotion recognition.
  • loading

Catalog

    Turn off MathJax
    Article Contents

    /

    DownLoad:  Full-Size Img  PowerPoint
    Return
    Return