The human voice can be characterized by several attributes such as pitch, timbre, loudness, and vocal tone. It has often been observed that humans express their emotions by varying different vocal attributes during speech generation. This paper presents an algorithmic approach for detection of human emotions with the help speech .The prime objective of this paper isto recognize emotionsin speech and classify them in 6 emotion output classes namely angry, fear, disgust, happy, sad and neutral. The proposed approach is based upon the Mel Frequency Cepstral coefficients (MFCC) uses Crema-D database of emotional speech. Data Augmention is perfomed on input data audio file,such as Noise, High Speed, Low Speed etc. are added, thus more the varied data is available to the model better the model understands. Feature extraction is done using MFCC and then the extracted features are Normalized(for Independent Variable), Label Encoding(for Dependent Variable(for SVM,RF)),One Hot Encoding(for Dependent Variable(for CNN))is done. After thisthe datasetis divided into Train, Test and given to different models such as Convolutional Neural Network(CNN),Support Vector Machine(SVM), Random Forest(RF) for Emotion prediction. We report accuracy, f-score, precision and recall for the different experiment settings we evaluated our models in. Convolutional Neural Network(CNN) was found to have the highest accuracy and predicted correct emotion 88.21%ofthe time. Hence, deduction of human emotionsthrough speech analysis has a practical plausibility and could potentially be beneficial for improving human conversational and persuasion skills.
Citations
APA: Mrs. T. Sunitha, K. Poojitha, S. Venkata Rakesh, G. Bhanu Swetha, B. Kasi Priyanka (2025). Enhancing the Technique of Speech Emotion recognition using Feature Learning. DOI: 10.86493/VEREDAS.2024/V14I4/03
AMA: Mrs. T. Sunitha, K. Poojitha, S. Venkata Rakesh, G. Bhanu Swetha, B. Kasi Priyanka. Enhancing the Technique of Speech Emotion recognition using Feature Learning. 2025. DOI: 10.86493/VEREDAS.2024/V14I4/03
Chicago: Mrs. T. Sunitha, K. Poojitha, S. Venkata Rakesh, G. Bhanu Swetha, B. Kasi Priyanka. "Enhancing the Technique of Speech Emotion recognition using Feature Learning." Published 2025. DOI: 10.86493/VEREDAS.2024/V14I4/03
IEEE: Mrs. T. Sunitha, K. Poojitha, S. Venkata Rakesh, G. Bhanu Swetha, B. Kasi Priyanka, "Enhancing the Technique of Speech Emotion recognition using Feature Learning," 2025, DOI: 10.86493/VEREDAS.2024/V14I4/03
ISNAD: Mrs. T. Sunitha, K. Poojitha, S. Venkata Rakesh, G. Bhanu Swetha, B. Kasi Priyanka. "Enhancing the Technique of Speech Emotion recognition using Feature Learning." DOI: 10.86493/VEREDAS.2024/V14I4/03
MLA: Mrs. T. Sunitha, K. Poojitha, S. Venkata Rakesh, G. Bhanu Swetha, B. Kasi Priyanka. "Enhancing the Technique of Speech Emotion recognition using Feature Learning." 2025, DOI: 10.86493/VEREDAS.2024/V14I4/03