A Database for Automatic Persian Speech Emotion Recognition: Collection, Processing and Evaluation

Authors

  • Ali Harimi Electrical Engineering Department, Semnan University
  • Zeynab Esmaileyan Department of Electrical engineering, Shahrood Sci, Shahrood Science and Research branch, Islamic Azad Univercity
Abstract:

Abstract   Recent developments in robotics automation have motivated researchers to improve the efficiency of interactive systems by making a natural man-machine interaction. Since speech is the most popular method of communication, recognizing human emotions from speech signal becomes a challenging research topic known as Speech Emotion Recognition (SER). In this study, we propose a Persian emotional speech corpus collected from emotional sentences of drama radio programs. Moreover, we proposed a new automatic speech emotion recognition system which is used both spectral and prosodic feature simultaneously. We compared the proposed database with the public and widely used Berlin database. The proposed SER system  is developed for females and  males separately. Then, irrelevant features are removed using Fisher Discriminant Ratio (FDR) filtering feature selection technique. The selected features are further reduced in dimensions using Linear Discriminant Analysis (LDA) embedding feature reduction scheme. Finally, the samples are classified by a LDA classifier. The overall recognition rate of 55.74% and 47.28% is achieved on proposed database for females and males, respectively. Also, the average recognition rate of 78.64% and 73.40% are obtained for Berlin database for females and males, respectively.

Upgrade to premium to download articles

Sign up to access the full text

Already have an account?login

similar resources

Designing and implementing a system for Automatic recognition of Persian letters by Lip-reading using image processing methods

For many years, speech has been the most natural and efficient means of information exchange for human beings. With the advancement of technology and the prevalence of computer usage, the design and production of speech recognition systems have been considered by researchers. Among this, lip-reading techniques encountered with many challenges for speech recognition, that one of the challenges b...

full text

Developing a Standardized Medical Speech Recognition Database for Reconstructive Hand Surgery

Fast and holistic access to the patients’ clinical record is a major requirement of modern medical decision support systems (DSS). While electronic health records (EHRs) have replaced the traditional paper-based records in most healthcare organization, the data entry into these systems remains largely manual. Speech recognition technology promises substitution of the more convenient speech-base...

full text

Design of Automatic Speech Emotion Recognition System

In this paper we describe a speech emotion recognition system by using k nearest neighbor classifier of statistic features of prosodic contours. We survey major approaches to emotion recognition and argue for using an algorithm dealing with a selection of statistic features of the prosodic contours with further reduction feature space by using SFFS, PCA and LDA and classification provided by k-...

full text

Improving automatic emotion recognition from speech signals

We present a speech signal driven emotion recognition system. Our system is trained and tested with the INTERSPEECH 2009 Emotion Challenge corpus, which includes spontaneous and emotionally rich recordings. The challenge includes classifier and feature sub-challenges with five-class and two-class classification problems. We investigate prosody related, spectral and HMM-based features for the ev...

full text

Real-time automatic emotion recognition from speech

Recently, the importance of reacting to the emotional state of a user has been generally accepted in the field of human-computer interaction and especially speech has received increased focus as a modality from which to automatically deduct information on emotion. So far, mainly academic and not very application-oriented offline studies based on previously recorded and annotated databases with ...

full text

Automatic Emotion Recognition by the Speech Signal

This paper dis cusses approaches to recognize the emotional user state by analyzing spoken utterances on both, the semantic and the signal level. We classify seven emotions: joy, anger, irritation, fear, disgust, sadness and neutral inner state. The introduced methods analyze the wording, the degree of verbosity, the temporal intention rate as well as the history of user utterances. As prosodic...

full text

My Resources

Save resource for easier access later

Save to my library Already added to my library

{@ msg_add @}


Journal title

volume 27  issue 1

pages  79- 90

publication date 2014-01-01

By following a journal you will be notified via email when a new issue of this journal is published.

Hosted on Doprax cloud platform doprax.com

copyright © 2015-2023