24
Performance Comparison of Speaker and Emotion Recognition

Performance Comparison of Speaker and Emotion Recognition

Embed Size (px)

Citation preview

Page 1: Performance Comparison of Speaker and Emotion Recognition

Performance Comparison of Speaker and Emotion Recognition

Page 2: Performance Comparison of Speaker and Emotion Recognition

Abstract

• This research paper mainly focus on the Hidden Markov Model tool kit (HTK) which recognizes the speech , speaker and emotional speeches using the MFCC.

• HTK based technique gives the good result when compared to other techniques for the voice recognition, emotional speech recognition and also evaluates the the noisy test speech.

Page 3: Performance Comparison of Speaker and Emotion Recognition

• The speech signal mainly focuses on the information like age, gender, social status, accent and emotional state of the speaker.

• The main challenging task is to recognize the speaker , speech and the emotion from the speeches.

• If the user experiences the negative emotion the system has to adjust itself the need for the user or pass the control for the human agents for alternate convenient reply from the user.

Page 4: Performance Comparison of Speaker and Emotion Recognition

Introduction

• The emotion of the speech focuses on the use of MFCC and HTK modelling technique for recognising the speech, speaker and emotion by the speech database.

• In this research paper volvo , white and f16 noises are considered for the work to evaluate the emotion of independent and speaker noisy emotional speech recognition system.

Page 5: Performance Comparison of Speaker and Emotion Recognition

• we can get better results through the implementation of the additional pre processing techniques using the adaptive technique of the RLS filters prior to the conventional pre processing stages.

• noisy data are evaluated through the combination of the short time energy and zero crossing rate parameters for developing a technique which reduces the effects of the noise speech

Page 6: Performance Comparison of Speaker and Emotion Recognition

Features based on CEPSTRUM

• The short-time speech spectrum for voiced speech sound which has two components: 1) harmonic peaks due to the periodicity of voiced speech 2) glottal pulse shape.

• The excitation source decides the periodicity of voiced speech.

• It reflects the characteristics of speaker. The spectral envelope is shaped by formants which reflect the resonances of vocal tract.

Page 7: Performance Comparison of Speaker and Emotion Recognition

• This represents the source characteristics of speech signal and based on the known variation of the human ear's.

• critical bandwidth with frequencies, filters spaced linearly at low frequencies and logarithmically at high frequencies preferred to extract phonetically important characteristics of speech.

Page 8: Performance Comparison of Speaker and Emotion Recognition

Characteristics of Emotional speech

• The structural part of the speech contains linguistic information which reveals the characteristics of the pronunciation of the utterances based on the rules of the language.

• Paralinguistic information refers to the internal structure messages such as emotional state of the speaker.

• Speeches of the emotions such as anger, fear and happy are displaying the psychological behavior of the speaker such as high blood pressure and high heart rate.

Page 9: Performance Comparison of Speaker and Emotion Recognition

• Speech signals are converted into frames and frequency analysis is done on the frames.

• Frequencies are calculated on the basis of choosing the frequency bin which has high spectral energy.

• It is indicated that emotions such as anger, fear and happy have more number of frames with high frequency energy and the emotion sadness has very few frames with high frequency energy

Page 10: Performance Comparison of Speaker and Emotion Recognition

Frequency distribution of speech in different emotions.

Page 11: Performance Comparison of Speaker and Emotion Recognition
Page 12: Performance Comparison of Speaker and Emotion Recognition

SPEECH RECOGNITION USING HTK

• Utterances are chosen from 10 different actors and ten different texts.

• Ten emotional utterances are collected from five male and• Female speakers respectively in the age ranging from 21 to

35 years.• They are required to utter ten different utterances in Berlin

in seven different emotions such as anger, boredom, disgust, fear, happy, neutral and sad.

• Speech recognition system generally involves the realization of speech signal as the message encoded as the sequence of one or more symbols.

Page 13: Performance Comparison of Speaker and Emotion Recognition

• Then the MFCC features are extracted. For each training model corresponding to continuous speeches, training set of K utterances are used, where each utterance constitutes an observation sequence of some appropriate spectral or temporal representation.

Page 14: Performance Comparison of Speaker and Emotion Recognition
Page 15: Performance Comparison of Speaker and Emotion Recognition
Page 16: Performance Comparison of Speaker and Emotion Recognition

• HCopy is the HTK tool used for MFCC extraction.

• HCompV is a tool for computing overall mean and variance and generating proto type HMM.

• HInit is a tool used to read all the bootstrapped training data for initializing a single Hidden markov model using a segmental K-means algorithm.

Page 17: Performance Comparison of Speaker and Emotion Recognition

• HVite is a general purpose speech recognizer. It matches the speech file against a network of HMMs and output a transcription for each speech.

• The HTK recognizer actually requires a network to be defined using a HTK lattice format in which word to word transcription is done.

Page 18: Performance Comparison of Speaker and Emotion Recognition

• HResuIt is HTK performance analysis tool which reads the set of label files and compares them with the corresponding reference transcription.

Page 19: Performance Comparison of Speaker and Emotion Recognition

Noise Characteristics

• White noise signal

• Frequency distribution of white noise

Page 20: Performance Comparison of Speaker and Emotion Recognition

Emotion independent speaker recognition

Page 21: Performance Comparison of Speaker and Emotion Recognition

NOISY SPEECH RECOGNITION

• Noisy speech recognition in this paper is such that performance of the noisy speech recognition system, noises such as volvo, white and F16 are taken from "Noisex-92" database.

• The noise reduced speech closely resembling the clean test speech is obtained. Then, conventional preprocessing techniques are applied on the noise reduced speeches and features are extracted.

Page 22: Performance Comparison of Speaker and Emotion Recognition

RESULTS AND DISCUSSION

• Hear in this paper, the performance of emotion independent speech recognition is evaluated by considering the ten utterances spoken by ten actors.

• Overall accuracy for emotion independent speech recognition and speaker independent speech recognition are 91 % and 87% respectively for F16 noise added to the test speeches.

Page 23: Performance Comparison of Speaker and Emotion Recognition

CONCLUSION

• The performance of speech and speaker recognition systems is found to be good and is slightly low for emotion recognition. This is probably due to the usage of same set of speech of same set of speakers in different emotions.

• The variance of the noise reduced speech is almost same as that of the clean speech. Features extracted from the noise-reduced speeches are applied to models of the clean data and performance is measured in terms of recognition accuracy.

Page 24: Performance Comparison of Speaker and Emotion Recognition

Thank you..!