时间:2024-12-22
Disne SIVALINGAM
(Department of Computer Science, Trincomalee Campus, EUSL, Sri Lanka)
Abstract: The interaction between humans and machines has become an issue of concern in recent years. Besides facial expressions or gestures, speech has been evidenced as one of the foremost promising modalities for automatic emotion recognition. Effective computing means to support HCI (Human-Computer Interaction) at a psychological level, allowing PCs to adjust their reactions as per human requirements. Therefore, the recognition of emotion is pivotal in High-level interactions. Each Emotion has distinctive properties that form us to recognize them. The acoustic signal produced for identical expression or sentence changes is essentially a direct result of biophysical changes, (for example, the stress instigated narrowing of the larynx) set off by emotions. This connection between acoustic cues and emotions made Speech Emotion Recognition one of the moving subjects of the emotive computing area. The most motivation behind a Speech Emotion Recognition algorithm is to observe the emotional condition of a speaker from recorded Speech signals. The results from the application of k-NN and OVA-SVM for MFCC features without and with a feature selection approach are presented in this research. The MFCC features from the audio signal were initially extracted to characterize the properties of emotional speech.Secondly, nine basic statistical measures were calculated from MFCC and 117-dimensional features were consequently obtained to train the classifiers for seven different classes (Anger, Happiness, Disgust, Fear, Sadness,Disgust, Boredom and Neutral) of emotions. Next, Classification was done in four steps. First, all the 117-features are classified using both classifiers. Second, the best classifier was found and then features were scaled to [-1, 1] and classified. In the third step, the with or without feature scaling which gives better performance was derived from the results of the second step and the classification was done for each of the basic statistical measures separately. Finally, in the fourth step, the combination of statistical measures which gives better performance was derived using the forward feature selection method Experiments were carried out using k-NN with different k values and a linear OVA-based SVM classifier with different optimal values. Berlin emotional speech database for the German language was utilized for testing the planned methodology and recognition rates as high as 60% accomplished for the recognition of emotion from voice signal for the set of statistical measures(median, maximum, mean, Inter-quartile range, skewness). OVA-SVM performs better than k-NN and the use of the feature selection technique gives a high rate.
Keywords: Mel Frequency Cepstral Coefficients (MFCC), Fast Fourier Transformation (FFT), Discrete Cosine Transformation (DCT), k Nearest Neighbors (k-NN), Support Vector Machine (SVM), One-Vs-All(OVA)
The speech signal is the most natural method of communication between humans[1]. This fact has motivated to think of speech as a fast and efficient method of interaction between human and machine. Speech signal contains emotional contents because of that and the body languages human can understand the emotional states of another[8]. However, for a natural human-machine interaction the machine should understand the emotional state of the speaker for satisfactory response to human emotions[5]. This leads to a research field namely Speech Emotion Recognition, which defined as extracting the emotional state of a speaker from his or her speech[6].
Speech emotion recognition is particularly useful for applications, which require natural human-machine interactions where the response of those systems to the user depends on the detected emotion. Such applications are E-learning, Call-center System, Music Player,and Sony Artificial Intelligent Robot (AIBO).
The task of speech emotion classification is incredibly difficult for subsequent reasons. Initially, it's not clear that speech features are most powerful in identifying between emotions, the acoustic variability introduced by the existence of various sentences,speaking rates, speaking styles of the speakers adds another obstacle. As a result, these properties have an effect directly on most of the common extracted speech features like energy contours and pitch.[1].Sentence may contain several kinds of emotions at the same time,and emotions may be associated with just parts of the sentence. It is very difficult to determine the boundaries between these parts. An example of this situation is a live telecast of a program. One sentence may have more than one emotion. One sentence can speak in more than one language. Such as Tanglish (Tamil and English). No clear boundary for each complex emotion such emotions are shame, pride, satisfaction. Speaker’s cultural backgrounds and the environments.
Lalitha et al. (2015)[11]suggested a technique in view of the impact of MFCC and Cepstrum features in emotion detection. In addition, they carried out a relative analysis of Cepstrum, Mel-Frequency Cepstral Coefficients, and frequency scaled MFCC on Emotion Recognition. Creators utilized the Berlin Emo-set[12],which includes 7 fundamental emotions, the furthermore utilized Neural Network to do the classification.A classification rate of 57% with MFCC features and 85.70% with the combination of MFCC features,Cepstrum, and Frequency, scaled MFCC mentioned as the most extreme outcomes.
Angel Urbano Romeu (2016)[4]proposed a technique of Emotion Recognition in view of speech, utilizing a Naïve Bayes Classifier, His research the outcomes from the use of a Naïve Bayer Classifier over different kinds of features. He has used pitch and MFCC related features and used Berlin Dataset focused only on five emotions (Happy, Anger, Fear, Sad and Neutral). OpenSMILE toolkit used to extract features from the audio signal. Initially, he has obtained an accuracy of 70%. For classification improvement, he has used the weighting method and obtained results around 75% accurate the primary contrast between his project and also the others is the programing language used in C language and helpful to develop a final software, which may used in a automaton for various kinds of needs. He has obtained roughly 78% accuracy using proper layers and weights in the classifier.Moreover, they concluded classifying emotions with Naïve bayes provides fast probabilistic results and performs better than classifiers that are more sophisticated.
Shajini Majuran (2017)[13]proposed a Hierarchical Classification technique utilizing MFCC(Mel-Frequency Cepstral Coefficients). Statistical measures of MFCC are investigated in the classification individually. Also, the best fit model for every decision node is developed utilizing a one-versusall-based SVM classifier. Two benchmark speech sets have been evaluated by the proposed framework: Danish and Berlin provide better performance and results in emotion classification. The statistical measures’subset gives the best recognition. The general results reveal those essential feelings like anger, sadness,happiness, and neutral can classified in a hierarchical structure. According to the analysis on both DES and Berlin datasets, they discovered the classification values obtained using Berlin datasets are relatively above the values obtained from the DES dataset. Test results of classification acquired on the Berlin dataset demonstrates that emotions can be organized in order:Sad, Disgust, Anger, Fear, Happiness, Bored, and Neutral, respectively. Classification rate using OVA-SVMs for Berlin and DES datasets obtained 70.88% and 55.35% respectively. Classification rate using Hierarchical Scheme (without feature selection)for Berlin and DES datasets obtained 84.17% and 72.21% respectively. Classification rate using Hierarchical Scheme (with feature selection) for Berlin and DES datasets obtained 87.20% and 78.54% respectively.
Many researchers have employed emotion classification from audio signal with different features and classifiers. It can see that a couple of works have focused in on frequency and energy. In such a way,MFCC are considered as the more instructive coefficients. Also, various sorts of features and its statistical characteristics have been taken and assessed all through the research in Emotion Recognition. Plus,successful characteristics from statistical measures can be recognized for a superior classification rate. Hence,this paper proposes a way to deal with Speech Emotion Classification utilizing statistical measures of MFCC features and Its viable subset of features.
Emotion classification from speeches consists of three parts; first, one is Speech processing (noise removal and feature extraction). This part Removes unwanted noise components from signals like wind,rain, etc. The second part is Feature selection. Some features are unhelpful for classification if we use these features, it will affect the classification rate. Feature selection is the process of reducing the no of features.Best feature selection reduces the time and increases the accuracy. Best feature selection allows the development of simpler and faster models. The third part is Classification. In this part classifying, the test set to predict which emotion to assign to the speech.
Extracting features is the first step in any automatic speech recognition-related system. That is determining the elements of the audio signal that are good for distinguishing the linguistic content and dropping all the other stuff that carries information like background noise, emotion, and so on the main point to know regarding speech is that the shape of the vocal tract including the tongue, teeth that filters the sounds generated by a human. This shape verifies what sound comes out[3].If we are able to decide the shape correctly, it could provide us with a precise representation of the phoneme delivered.
The shape of the vocal tract appears in the envelope of the short-term power spectrum, and the Mel Frequency Cepstral Coefficients (MFCC) work has precisely portrayed this envelope[9].
The Mel Frequency Cepstral Coefficients (MFCC)are widely used features in automatic speech and speaker recognition. They were introduced by Davis and Mermelstein in the 1980s[9]and have been state-of-the-art ever since. Major steps of computation used to generate MFCC features shown in Fig.2.
Fig.1 A Generic Framework for State of the Art Approach in Emotion Classification
Fig.2 Block Diagram for Generating MFCC Features
a) Pre-emphasis
In the speech process, this signal filter is expected after the sampling to get a smoother spectral shape of the frequency of the speech signal. In other words, this is performed to reduce noise during sound recording.
The filter distortion is based on the input/output relationship in the time domain, expressed in the following equation:
From equation 1, a- is a filter constant, it is usually 0.9 < b < 1.0.
b) Framing
In this step, the sound signals have been sectioned into many overlapping so that no single signal is eliminated. This step goes on until all signals have entered at least one frames as displayed (Fig.2). Speech analysis was performed by short-term analysis. A sound signal is continually changing, and to keep things simple we should accept that the sound signal doesn't change much on short time frame scales.
Therefore, we outline the signal in 20-40 ms frames. Assuming that the frame is short, we need more samples to get a reliable estimation of the spectral,assuming it is longer the signal changes a lot all through the frame.
c) Windowing
Windowing is a process for examining long sound signals by taking an adequate delegate area. Windowing is a Finite Impulse Response (FIR) digital filter approach. This process eliminates the associating signal because of the intermittence of the sign pieces.
The discontinuities happen because of the framing process. Assuming that we characterize the window as w (n), 0 ≤ n ≤ N - 1, where N is the number of samples in each frame, the result of windowing is a signal (equation 2):
From Equation 2, y(n) is the signal resulting from the convolution between the input signal and the window function, and x(n) represents the signal that the window function must convolve. Where w(n)generally uses the Hamming window, which has the form (Equation 3):
Hamming window is most popular due to its high-resolution performance and it powerfully shapes the separation in amplitude at its ends.
d) Fast Fourier Transformation (FFT)
Fourier transform converts a time series of bounded time-domain signals into a frequency spectrum. After the windowing method, this step will be applied, and the resulting frame is converted into a frequency spectrum.
FFT is a quick algorithm of DFT (Discrete Fourier transform) that is advantageous for changing each frame to N samples from the time domain into the frequency domain.
Fourier Transformation equation 4, shows n = 0, 1,2…, N-1 and j = sqrt-1. Xnis the n-frequency pattern produced from the FT, xkis the signal of a frame. The result of this stage is typically called Spectrum.
Throw it in the pot, said the old woman, as soon as he had told his story; put it in that basket, and as soon as the water boils in the pot we will hang it over the fire! Better kill it first, said the old man; and the hare listened, horribly frightened, but still looking secretly to see if there was no hole through which he could escape, if he had a chance of doing so
e) Mel filter bank
The Spectrum still has many data that isn't needed for ASR (Automatic Speech Recognition). Specifically,the cochlea cannot decide the difference between intently spaced frequencies. This impact is more articulated as the frequencies increase. Thus, we take clumps of period gram bins and sum them to get an idea of ways much energy exists in different frequency areas.
This is performed by our Mel filter bank: the basic channel is narrow and offers an indication of what extent of energy exists near zero Hertz. as the frequencies get higher, our filters get wider as we have a tendency to become less involved in variations.
The human ear also propels this: we don't hear tumult on a linear scale. To double the perceived loudness of a sound, we generally have to put eight times more energy into it. This implies that enormous varieties in energy probably won't sound different assuming the tone is loud. This compression process makes our functions more closely match what individuals really hear.
The formula for calculating the Mel frequency for some random frequency f in Hz is given in Equation 5:
f) Discrete Cosine Transformation (DCT)
The final step is to calculate the DCT of the filter bank energies. Still, filter banks are overlapping, the filter bank energies are quite connected with every other. The DCT decorrelates the energies. In any case,notice that only 13 of the 26 DCT coefficients were kept. This is for the reason that the higher DCT coefficients represent quick changes in the filter bank energies, and it appears to be that these quick changes really reduce the performance, thus we get a little improvement by dropping them[17].
This is a process of normalizing the scope of independent factors or features in the data. In Data Preprocessing, this is called Normalization and is normally performed during the data pre-processing step.Since the range of values of the raw data is much different, in some AI algorithms the objective functions won't work accurately without Normalization. For instance, most classifiers find the distance between two points using Euclidean distance. Assuming one of the features has a wide range of values, that specific feature determines the distance. Therefore, the range of all features should normalize so that each feature contributes approximately proportionately to the final distance. Another reason why feature scaling applied is that gradient descent converges much faster with feature scaling than without it[10].There are various methods to do feature scaling such as Rescaling (min-max normalization), Mean normalization, Standardization,Scaling to unit length. For this research, rescaling method is used.
Rescaling: Also referred to as minimum-maximum scaling or minimum-maximum normalization, is that the simplest technique and consists in rescaling the range of features to scale the range in [0,1] or [−1, 1]. Picking the target range relies upon the nature of the data. The overall formula is given as(equation 6):
Where x - original value, x’ - normalized value.
It is important when the number of features is very large there is no need to use every feature. We can use features only those that are important. The top reasons are to use feature selections: It enables the machine-learning algorithm to train faster, it reduces the complexity of a model and makes it easier to interpret, it improves the accuracy of a model if the right subset chosen, and it reduces overfitting. There are various techniques and methodologies that can use to subset feature space and help the models perform better and more efficiently. Such one method is the wrapper method[14].
In wrapper methods, we attempt to use a subset of features and train a model using them. Based on the inferences that draw from the previous model, decide to add or remove features from the subset. The problem essentially reduced to a search problem. These methods are typically computationally very expensive.Some common samples of wrapper methods are forward feature selection, backward feature elimination,recursive feature elimination, and so forth.
Forward Selection: Forward selection is an iterative technique in which we start with having no feature in the model. In every iteration, we keep adding the feature which best improves our model till an addition of a new variable doesn't improve the performance of the model.
a) k-Nearest Neighbors
k-NN classification is one of the most basic and simple classification methods and should be one of the primary selections for a classification study once there's very little or no previous data about the distribution of the data. k-nearest-neighbor classification developed from the necessity to perform discriminant analysis when reliable parametric estimates of probability densities are unknown or difficult to determine.k-NN is a non-parametric method used for classification and regression. In each case, the input consists of the k closest training examples in the feature space.The output depends on whether k-NN used for classification or regression.
In the kNN classification, the result is a class membership. An object is classified by the majority vote of its neighbors, with the object being assigned the most common class among its k nearest neighbors (k is a positive integer, usually small). If k = 1, then the object is simply assigned to that nearest neighbor’s class. In kNN regression, the output is the object's property value. This value is the average of the values of its nearest neighbors[18].
b) SVM-based multi-class classifiers
SVM is a supervised machine-learning algorithm,which may be, used for both classification and regression challenges. However, it's mostly used in classification problems. In this algorithm, we plot each data item as a point in n-dimensional space (where n is that the number of features you have) with the value of each feature being the value of a particular coordinate[1].Then, we perform classification by finding the hyper-plane that differentiates the two classes very well[7]. There are three standard techniques often used by SVMs to tackle multi-class problems, namely one-versus-one (OVO), one-versus-all (OVA), and directed acyclic graph (DAG). Here we tend to consider the OVA method.
One-versus-all (OVA) method is implemented using a “winner-takes-all” strategy. It constructs k separate binary classifiers for k-class classification.The mthbinary classifier trained using the data from the mthclass as
Positive examples and the remaining k − 1 classes as negative examples. During test, the binary classifier that gives maximum output value determines the class label.
For an example, the winner-takes-all strategy assigns it to the class with the boundary function highest classification. Fig.4 shows the OVA architecture and also
Fig.4 OVA Architecture
the classification problems at each node for finding the best class out of four classes[19].
Feature extraction technique is the first step in a speech emotion classification. The feature generating process from the dataset called as feature extraction.There are many features available to extract from an audio file. The specialists are as yet examining which ones are best for each particular instance of the speech signal are processing. The most useful and used one is the MFCC features[16].
Therefore, MFCC features selected to use for this research work. MFCC feature extraction shown in Fig.6 and the Fig.5 illustrate the method of emotion recognition system through speech.
Fig.5 Diagram of Emotion Recognition System through speech
Fig.6 MFCC Extraction
At first, a high-pass filter was applied along with a pre-emphasis coefficient value of 0.97. From there on speech signals changed over into a sequence of feature frames having a window length of 25 msec. Then, at that point, the Hamming Window was applied utilizing 10 ms of hop time. After that, the signals are converted from the time domain to the frequency domain by Fast Fourier Transformation. Then, the filter bank according to the Mel scale is used to calculate the weighted sum of the filter spectral components so the result of the method for a Mel scale approaches. When the channel bank energies are gotten, the logarithm of these energies is calculated.
Finally, the log Mel range spectrum process is changed over into a time-domain utilizing DCT (Discrete Cosine Transformation). This change shows the outcome as MFCC, that depicted as a sequential arrangement of cepstral vectors.13 MFCC was used for this research[16]. Finally, 9 basic statistical measures(median, maximum, minimum, mean, standard deviation, inter-quartile range, skewness, kurtosis, and range)were calculated for each coefficient. And also grouped together resulting in a 117-dimensional feature set.Classification results using k-NN are shown in Table 1 and OVA-SVM showed in Table 2.
Each of the basic nine statistical measures selected individually and tested, results shown in Table 3.Then forward feature selection done and obtained results shown in Table 4.
First K-Nearest neighbor approach used for classification. That is the simplest classification approach.Would be a nearest neighbor voting strategy that computes all pairwise Euclidean distances between key point representations of a test set to all or any labeled sets within the dataset. Then we used the SVMs approach for classification. SVMs are a supervised learning technique based on a statistical learning theory which will use for pattern classification and regression. Once applied to classification, a linear SVM finds the hyperplane leaving the largest possible fraction of points of the same class on the same side,whereas maximizing the distance of either class from the hyperplane. In this project, we've got used OVASVM.
Overview of the experiment shown in Fig.7. The aim of the experiment was to clarify the accuracy of selected groups of features, as well as classification ability of selected classification methods for speech emotion classification. Samples of examination formed from recordings of human speech with various emotional characters. The following settings used in the experiment.
Fig.7 Architecture of Methodology
Berlin database of emotional utterances used as Dataset. The database contains of 535 Utterances spoken in German which consisting of seven emotions such as angry (one hundred and twenty-seven), bored (eighty-one), disgust (forty-six), fearful (sixty-nine), happy(seventy-one) emotions, sad (sixty-two) and neutral(seventy-nine) emotions. Numbers in parentheses indicate the number of utterances per emotion, each sentence consists of an average of ten words with an average duration of approximately five seconds, recorded by ten different actors (both genders). The audio files are in WAV file format.
For classification, the k-nearest neighbors’ classifier was used for 70% of training data and 30% of testing data. For the classification, the value of k is used as k=1, 3, 5, 7, 11, 13. OVA-SVMs used for 70% of training data and 30% of testing data. The optimal values c was used 2-14to 210. Then the classifications done in three categories. They are classification without scaling data, classification after Scaling data to [-1, 1] and select appropriate features and classify using SVMs.
Valence of the Berlin Database is shown in Fig.8.
useless for classification. Therefore, we appropriate feature selection. Some features are unhelpful for classification; if we use such features, it will affect the classification rate. Best feature selection allows the development of simpler and faster models. For this research, I have used the forward feature selection method. Classification rate using the subset of statistical measures (median, maximum, mean, Inter-quartile range, skewness) set gives better classification rate of 60%. The performance increased with feature selection.
我们致力于保护作者版权,注重分享,被刊用文章因无法核实真实出处,未能及时与作者取得联系,或有版权异议的,请联系管理员,我们会立即处理! 部分文章是来自各大过期杂志,内容仅供学习参考,不准确地方联系删除处理!