scholarly journals Heart Sound Classification Based on Fractal Dimension and MFCC Features Using Hidden Markov Model

Author(s):  
Mahbubeh Bahreini ◽  
Ramin Barati ◽  
Abbas Kamaly

Abstract Early diagnosis is crucial in the treatment of heart diseases. Researchers have applied a variety of techniques for cardiovascular disease diagnosis, including the detection of heart sounds. It is an efficient and affordable diagnosis technique. Body organs, including the heart, generate several sounds. These sounds are different in different individuals. A number of methodologies have been recently proposed to detect and diagnose normal/abnormal sounds generated by the heart. The present study proposes a technique on the basis of the Mel-frequency cepstral coefficients, fractal dimension, and hidden Markov model. It uses the fractal dimension to identify sounds S1 and S2. Then, the Mel-frequency cepstral coefficients and the first- and second-order difference Mel-frequency cepstral coefficients are employed to extract the features of the signals. The adaptive Hemming window length is a major advantage of the methodology. The S1-S2 interval determines the adaptive length. Heart sounds are divided into normal and abnormal through the improved hidden Markov model and Baum-Welch and Viterbi algorithms. The proposed framework is evaluated using a number of datasets under various scenarios.

2014 ◽  
Vol 2014 ◽  
pp. 1-7 ◽  
Author(s):  
Md. Rabiul Islam ◽  
Md. Abdus Sobhan

The aim of the paper is to propose a feature fusion based Audio-Visual Speaker Identification (AVSI) system with varied conditions of illumination environments. Among the different fusion strategies, feature level fusion has been used for the proposed AVSI system where Hidden Markov Model (HMM) is used for learning and classification. Since the feature set contains richer information about the raw biometric data than any other levels, integration at feature level is expected to provide better authentication results. In this paper, both Mel Frequency Cepstral Coefficients (MFCCs) and Linear Prediction Cepstral Coefficients (LPCCs) are combined to get the audio feature vectors and Active Shape Model (ASM) based appearance and shape facial features are concatenated to take the visual feature vectors. These combined audio and visual features are used for the feature-fusion. To reduce the dimension of the audio and visual feature vectors, Principal Component Analysis (PCA) method is used. The VALID audio-visual database is used to measure the performance of the proposed system where four different illumination levels of lighting conditions are considered. Experimental results focus on the significance of the proposed audio-visual speaker identification system with various combinations of audio and visual features.


he proposed research is dedicated to verifying the claimed emotion of speaker-independent and text-independent formed on three dissimilar classifiers. The HMM3 short for Third-Order Hidden Markov Model, HMM2 short for Second-Order Hidden Markov Model, and HMM1 short for First-Order Hidden Markov Model are the three classifiers utilized in this study. Our work has been evaluated on our collected Emirati-accented speech corpus which entails 50 speakers of Emirati origin (25 female and 25 male) uttering sentences in six emotions by means of the extracted features by Mel-Frequency Cepstral Coefficients (MFCCs). Our outcomes prove that HMM3 is superior to each of HMM1 and HMM2 to authenticate the claimed emotion. The achieved results formed on HMM3 are very similar to the outcomes attained in the subjective valuation by Arab listeners.


Sign in / Sign up

Export Citation Format

Share Document