scholarly journals A Low Complex Algorithm for Detection of Sleep Apnea from ECG Signals using Sliding SSA Combined with PCA, KPCA, and SPPCA

Author(s):  
Muhammad Zubair

Sleep apnea is a potentially life-threatening sleep condition in which breathing stops and resumes repeatedly. It is caused by breathing pauses during sleep, which leads to frequent awakenings. As we all know, computational time and efficiency are essential in the healthcare industry; to address this issue, we proposed an algorithm that performs more computations in less time without compromising the machine learning model’s performance. This study employs a unique technique called Sliding Singular Spectrum Analysis (SSSA) to decompose and de-noise the ECG signals. To identify the significant apnea and non-apnea components from the pre-processed ECG data and to reduce the dimensionality, we used Principal Component Analysis (PCA), Kernal PCA (KPCA), and Sub-Pattern based PCA (SPPCA). These characteristics were then used to train and evaluate various machine learning models, including KNN, SVM, GaussianNB, SGD, and XGBoost, to distinguish between apnea and non-apnea ECG data. The publicly available Physionet Apnea-ECG database is used for the simulation of the proposed algorithm. To verify the performance of machine learning models, we have calculated various metrics like accuracy, precision, recall and F1 score. The validation of the proposed method is done by comparing the classification metrics with the latest state-of-the-art works.

2021 ◽  
Author(s):  
Muhammad Zubair

Sleep apnea is a potentially life-threatening sleep condition in which breathing stops and resumes repeatedly. It is caused by breathing pauses during sleep, which leads to frequent awakenings. As we all know, computational time and efficiency are essential in the healthcare industry; to address this issue, we proposed an algorithm that performs more computations in less time without compromising the machine learning model’s performance. This study employs a unique technique called Sliding Singular Spectrum Analysis (SSSA) to decompose and de-noise the ECG signals. To identify the significant apnea and non-apnea components from the pre-processed ECG data and to reduce the dimensionality, we used Principal Component Analysis (PCA), Kernal PCA (KPCA), and Sub-Pattern based PCA (SPPCA). These characteristics were then used to train and evaluate various machine learning models, including KNN, SVM, GaussianNB, SGD, and XGBoost, to distinguish between apnea and non-apnea ECG data. The publicly available Physionet Apnea-ECG database is used for the simulation of the proposed algorithm. To verify the performance of machine learning models, we have calculated various metrics like accuracy, precision, recall and F1 score. The validation of the proposed method is done by comparing the classification metrics with the latest state-of-the-art works.


2021 ◽  
Author(s):  
Muhammad Zubair

<pre>Alcoholism is a widely affected disorder that leads to critical brain deficiencies such as emotional and behavioural impairments. One of the prominent sources to detect alcoholism is by analysing Electroencephalogram (EEG) signals. Previously, most of the works have focused on detecting alcoholism using various machine and deep learning algorithms. This paper has used a novel algorithm named Sliding Singular Spectrum Analysis (S-SSA) to decompose and de-noise the EEG signals. We have considered independent component analysis (ICA) to select the prominent alcoholic and non-alcoholic components from the preprocessed EEG data. Later, these components were used to train and test various machine learning models like SVM, KNN, ANN, GBoost, AdaBoost and XGBoost to classify alcoholic and non-alcoholic EEG signals. The sliding SSA-ICA algorithm helps in reducing the computational time and complexity of the machine learning models. To validate the performance of the ICA algorithm, we have compared the computational time and accuracy of ICA with its counterpart, like principal component analysis (PCA). The proposed algorithm is tested on a publicly available UCI alcoholic EEG dataset. To verify the performance of machine learning models, we have calculated various metrics like accuracy, precision, recall and F1 score. Our work reported the highest accuracy of 98.97% with the XGBoost classifier. The validation of the proposed method is done by comparing the classification metrics with the latest state-of-the-art works.</pre>


2021 ◽  
Author(s):  
Muhammad Zubair

<pre>Alcoholism is a widely affected disorder that leads to critical brain deficiencies such as emotional and behavioural impairments. One of the prominent sources to detect alcoholism is by analysing Electroencephalogram (EEG) signals. Previously, most of the works have focused on detecting alcoholism using various machine and deep learning algorithms. This paper has used a novel algorithm named Sliding Singular Spectrum Analysis (S-SSA) to decompose and de-noise the EEG signals. We have considered independent component analysis (ICA) to select the prominent alcoholic and non-alcoholic components from the preprocessed EEG data. Later, these components were used to train and test various machine learning models like SVM, KNN, ANN, GBoost, AdaBoost and XGBoost to classify alcoholic and non-alcoholic EEG signals. The sliding SSA-ICA algorithm helps in reducing the computational time and complexity of the machine learning models. To validate the performance of the ICA algorithm, we have compared the computational time and accuracy of ICA with its counterpart, like principal component analysis (PCA). The proposed algorithm is tested on a publicly available UCI alcoholic EEG dataset. To verify the performance of machine learning models, we have calculated various metrics like accuracy, precision, recall and F1 score. Our work reported the highest accuracy of 98.97% with the XGBoost classifier. The validation of the proposed method is done by comparing the classification metrics with the latest state-of-the-art works.</pre>


SLEEP ◽  
2021 ◽  
Vol 44 (Supplement_2) ◽  
pp. A164-A164
Author(s):  
Pahnwat Taweesedt ◽  
JungYoon Kim ◽  
Jaehyun Park ◽  
Jangwoon Park ◽  
Munish Sharma ◽  
...  

Abstract Introduction Obstructive sleep apnea (OSA) is a common sleep-related breathing disorder with an estimation of one billion people. Full-night polysomnography is considered the gold standard for OSA diagnosis. However, it is time-consuming, expensive and is not readily available in many parts of the world. Many screening questionnaires and scores have been proposed for OSA prediction with high sensitivity and low specificity. The present study is intended to develop models with various machine learning techniques to predict the severity of OSA by incorporating features from multiple questionnaires. Methods Subjects who underwent full-night polysomnography in Torr sleep center, Texas and completed 5 OSA screening questionnaires/scores were included. OSA was diagnosed by using Apnea-Hypopnea Index ≥ 5. We trained five different machine learning models including Deep Neural Networks with the scaled principal component analysis (DNN-PCA), Random Forest (RF), Adaptive Boosting classifier (ABC), and K-Nearest Neighbors classifier (KNC) and Support Vector Machine Classifier (SVMC). Training:Testing subject ratio of 65:35 was used. All features including demographic data, body measurement, snoring and sleepiness history were obtained from 5 OSA screening questionnaires/scores (STOP-BANG questionnaires, Berlin questionnaires, NoSAS score, NAMES score and No-Apnea score). Performance parametrics were used to compare between machine learning models. Results Of 180 subjects, 51.5 % of subjects were male with mean (SD) age of 53.6 (15.1). One hundred and nineteen subjects were diagnosed with OSA. Area Under the Receiver Operating Characteristic Curve (AUROC) of DNN-PCA, RF, ABC, KNC, SVMC, STOP-BANG questionnaire, Berlin questionnaire, NoSAS score, NAMES score, and No-Apnea score were 0.85, 0.68, 0.52, 0.74, 0.75, 0.61, 0.63, 0,61, 0.58 and 0,58 respectively. DNN-PCA showed the highest AUROC with sensitivity of 0.79, specificity of 0.67, positive-predictivity of 0.93, F1 score of 0.86, and accuracy of 0.77. Conclusion Our result showed that DNN-PCA outperforms OSA screening questionnaires, scores and other machine learning models. Support (if any):


2019 ◽  
Vol 111 ◽  
pp. 05019
Author(s):  
Brian de Keijzer ◽  
Pol de Visser ◽  
Víctor García Romillo ◽  
Víctor Gómez Muñoz ◽  
Daan Boesten ◽  
...  

Machine learning models have proven to be reliable methods in the forecasting of energy use in commercial and office buildings. However, little research has been done on energy forecasting in dwellings, mainly due to the difficulty of obtaining household level data while keeping the privacy of inhabitants in mind. Gaining insight into the energy consumption in the near future can be helpful in balancing the grid and insights in how to reduce the energy consumption can be received. In collaboration with OPSCHALER, a measurement campaign on the influence of housing characteristics on energy costs and comfort, several machine learning models were compared on forecasting performance and the computational time needed. Nine months of data containing the mean gas consumption of 52 dwellings on a one hour resolution was used for this research. The first 6 months were used for training, whereas the last 3 months were used to evaluate the models. The results showed that the Deep Neural Network (DNN) performed best with a 50.1 % Mean Absolute Percentage Error (MAPE) on a one hour resolution. When comparing daily and weekly resolutions, the Multivariate Linear Regression (MVLR) outperformed other models, with a 20.1 % and 17.0 % MAPE, respectively. The models were programmed in Python.


2021 ◽  
Author(s):  
Chang H Kim ◽  
Sadeer Al-Kindi ◽  
Yasir Tarabichi ◽  
Suril Gohel ◽  
Riddhi Vyas ◽  
...  

Background: The value of the electrocardiogram (ECG) for predicting long-term cardiovascular outcomes is not well defined. Machine learning methods are well suited for analysis of highly correlated data such as that from the ECG. Methods: Using demographic, clinical, and 12-lead ECG data from the Third National Health and Nutrition Examination Survey (NHANES III), machine learning models were trained to predict 10-year cardiovascular mortality in ambulatory U.S. adults. Predictive performance of each model was assessed using area under receiver operating characteristic curve (AUROC), area under precision-recall curve (AUPRC), sensitivity, and specificity. These were compared to the 2013 American College of Cardiology/American Heart Association Pooled Cohort Equations (PCE). Results: 7,067 study participants (mean age: 59.2 +/- 13.4 years, female: 52.5%, white: 73.9%, black: 23.3%) were included. At 10 years of follow up, 338 (4.8%) had died from cardiac causes. Compared to the PCE (AUROC: 0.668, AUPRC: 0.125, sensitivity: 0.492, specificity: 0.859), machine learning models only required demographic and ECG data to achieve comparable performance: logistic regression (AUROC: 0.754, AUPRC: 0.141, sensitivity: 0.747, specificity: 0.759), neural network (AUROC: 0.764, AUPRC: 0.149, sensitivity: 0.722, specificity: 0.787), and ensemble model (AUROC: 0.695, AUPRC: 0.166, sensitivity: 0.468, specificity: 0.912). Additional clinical data did not improve the predictive performance of machine learning models. In variable importance analysis, important ECG features clustered in inferior and lateral leads. Conclusions: Machine learning can be applied to demographic and ECG data to predict 10-year cardiovascular mortality in ambulatory adults, with potentially important implications for primary prevention.


2021 ◽  
Vol 12 (1) ◽  
pp. 269
Author(s):  
Máté Szűcs ◽  
Tamás Szepesi ◽  
Christoph Biedermann ◽  
Gábor Cseh ◽  
Marcin Jakubowski ◽  
...  

The detachment regime has a high potential to play an important role in fusion devices on the road to a fusion power plant. Complete power detachment has been observed several times during the experimental campaigns of the Wendelstein 7-X (W7-X) stellarator. Automatic observation and signaling of such events could help scientists to better understand these phenomena. With the growing discharge times in fusion devices, machine learning models and algorithms are a powerful tool to process the increasing amount of data. We investigate several classical supervised machine learning models to detect complete power detachment in the images captured by the Event Detection Intelligent Camera System (EDICAM) at the W7-X at each given image frame. In the dedicated detached state the plasma is stable despite its reduced contact with the machine walls and the radiation belt stays close to the separatrix, without exhibiting significant heat load onto the divertor. To decrease computational time and resources needed we propose certain pixel intensity profiles (or intensity values along lines) as the input to these models. After finding the profile that describes the images best in terms of detachment, we choose the best performing machine learning algorithm. It achieves an F1 score of 0.9836 on the training dataset and 0.9335 on the test set. Furthermore, we investigate its predictions in other scenarios, such as plasmas with substantially decreased minor radius and several magnetic configurations.


Author(s):  
Andrew McDonald ◽  

Decades of subsurface exploration and characterization have led to the collation and storage of large volumes of well-related data. The amount of data gathered daily continues to grow rapidly as technology and recording methods improve. With the increasing adoption of machine-learning techniques in the subsurface domain, it is essential that the quality of the input data is carefully considered when working with these tools. If the input data are of poor quality, the impact on precision and accuracy of the prediction can be significant. Consequently, this can impact key decisions about the future of a well or a field. This study focuses on well-log data, which can be highly multidimensional, diverse, and stored in a variety of file formats. Well-log data exhibits key characteristics of big data: volume, variety, velocity, veracity, and value. Well data can include numeric values, text values, waveform data, image arrays, maps, and volumes. All of which can be indexed by time or depth in a regular or irregular way. A significant portion of time can be spent gathering data and quality checking it prior to carrying out petrophysical interpretations and applying machine-learning models. Well-log data can be affected by numerous issues causing a degradation in data quality. These include missing data ranging from single data points to entire curves, noisy data from tool-related issues, borehole washout, processing issues, incorrect environmental corrections, and mislabeled data. Having vast quantities of data does not mean it can all be passed into a machine-learning algorithm with the expectation that the resultant prediction is fit for purpose. It is essential that the most important and relevant data are passed into the model through appropriate feature selection techniques. Not only does this improve the quality of the prediction, but it also reduces computational time and can provide a better understanding of how the models reach their conclusion. This paper reviews data quality issues typically faced by petrophysicists when working with well-log data and deploying machine-learning models. This is achieved by first providing an overview of machine learning and big data within the petrophysical domain, followed by a review of the common well-log data issues, their impact on machine-learning algorithms, and methods for mitigating their influence.


Sign in / Sign up

Export Citation Format

Share Document