expectation maximization
Recently Published Documents


TOTAL DOCUMENTS

1732
(FIVE YEARS 431)

H-INDEX

55
(FIVE YEARS 10)

2022 ◽  
Author(s):  
Lenore Pipes ◽  
Zihao Chen ◽  
Svetlana Afanaseva ◽  
Rasmus Nielsen

Wastewater surveillance has become essential for monitoring the spread of SARS-CoV-2. The quantification of SARS-CoV-2 RNA in wastewater correlates with the Covid-19 caseload in a community. However, estimating the proportions of different SARS-CoV-2 strains has remained technically difficult. We present a method for estimating the relative proportions of SARS-CoV-2 strains from wastewater samples. The method uses an initial step to remove unlikely strains, imputation of missing nucleotides using the global SARS-CoV-2 phylogeny, and an Expectation-Maximization (EM) algorithm for obtaining maximum likelihood estimates of the proportions of different strains in a sample. Using simulations with a reference database of >3 million SARS-CoV-2 genomes, we show that the estimated proportions accurately reflect the true proportions given sufficiently high sequencing depth and that the phylogenetic imputation is highly accurate and substantially improves the reference database.


2022 ◽  
Vol 2022 ◽  
pp. 1-10
Author(s):  
Shaowen Tan ◽  
Zili Xu

In this study, dictionary learning and expectation maximization reconstruction (DLEM) was combined to denoise 64-slice spiral CT images, and results of coronary angiography (CAG) were used as standard to evaluate its clinical value in diagnosing coronary artery diseases. 120 patients with coronary heart disease (CHD) confirmed by CAG examination were retrospectively selected as the research subjects. According to the random number table method, the patients were divided into two groups: the control group was diagnosed by conventional 64-slice spiral CT images, and the observation group was diagnosed by 64-slice spiral CT images based on the DLEM algorithm, with 60 cases in both groups. With CAG examination results as the standard, the diagnostic effects of the two CT examination methods were compared. The results showed that when the number of iterations of maximum likelihood expectation maximization (MLEM) algorithm reached 50, the root mean square error (RMSE) and peak signal to noise ratio (PSNR) values were similar to the results obtained by the DLEM algorithm under a number of iterations of 10 when the RMSE and PSNR values were 18.9121 dB and 74.9911 dB, respectively. In the observation group, 28.33% (17/60) images were of grade 4 or above before processing; after processing, it was 70% (42/60), significantly higher than the proportion of high image quality before processing. The overall diagnostic consistency, sensitivity, specificity, and accuracy (88.33%, 86.67%, 80%, and 85%) of the observation group were better than those in the control group (60.46%, 62.5%, 58.33%, and 61.66%). In conclusion, the DLEM algorithm has good denoising effect on 64-slice spiral CT images, which significantly improves the accuracy in the diagnosis of coronary artery stenosis and has good clinical diagnostic value and is worth promoting.


2022 ◽  
pp. 1-32
Author(s):  
Martin Bladt

Abstract This paper addresses the task of modeling severity losses using segmentation when the data distribution does not fall into the usual regression frameworks. This situation is not uncommon in lines of business such as third-party liability insurance, where heavy-tails and multimodality often hamper a direct statistical analysis. We propose to use regression models based on phase-type distributions, regressing on their underlying inhomogeneous Markov intensity and using an extension of the expectation–maximization algorithm. These models are interpretable and tractable in terms of multistate processes and generalize the proportional hazards specification when the dimension of the state space is larger than 1. We show that the combination of matrix parameters, inhomogeneity transforms, and covariate information provides flexible regression models that effectively capture the entire distribution of loss severities.


SoftwareX ◽  
2022 ◽  
Vol 17 ◽  
pp. 100944
Author(s):  
Parichit Sharma ◽  
Hasan Kurban ◽  
Mehmet Dalkilic

2022 ◽  
pp. 108180
Author(s):  
Aiman Solyman ◽  
Wang Zhenyu ◽  
Tao Qian ◽  
Arafat Abdulgader Mohammed Elhag ◽  
Zhang Rui ◽  
...  

Author(s):  
S. Nickolas ◽  
K. Shobha

Data pre-processing plays a vital role in the life cycle of data mining for accomplishing quality outcomes. In this paper, it is experimentally shown the importance of data pre-processing to achieve highly accurate classifier outcomes by imputing missing values using a novel imputation method, CLUSTPRO, by selecting highly correlated features using Correlation-based Variable Selection (CVS) and by handling imbalanced data using Synthetic Minority Over-sampling Technique (SMOTE). The proposed CLUSTPRO method makes use of Random Forest (RF) and Expectation Maximization (EM) algorithms to impute missing. The imputed results are evaluated using standard evaluation metrics. The CLUSTPRO imputation method outperforms existing, state-of-the-art imputation methods. The combined approach of imputation, feature selection, and imbalanced data handling techniques has significantly contributed to attaining an improved classification accuracy (AUC curve) of 40%–50% in comparison with results obtained without any pre-processing.


2021 ◽  
Vol 2021 ◽  
pp. 1-13
Author(s):  
Weisan Wu

In this paper, we give a modified gradient EM algorithm; it can protect the privacy of sensitive data by adding discrete Gaussian mechanism noise. Specifically, it makes the high-dimensional data easier to process mainly by scaling, truncating, noise multiplication, and smoothing steps on the data. Since the variance of discrete Gaussian is smaller than that of the continuous Gaussian, the difference privacy of data can be guaranteed more effectively by adding the noise of the discrete Gaussian mechanism. Finally, the standard gradient EM algorithm, clipped algorithm, and our algorithm (DG-EM) are compared with the GMM model. The experiments show that our algorithm can effectively protect high-dimensional sensitive data.


Sign in / Sign up

Export Citation Format

Share Document