Adaptive Optimization-Enabled Neural Networks to Handle the Imbalance Churn Data in Churn Prediction

Author(s):  
Bharathi Garimella ◽  
G. V. S. N. R. V. Prasad ◽  
M. H. M. Krishna Prasad

The churn prediction based on telecom data has been paid great attention because of the increasing the number telecom providers, but due to inconsistent data, sparsity, and hugeness, the churn prediction becomes complicated and challenging. Hence, an effective and optimal prediction of churns mechanism, named adaptive firefly-spider optimization (adaptive FSO) algorithm, is proposed in this research to predict the churns using the telecom data. The proposed churn prediction method uses telecom data, which is the trending domain of research in predicting the churns; hence, the classification accuracy is increased. However, the proposed adaptive FSO algorithm is designed by integrating the spider monkey optimization (SMO), firefly optimization algorithm (FA), and the adaptive concept. The input data is initially given to the master node of the spark framework. The feature selection is carried out using Kendall’s correlation to select the appropriate features for further processing. Then, the selected unique features are given to the master node to perform churn prediction. Here, the churn prediction is made using a deep convolutional neural network (DCNN), which is trained by the proposed adaptive FSO algorithm. Moreover, the developed model obtained better performance using the metrics, like dice coefficient, accuracy, and Jaccard coefficient by varying the training data percentage and selected features. Thus, the proposed adaptive FSO-based DCNN showed improved results with a dice coefficient of 99.76%, accuracy of 98.65%, Jaccard coefficient of 99.52%.

Genetics ◽  
2021 ◽  
Author(s):  
Marco Lopez-Cruz ◽  
Gustavo de los Campos

Abstract Genomic prediction uses DNA sequences and phenotypes to predict genetic values. In homogeneous populations, theory indicates that the accuracy of genomic prediction increases with sample size. However, differences in allele frequencies and in linkage disequilibrium patterns can lead to heterogeneity in SNP effects. In this context, calibrating genomic predictions using a large, potentially heterogeneous, training data set may not lead to optimal prediction accuracy. Some studies tried to address this sample size/homogeneity trade-off using training set optimization algorithms; however, this approach assumes that a single training data set is optimum for all individuals in the prediction set. Here, we propose an approach that identifies, for each individual in the prediction set, a subset from the training data (i.e., a set of support points) from which predictions are derived. The methodology that we propose is a Sparse Selection Index (SSI) that integrates Selection Index methodology with sparsity-inducing techniques commonly used for high-dimensional regression. The sparsity of the resulting index is controlled by a regularization parameter (λ); the G-BLUP (the prediction method most commonly used in plant and animal breeding) appears as a special case which happens when λ = 0. In this study, we present the methodology and demonstrate (using two wheat data sets with phenotypes collected in ten different environments) that the SSI can achieve significant (anywhere between 5-10%) gains in prediction accuracy relative to the G-BLUP.


2022 ◽  
Author(s):  
Maxat Kulmanov ◽  
Robert Hoehndorf

Motivation: Protein functions are often described using the Gene Ontology (GO) which is an ontology consisting of over 50,000 classes and a large set of formal axioms. Predicting the functions of proteins is one of the key challenges in computational biology and a variety of machine learning methods have been developed for this purpose. However, these methods usually require significant amount of training data and cannot make predictions for GO classes which have only few or no experimental annotations. Results: We developed DeepGOZero, a machine learning model which improves predictions for functions with no or only a small number of annotations. To achieve this goal, we rely on a model-theoretic approach for learning ontology embeddings and combine it with neural networks for protein function prediction. DeepGOZero can exploit formal axioms in the GO to make zero-shot predictions, i.e., predict protein functions even if not a single protein in the training phase was associated with that function. Furthermore, the zero-shot prediction method employed by DeepGOZero is generic and can be applied whenever associations with ontology classes need to be predicted. Availability: http://github.com/bio-ontology-research-group/deepgozero


Kybernetes ◽  
2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Lantian Li ◽  
Bahareh Pahlevanzadeh

PurposeCloud eases information processing, but it holds numerous risks, including hacking and confidentiality problems. It puts businesses at risk in terms of data security and compliance. This paper aims to maximize the covered human resource (HR) vulnerabilities and minimize the security costs in the enterprise cloud using a fuzzy-based method and firefly optimization algorithm.Design/methodology/approachCloud computing provides a platform to improve the quality and availability of IT resources. It changes the way people communicate and conduct their businesses. However, some security concerns continue to derail the expansion of cloud-based systems into all parts of human life. Enterprise cloud security is a vital component in ensuring the long-term stability of cloud technology by instilling trust. In this paper, a fuzzy-based method and firefly optimization algorithm are suggested for optimizing HR vulnerabilities while mitigating security expenses in organizational cloud environments. MATLAB is employed as a simulation tool to assess the efficiency of the suggested recommendation algorithm. The suggested approach is based on the firefly algorithm (FA) since it is swift and reduces randomization throughout the lookup for an optimal solution, resulting in improved performance.FindingsThe fuzzy-based method and FA unveil better performance than existing met heuristic algorithms. Using a simulation, all the results are verified. The study findings showed that this method could simulate complex and dynamic security problems in cloud services.Practical implicationsThe findings may be utilized to assist the cloud provider or tenant of the cloud infrastructure system in taking appropriate risk mitigation steps.Originality/valueUsing a fuzzy-based method and FA to maximize the covered HR vulnerabilities and minimize the security costs in the enterprise cloud is the main novelty of this paper.


Sign in / Sign up

Export Citation Format

Share Document