scholarly journals Mean Received Resources Meet Machine Learning Algorithms to Improve Link Prediction Methods

Information ◽  
2022 ◽  
Vol 13 (1) ◽  
pp. 35
Author(s):  
Jibouni Ayoub ◽  
Dounia Lotfi ◽  
Ahmed Hammouch

The analysis of social networks has attracted a lot of attention during the last two decades. These networks are dynamic: new links appear and disappear. Link prediction is the problem of inferring links that will appear in the future from the actual state of the network. We use information from nodes and edges and calculate the similarity between users. The more users are similar, the higher the probability of their connection in the future will be. The similarity metrics play an important role in the link prediction field. Due to their simplicity and flexibility, many authors have proposed several metrics such as Jaccard, AA, and Katz and evaluated them using the area under the curve (AUC). In this paper, we propose a new parameterized method to enhance the AUC value of the link prediction metrics by combining them with the mean received resources (MRRs). Experiments show that the proposed method improves the performance of the state-of-the-art metrics. Moreover, we used machine learning algorithms to classify links and confirm the efficiency of the proposed combination.

Author(s):  
Prof. Gowrishankar B S

Stock market is one of the most complicated and sophisticated ways to do business. Small ownerships, brokerage corporations, banking sectors, all depend on this very body to make revenue and divide risks; a very complicated model. However, this paper proposes to use machine learning algorithms to predict the future stock price for exchange by using pre-existing algorithms to help make this unpredictable format of business a little more predictable. The use of machine learning which makes predictions based on the values of current stock market indices by training on their previous values. Machine learning itself employs different models to make prediction easier and authentic. The data has to be cleansed before it can be used for predictions. This paper focuses on categorizing various methods used for predictive analytics in different domains to date, their shortcomings.


2019 ◽  
Vol 19 (03) ◽  
pp. 1950014
Author(s):  
ALFREDO ARANDA ◽  
ALVARO VALENCIA

Fluid-mechanical and morphological parameters are recognized as major factors in the rupture risk of human aneurysms. On the other hand, it is well known that a lot of machine learning tools are available to study a variety of problems in many fields. In this work, fluid–structure interaction (FSI) simulations were carried out to examine a database of 60 real saccular cerebral aneurysms (30 ruptured and 30 unruptured) using reconstructions by angiography images. With the results of the simulations and geometric analyses, we studied the analysis of variance (ANOVA) statistic test in many variables and we obtained that aspect ratio (AR), bottleneck factor (BNF), maximum height of the aneurysms (MH), relative residence time (RRT), Womersley number (WN) and Von-Mises strain (VMS) are statically significant and good predictors for the models. In consequence, these ones were used in five machine learning algorithms to determine the rupture risk predictions of the aneurysms, where the adaptative boosting (AdaBoost) was calculated with the highest area under the curve (AUC) in the receiver operating characteristic (ROC) curve (AUC 0.944).


Cancers ◽  
2020 ◽  
Vol 12 (10) ◽  
pp. 3010
Author(s):  
Johannes Uhlig ◽  
Andreas Leha ◽  
Laura M. Delonge ◽  
Anna-Maria Haack ◽  
Brian Shuch ◽  
...  

This study evaluates the diagnostic performance of radiomic features and machine learning algorithms for renal tumor subtype assessment in venous computed tomography (CT) studies from clinical routine. Patients undergoing surgical resection and histopathological assessment of renal tumors at a tertiary referral center between 2012 and 2019 were included. Preoperative venous-phase CTs from multiple referring imaging centers were segmented, and standardized radiomic features extracted. After preprocessing, class imbalance handling, and feature selection, machine learning algorithms were used to predict renal tumor subtypes using 10-fold cross validation, assessed as multiclass area under the curve (AUC). In total, n = 201 patients were included (73.7% male; mean age 66 ± 11 years), with n = 131 clear cell renal cell carcinomas (ccRCC), n = 29 papillary RCC, n = 11 chromophobe RCC, n = 16 oncocytomas, and n = 14 angiomyolipomas (AML). An extreme gradient boosting algorithm demonstrated the highest accuracy (multiclass area under the curve (AUC) = 0.72). The worst discrimination was evident for oncocytomas vs. AML and oncocytomas vs. chromophobe RCC (AUC = 0.55 and AUC = 0.45, respectively). In sensitivity analyses excluding oncocytomas, a random forest algorithm showed the highest accuracy, with multiclass AUC = 0.78. Radiomic feature analyses from venous-phase CT acquired in clinical practice with subsequent machine learning can discriminate renal tumor subtypes with moderate accuracy. The classification of oncocytomas seems to be the most complex with the lowest accuracy.


2019 ◽  
Vol 37 (15_suppl) ◽  
pp. 2581-2581 ◽  
Author(s):  
Paul Johannet ◽  
Nicolas Coudray ◽  
George Jour ◽  
Douglas MacArthur Donnelly ◽  
Shirin Bajaj ◽  
...  

2581 Background: There is growing interest in optimizing patient selection for treatment with immune checkpoint inhibitors (ICIs). We postulate that phenotypic features present in metastatic melanoma tissue reflect the biology of tumor cells, immune cells, and stromal tissue, and hence can provide predictive information about tumor behavior. Here, we test the hypothesis that machine learning algorithms can be trained to predict the likelihood of response and/or toxicity to ICIs. Methods: We examined 124 stage III/IV melanoma patients who received anti-CTLA-4 (n = 81), anti-PD-1 (n = 25), or combination (n = 18) therapy as first line. The tissue analyzed was resected before treatment with ICIs. In total, 340 H&E slides were digitized and annotated for three regions of interest: tumor, lymphocytes, and stroma. The slides were then partitioned into training (n = 285), validation (n = 26), and test (n = 29) sets. Slides were tiled (299x299 pixels) at 20X magnification. We trained a deep convolutional neural network (DCNN) to automatically segment the images into each of the three regions and then deconstruct images into their component features to detect non-obvious patterns with objectivity and reproducibility. We then trained the DCNN for two classifications: 1) complete/partial response versus progression of disease (POD), and 2) severe versus no immune-related adverse events (irAEs). Predictive accuracy was estimated by area under the curve (AUC) of receiver operating characteristics (ROC). Results: The DCNN identified tumor within LN with AUC 0.987 and within ST with AUC 0.943. Prediction of POD based on ST-only always performed better than prediction based on LN-only (AUC 0.84 compared to 0.61, respectively). The DCNN had an average AUC 0.69 when analyzing only tumor regions from both LN and ST data sets and AUC 0.68 when analyzing tumor and lymphocyte regions. Severe irAEs were predicted with limited accuracy (AUC 0.53). Conclusions: Our results support the potential application of machine learning on pre-treatment histologic slides to predict response to ICIs. It also revealed their limited value in predicting toxicity. We are currently investigating whether the predictive capability of the algorithm can be further improved by incorporating additional immunologic biomarkers.


2020 ◽  
Author(s):  
Nida Fatima

Abstract Background: Preoperative prognostication of clinical and surgical outcome in patients with neurosurgical diseases can improve the risk stratification, thus can guide in implementing targeted treatment to minimize these events. Therefore, the author aims to highlight the development and validation of predictive models determining neurosurgical outcomes through machine learning algorithms using logistic regression.Methods: Logistic regression (enter, backward and forward) and least absolute shrinkage and selection operator (LASSO) method for selection of variables from selected database can eventually lead to multiple candidate models. The final model with a set of predictive variables must be selected based upon the clinical knowledge and numerical results.Results: The predictive model which performed best on the discrimination, calibration, Brier score and decision curve analysis must be selected to develop machine learning algorithms. Logistic regression should be compared with the LASSO model. Usually for the big databases, the predictive model selected through logistic regression gives higher Area Under the Curve (AUC) than those with LASSO model. The predictive probability derived from the best model could be uploaded to an open access web application which is easily deployed by the patients and surgeons to make a risk assessment world-wide.Conclusions: Machine learning algorithms provide promising results for the prediction of outcomes following cranial and spinal surgery. These algorithms can provide useful factors for patient-counselling, assessing peri-operative risk factors, and predicting post-operative outcomes after neurosurgery.


2020 ◽  
Vol 10 (1) ◽  
Author(s):  
Bum-Joo Cho ◽  
Kyoung Min Kim ◽  
Sanchir-Erdene Bilegsaikhan ◽  
Yong Joon Suh

Abstract Febrile neutropenia (FN) is one of the most concerning complications of chemotherapy, and its prediction remains difficult. This study aimed to reveal the risk factors for and build the prediction models of FN using machine learning algorithms. Medical records of hospitalized patients who underwent chemotherapy after surgery for breast cancer between May 2002 and September 2018 were selectively reviewed for development of models. Demographic, clinical, pathological, and therapeutic data were analyzed to identify risk factors for FN. Using machine learning algorithms, prediction models were developed and evaluated for performance. Of 933 selected inpatients with a mean age of 51.8 ± 10.7 years, FN developed in 409 (43.8%) patients. There was a significant difference in FN incidence according to age, staging, taxane-based regimen, and blood count 5 days after chemotherapy. The area under the curve (AUC) built based on these findings was 0.870 on the basis of logistic regression. The AUC improved by machine learning was 0.908. Machine learning improves the prediction of FN in patients undergoing chemotherapy for breast cancer compared to the conventional statistical model. In these high-risk patients, primary prophylaxis with granulocyte colony-stimulating factor could be considered.


Author(s):  
Gerardo Mario Ortigoza Capetillo ◽  
Alberto Pedro Lorandi Medina

En este trabajo analizamos escenarios hipotéticos para contagios de COVID-19 durante la elección 2021 en México. Del 2 de abril al 2 de junio 2021 se llevarán a cabo elecciones de diputados federales, diputados locales, gubernaturas y presidencias municipales en lo que es considerada como la elección más grande en la historia de México; se estima que las actividades de las campañas electorales y el día de la votación se incrementará la movilidad de las personas y con ello su riesgo de contagio por COVID-19. Usando datos históricos de razones de contagios se define la media de estos datos, su desviación estándar y mediante una distribución t-Student se obtiene un intervalo de 90% de confianza para la media. Se utilizan el centro y ambos extremos de este intervalo como tasas de incremento para simular el crecimiento de casos en dos periodos (primer mes: elección diputados federales; segundo mes: elección gubernaturas, diputados locales y ayuntamientos); se reportan simulaciones usando algoritmos de aprendizaje de máquina a 2 meses pasadas las elecciones.Palabras clave: aprendizaje máquina, proyecciones COVID-19, elección 2021 México.SUMMARYIn this work we analyze hypothetical scenarios for COVID-19 infections during the 2021 election in Mexico; from april 2 to june 2, 2021, elections for federal deputies, local deputies, governorships and municipal presidencies will be held in what is considered the largest election in Mexico´s history; it is estimated that the activities of the electoral campaigns and the election day will increase the mobility of people and with it their risk of contagion by COVID-19. Using historical data on infection rates, the mean of these data is defined, its standard deviation and a t-Student distribution is used to obtain a 90% confidence interval for the mean. The center and both ends of this interval are used as rates of increase to simulate the growth of cases in two periods (first month; election of federal deputies; second month; election of governorships, local deputies and municipalities), simulations are reported using machine learning algorithms 2 monts after the elections.Keywords: machine learning, COVID-19 projections, Mexico 2021 electionINTRODUCCIÓNAl momento de escribir este trabajo, se han confirmado alrededor de 110 millones de casos de


Author(s):  
Julien Lepine ◽  
Vincent Rouillard ◽  
Michael Sek

Road surface imperfections and aberrations generate shocks causing vehicles to sustain structural fatigue and functional defects, driver and passenger discomfort, injuries, and damage to freight. The harmful effect of shocks can be mitigated at different levels, for example, by improving road surfaces, vehicle suspension and protective packaging of freight. The efficiency of these methods partly depends on the identification and characterisation of the shocks. An assessment of four machine learning algorithms (Classifiers) that can be used to identify shocks produced on different roads and test tracks is presented in this paper. The algorithms were trained using synthetic signals. These were created from a model made from acceleration measurements on a test vehicle. The trained Classifiers were assessed on different measurement signals made on the same vehicle. The results show that the Support Vector Machine detection algorithm used in conjunction with a Gaussian Kernel Transform can accurately detect shocks generated on the test track with an area under the curve (AUC) of 0.89 and a Pseudo Energy Ratio Fall-Out (PERFO) of 8%.


Forests ◽  
2021 ◽  
Vol 12 (4) ◽  
pp. 395
Author(s):  
Milan Koreň ◽  
Rastislav Jakuš ◽  
Martin Zápotocký ◽  
Ivan Barka ◽  
Jaroslav Holuša ◽  
...  

Machine learning algorithms (MLAs) are used to solve complex non-linear and high-dimensional problems. The objective of this study was to identify the MLA that generates an accurate spatial distribution model of bark beetle (Ips typographus L.) infestation spots. We first evaluated the performance of 2 linear (logistic regression, linear discriminant analysis), 4 non-linear (quadratic discriminant analysis, k-nearest neighbors classifier, Gaussian naive Bayes, support vector classification), and 4 decision trees-based MLAs (decision tree classifier, random forest classifier, extra trees classifier, gradient boosting classifier) for the study area (the Horní Planá region, Czech Republic) for the period 2003–2012. Each MLA was trained and tested on all subsets of the 8 explanatory variables (distance to forest damage spots from previous year, distance to spruce forest edge, potential global solar radiation, normalized difference vegetation index, spruce forest age, percentage of spruce, volume of spruce wood per hectare, stocking). The mean phi coefficient of the model generated by extra trees classifier (ETC) MLA with five explanatory variables for the period was significantly greater than that of most forest damage models generated by the other MLAs. The mean true positive rate of the best ETC-based model was 80.4%, and the mean true negative rate was 80.0%. The spatio-temporal simulations of bark beetle-infested forests based on MLAs and GIS tools will facilitate the development and testing of novel forest management strategies for preventing forest damage in general and bark beetle outbreaks in particular.


2021 ◽  
Vol 271 ◽  
pp. 01034
Author(s):  
Yushan Min

If the retinal images show evidences of abnormalities such as change in volume, diameter, and unusual spots in the retina, then there is a positive correlation to the diabetic progress. Mathematical and statistical theories behind the machine learning algorithms are powerful enough to detect signs of diabetes through retinal images. Several machine learning algorithms: Logistic Regression, Support Vector Machine, Random Forest, and Neural Networks were applied to predict whether images contain signs of diabetic retinopathy or not. After building the models, the computed results of these algorithms were compared by confusion matrixes, receiver operating characteristic curves, and Precision-Recall curves. The performance of the Support Vector Machine algorithm was the best since it had the highest true-positive rate, area under the curve for ROC curve, and area under the curve for Precision-Recall curve. This conclusion shows that the most complex algorithms doesn’t always give the best performance, the final accuracy also depends on the dataset. For this dataset of retinal imaging, the Support Vector Machine algorithm achieved the best results. Detecting signs of diabetic retinopathy is helpful for detecting for diabetes since more than 60% of patients with diabetes have signs of diabetic retinopathy. Machine learning algorithms can speed up the process and improve the accuracy of diagnosis. When the method is reliable enough, it can be utilized in diabetes diagnosis directly in clinics. Current methods require going on diets and taking blood samples, which could be very time consuming and inconvenient. Using machine learning algorithms is fast and noninvasive compared to the existing methods. The purpose of this research was to build an optimized model by machine learning algorithms that can improve the diagnosis accuracy and classification of patients at high risk of diabetes using retinal imaging.


Sign in / Sign up

Export Citation Format

Share Document