Distributed probability density based multi-objective routing for Opp-IoT networks enabled by machine learning

2021 ◽  
pp. 1-13
Author(s):  
S.P. Ajith Kumar ◽  
Siddhant Banyal ◽  
Kartik Krishna Bharadwaj ◽  
Hardeo Kumar Thakur ◽  
Deepak Kumar Sharma

 Opportunistic IoT networks operate in an intermittent, mobile communication topology, employing peer-to-peer transmission hops on a store-carry-forward basis. Such a network suffers from intermittent connectivity, lack of end-to-end route definition, resource constraints and uncertainties arising from a dynamic topology, given the mobility of participating nodes. Machine learning is an instrumental tool for learning and many histories-based machine learning paradigms like MLPROPH, KNNR and GMMR have been proposed for digital transformations in the field with varying degrees of success. This paper explores the dynamic topology with a plethora of characteristics guiding the node interactions, and consequently, the routing decisions. Further, the study ascertains the need for better representation of the versatility of node characteristics that guide their behavior. The proposed scheme Opportunistic Fuzzy Clustering Routing (OFCR) protocol employs a three-tiered intelligent fuzzy clustering-based paradigm that allows representation of multiple properties of a single entity and the degree of association of the entity with each property group that it is represented by. Such quantification of the extent of association allows OFCR a proper representation of multiple node characteristics, allowing a better judgement for message routing decisions based on these characteristics. OFCR performed 33.77%, 6.07%, 3.69%, 6.88% and 78.14% better than KNNR, GMMR, CAML, MLPRoPH and HBPR respectively across Message Delivery probability. OFCR, not only shows improved performance from the compared protocols but also shows relatively more consistency across the change in simulation time, message TTL and message generation interval across performance metrics.

2013 ◽  
Vol 4 (3) ◽  
pp. 31-46 ◽  
Author(s):  
Pedro Vieira ◽  
António Costa ◽  
Joaquim Macedo

Opportunistic networking differs from conventional architectures in the lack of existing network infrastructure, which can cause intermittent connectivity or increased communication delay between nodes. From a message routing perspective, solving these problems require a different set of techniques than those used in more traditional network schemes. Forwarding algorithms in these scenarios aim to improve performance metrics such as message delivery ratio and message delay time, while trying to keep the number of message copies small. A common approach used for testing the performance of opportunistic protocols relies on existing opportunistic contact traces. These datasets are widely available on the Internet, and provide a convenient way of simulating realistic usage scenarios. As such, studying the contact patterns between nodes can lead to useful observations to take into account in future experiments. This paper presents the results of a study on four different datasets. First, the authors describe the main characteristics of each trace. Then, they propose a graphical representation of the contact behavior for each pair of nodes. Further analysis of the results in terms of connectivity distribution among nodes reveals that contacts follow a roughly lognormal distribution and that there is a small group of nodes in each set which is seemingly much more popular than the rest. Finally, the authors introduce a temporal analysis that was made over the duration of each collection experiment. It was noticeable that individual nodes have repetitive contact patterns over time, apart from some observed cyclic variation over time (namely on weekends). By modeling the data traces as time-varying graphs, a performance decrease was observed with the absence of the most popular nodes.


2021 ◽  
pp. 1-10
Author(s):  
Ahmet Tezcan Tekin ◽  
Tolga Kaya ◽  
Ferhan Cebi

The use of fuzzy logic in machine learning is becoming widespread. In machine learning problems, the data, which have different characteristics, are trained and predicted together. Training the model consisting of data with different characteristics can increase the rate of error in prediction. In this study, we suggest a new approach to assembling prediction with fuzzy clustering. Our approach aims to cluster the data according to their fuzzy membership value and model it with similar characteristics. This approach allows for efficient clustering of objects with more than one cluster characteristic. On the other hand, our approach will enable us to combine boosting type ensemble algorithms, which are various forms of assemblies that are widely used in machine learning due to their excellent success in the literature. We used a mobile game’s customers’ marketing and gameplay data for predicting their customer lifetime value for testing our approach. Customer lifetime value prediction for users is crucial for determining the marketing cost cap for companies. The findings reveal that using a fuzzy method to ensemble the algorithms outperforms implementing the algorithms individually.


2021 ◽  
Vol 10 (4) ◽  
pp. 199
Author(s):  
Francisco M. Bellas Aláez ◽  
Jesus M. Torres Palenzuela ◽  
Evangelos Spyrakos ◽  
Luis González Vilas

This work presents new prediction models based on recent developments in machine learning methods, such as Random Forest (RF) and AdaBoost, and compares them with more classical approaches, i.e., support vector machines (SVMs) and neural networks (NNs). The models predict Pseudo-nitzschia spp. blooms in the Galician Rias Baixas. This work builds on a previous study by the authors (doi.org/10.1016/j.pocean.2014.03.003) but uses an extended database (from 2002 to 2012) and new algorithms. Our results show that RF and AdaBoost provide better prediction results compared to SVMs and NNs, as they show improved performance metrics and a better balance between sensitivity and specificity. Classical machine learning approaches show higher sensitivities, but at a cost of lower specificity and higher percentages of false alarms (lower precision). These results seem to indicate a greater adaptation of new algorithms (RF and AdaBoost) to unbalanced datasets. Our models could be operationally implemented to establish a short-term prediction system.


2021 ◽  
Vol 28 (1) ◽  
pp. e100262
Author(s):  
Mustafa Khanbhai ◽  
Patrick Anyadi ◽  
Joshua Symons ◽  
Kelsey Flott ◽  
Ara Darzi ◽  
...  

ObjectivesUnstructured free-text patient feedback contains rich information, and analysing these data manually would require a lot of personnel resources which are not available in most healthcare organisations.To undertake a systematic review of the literature on the use of natural language processing (NLP) and machine learning (ML) to process and analyse free-text patient experience data.MethodsDatabases were systematically searched to identify articles published between January 2000 and December 2019 examining NLP to analyse free-text patient feedback. Due to the heterogeneous nature of the studies, a narrative synthesis was deemed most appropriate. Data related to the study purpose, corpus, methodology, performance metrics and indicators of quality were recorded.ResultsNineteen articles were included. The majority (80%) of studies applied language analysis techniques on patient feedback from social media sites (unsolicited) followed by structured surveys (solicited). Supervised learning was frequently used (n=9), followed by unsupervised (n=6) and semisupervised (n=3). Comments extracted from social media were analysed using an unsupervised approach, and free-text comments held within structured surveys were analysed using a supervised approach. Reported performance metrics included the precision, recall and F-measure, with support vector machine and Naïve Bayes being the best performing ML classifiers.ConclusionNLP and ML have emerged as an important tool for processing unstructured free text. Both supervised and unsupervised approaches have their role depending on the data source. With the advancement of data analysis tools, these techniques may be useful to healthcare organisations to generate insight from the volumes of unstructured free-text data.


2021 ◽  
Vol 10 (4) ◽  
pp. 58-75
Author(s):  
Vivek Sen Saxena ◽  
Prashant Johri ◽  
Avneesh Kumar

Skin lesion melanoma is the deadliest type of cancer. Artificial intelligence provides the power to classify skin lesions as melanoma and non-melanoma. The proposed system for melanoma detection and classification involves four steps: pre-processing, resizing all the images, removing noise and hair from dermoscopic images; image segmentation, identifying the lesion area; feature extraction, extracting features from segmented lesion and classification; and categorizing lesion as malignant (melanoma) and benign (non-melanoma). Modified GrabCut algorithm is employed to generate skin lesion. Segmented lesions are classified using machine learning algorithms such as SVM, k-NN, ANN, and logistic regression and evaluated on performance metrics like accuracy, sensitivity, and specificity. Results are compared with existing systems and achieved higher similarity index and accuracy.


Author(s):  
Chenxi Huang ◽  
Shu-Xia Li ◽  
César Caraballo ◽  
Frederick A. Masoudi ◽  
John S. Rumsfeld ◽  
...  

Background: New methods such as machine learning techniques have been increasingly used to enhance the performance of risk predictions for clinical decision-making. However, commonly reported performance metrics may not be sufficient to capture the advantages of these newly proposed models for their adoption by health care professionals to improve care. Machine learning models often improve risk estimation for certain subpopulations that may be missed by these metrics. Methods and Results: This article addresses the limitations of commonly reported metrics for performance comparison and proposes additional metrics. Our discussions cover metrics related to overall performance, discrimination, calibration, resolution, reclassification, and model implementation. Models for predicting acute kidney injury after percutaneous coronary intervention are used to illustrate the use of these metrics. Conclusions: We demonstrate that commonly reported metrics may not have sufficient sensitivity to identify improvement of machine learning models and propose the use of a comprehensive list of performance metrics for reporting and comparing clinical risk prediction models.


2018 ◽  
Vol 7 (2.26) ◽  
pp. 25
Author(s):  
E Ramya ◽  
R Gobinath

Data mining plays an important role in analysis of data in modern sensor networks. A sensor network is greatly constrained by the various challenges facing a modern Wireless Sensor Network. This survey paper focuses on basic idea about the algorithms and measurements taken by the Researchers in the area of Wireless Sensor Network with Health Care. This survey also catego-ries various constraints in Wireless Body Area Sensor Networks data and finds the best suitable techniques for analysing the Sensor Data. Due to resource constraints and dynamic topology, the quality of service is facing a challenging issue in Wireless Sensor Networks. In this paper, we review the quality of service parameters with respect to protocols, algorithms and Simulations. 


2018 ◽  
Vol 11 (1) ◽  
pp. 105 ◽  
Author(s):  
Syed Abidi ◽  
Mushtaq Hussain ◽  
Yonglin Xu ◽  
Wu Zhang

Incorporating substantial, sustainable development issues into teaching and learning is the ultimate task of Education for Sustainable Development (ESD). The purpose of our study was to identify the confused students who had failed to master the skill(s) given by the tutors as homework using the Intelligent Tutoring System (ITS). We have focused ASSISTments, an ITS in this study, and scrutinized the skill-builder data using machine learning techniques and methods. We used seven candidate models including: Naïve Bayes (NB), Generalized Linear Model (GLM), Logistic Regression (LR), Deep Learning (DL), Decision Tree (DT), Random Forest (RF), and Gradient Boosted Trees (XGBoost). We trained, validated, and tested learning algorithms, performed stratified cross-validation, and measured the performance of the models through various performance metrics, i.e., ROC (Receiver Operating Characteristic), Accuracy, Precision, Recall, F-Measure, Sensitivity, and Specificity. We found RF, GLM, XGBoost, and DL were high accuracy-achieving classifiers. However, other perceptions such as detecting unexplored features that might be related to the forecasting of outputs can also boost the accuracy of the prediction model. Through machine learning methods, we identified the group of students that were confused when attempting the homework exercise, to help foster their knowledge and talent to play a vital role in environmental development.


2021 ◽  
Vol 35 (1) ◽  
pp. 11-21
Author(s):  
Himani Tyagi ◽  
Rajendra Kumar

IoT is characterized by communication between things (devices) that constantly share data, analyze, and make decisions while connected to the internet. This interconnected architecture is attracting cyber criminals to expose the IoT system to failure. Therefore, it becomes imperative to develop a system that can accurately and automatically detect anomalies and attacks occurring in IoT networks. Therefore, in this paper, an Intrsuion Detection System (IDS) based on extracted novel feature set synthesizing BoT-IoT dataset is developed that can swiftly, accurately and automatically differentiate benign and malicious traffic. Instead of using available feature reduction techniques like PCA that can change the core meaning of variables, a unique feature set consisting of only seven lightweight features is developed that is also IoT specific and attack traffic independent. Also, the results shown in the study demonstrates the effectiveness of fabricated seven features in detecting four wide variety of attacks namely DDoS, DoS, Reconnaissance, and Information Theft. Furthermore, this study also proves the applicability and efficiency of supervised machine learning algorithms (KNN, LR, SVM, MLP, DT, RF) in IoT security. The performance of the proposed system is validated using performance Metrics like accuracy, precision, recall, F-Score and ROC. Though the accuracy of Decision Tree (99.9%) and Randon Forest (99.9%) Classifiers are same but other metrics like training and testing time shows Random Forest comparatively better.


Sign in / Sign up

Export Citation Format

Share Document