scholarly journals Link Prediction in Facebook using Web Scrapping and Deep Learning Techniques

As Internet technologies develop continuously social networks are getting more popular day by day. People are connected with each other via virtual applications. Using the Link Prediction in social networks more people get connected, may be they are friends, may be work together at the same workplace and may be their education are. Machine learning techniques are used to analyze the link between the nodes of the network and also create a better link prediction model through deep learning. The objective of this research is to measure the performance using the different techniques to predict link between the social networks. Using deep learning, feature engineering can be reduced for link prediction. In this research, the feature based learning is used to predict the link for better performance. Dataset is obtained by scraping the profile of Facebook users and they are used along with the random forest and graph convolution neural network to measure the performance of link prediction in social networks.

Author(s):  
Putra Wanda ◽  
Marselina Endah Hiswati ◽  
Huang J. Jie

Manual analysis for malicious prediction in Online Social Networks (OSN) is time-consuming and costly. With growing users within the environment, it becomes one of the main obstacles. Deep learning is growing algorithm that gains a big success in computer vision problem. Currently, many research communities have proposed deep learning techniques to automate security tasks, including anomalous detection, malicious link prediction, and intrusion detection in OSN. Notably, this article describes how deep learning makes the OSN security technique more intelligent for detecting malicious activity by establishing a classifier model.


2021 ◽  
Vol 9 (1) ◽  
pp. 1315-1320
Author(s):  
Dr. Mohammed Ali Alhariri

The duplicate fake accounts are detected in this work the data from the social media platform is accessed. The platform choose to use the analysis on social media platform is selected as twitter. The twitter data is accessed using Twitter API, with using some selected features that remain the most appropriate regarding the reason of duplicate fake account. The feature based analysis is compared using machine learning techniques, Random Forest, Decision Tree, and SVM. The performance is further analyzed based on accuracy SVM performed 93.3% accuracy, where decision tree performed as 89.0% and random forest performed as 85.5%. The better performance observed using feature-based analysis is of SVM.  


More and more individuals are now using online social networks and resources throughout this day and age to not only interact and to communicate but also for sharing their views, experiences, ideas, impression about anything. The analysis of sentiments is the identification and categorization of these views to evaluate public opinions on a specific subject, question, product, etc. Day by day, the relevance of sentiment analysis is growing up. Machine learning is an area or field of computer science where, without being specifically programmed, computers can learn. Deep learning is the part of machine learning and deals with the algorithm, which is most widely used as Neural network, neural belief, etc., in which neuronal implementations are considered. For sentiment analysis, it compares their performance and accuracy so then it can be inferred that deep learning techniques in most of the cases provide better results. The gap in the precision of these two approaches, however, is not as important enough in certain situations, and so it is best to apply and use the machine learning approaches and methods because these are simpler in terms of Implementation


2019 ◽  
Vol 2019 (3) ◽  
pp. 191-209 ◽  
Author(s):  
Se Eun Oh ◽  
Saikrishna Sunkam ◽  
Nicholas Hopper

Abstract Recent advances in Deep Neural Network (DNN) architectures have received a great deal of attention due to their ability to outperform state-of-the-art machine learning techniques across a wide range of application, as well as automating the feature engineering process. In this paper, we broadly study the applicability of deep learning to website fingerprinting. First, we show that unsupervised DNNs can generate lowdimensional informative features that improve the performance of state-of-the-art website fingerprinting attacks. Second, when used as classifiers, we show that they can exceed performance of existing attacks across a range of application scenarios, including fingerprinting Tor website traces, fingerprinting search engine queries over Tor, defeating fingerprinting defenses, and fingerprinting TLS-encrypted websites. Finally, we investigate which site-level features of a website influence its fingerprintability by DNNs.


2019 ◽  
pp. 016555151989134 ◽  
Author(s):  
Mohammad Mehdi Keikha ◽  
Maseud Rahgozar ◽  
Masoud Asadpour

Recently, link prediction has attracted more attention from various disciplines such as computer science, bioinformatics and economics. In link prediction, numerous information such as network topology, profile information and user-generated contents are considered to discover missing links between nodes. Whereas numerous previous researches had focused on the structural features of the networks for link prediction, recent studies have shown more interest in profile and content information, too. So, some of these researches combine structural and content information. However, some issues such as scalability and feature engineering need to be investigated to solve a few remaining problems. Moreover, most of the previous researches are presented only for undirected and unweighted networks. In this article, a novel link prediction framework named ‘DeepLink’ is presented, which is based on deep learning techniques. While deep learning has the advantage of extracting automatically the best features for link prediction, many other link prediction algorithms need manual feature engineering. Moreover, in the proposed framework, both structural and content information are employed. The framework is capable of using different structural feature vectors that are prepared by various link prediction methods. It learns all proximity orders that are presented on a network during the structural feature learning. We have evaluated the effectiveness of DeepLink on two real social network datasets, Telegram and irBlogs. On both datasets, the proposed framework outperforms several other structural and hybrid approaches for link prediction.


Mathematics ◽  
2020 ◽  
Vol 8 (12) ◽  
pp. 2258
Author(s):  
Madhab Raj Joshi ◽  
Lewis Nkenyereye ◽  
Gyanendra Prasad Joshi ◽  
S. M. Riazul Islam ◽  
Mohammad Abdullah-Al-Wadud ◽  
...  

Enhancement of Cultural Heritage such as historical images is very crucial to safeguard the diversity of cultures. Automated colorization of black and white images has been subject to extensive research through computer vision and machine learning techniques. Our research addresses the problem of generating a plausible colored photograph of ancient, historically black, and white images of Nepal using deep learning techniques without direct human intervention. Motivated by the recent success of deep learning techniques in image processing, a feed-forward, deep Convolutional Neural Network (CNN) in combination with Inception- ResnetV2 is being trained by sets of sample images using back-propagation to recognize the pattern in RGB and grayscale values. The trained neural network is then used to predict two a* and b* chroma channels given grayscale, L channel of test images. CNN vividly colorizes images with the help of the fusion layer accounting for local features as well as global features. Two objective functions, namely, Mean Squared Error (MSE) and Peak Signal-to-Noise Ratio (PSNR), are employed for objective quality assessment between the estimated color image and its ground truth. The model is trained on the dataset created by ourselves with 1.2 K historical images comprised of old and ancient photographs of Nepal, each having 256 × 256 resolution. The loss i.e., MSE, PSNR, and accuracy of the model are found to be 6.08%, 34.65 dB, and 75.23%, respectively. Other than presenting the training results, the public acceptance or subjective validation of the generated images is assessed by means of a user study where the model shows 41.71% of naturalness while evaluating colorization results.


Vibration ◽  
2021 ◽  
Vol 4 (2) ◽  
pp. 341-356
Author(s):  
Jessada Sresakoolchai ◽  
Sakdirat Kaewunruen

Various techniques have been developed to detect railway defects. One of the popular techniques is machine learning. This unprecedented study applies deep learning, which is a branch of machine learning techniques, to detect and evaluate the severity of rail combined defects. The combined defects in the study are settlement and dipped joint. Features used to detect and evaluate the severity of combined defects are axle box accelerations simulated using a verified rolling stock dynamic behavior simulation called D-Track. A total of 1650 simulations are run to generate numerical data. Deep learning techniques used in the study are deep neural network (DNN), convolutional neural network (CNN), and recurrent neural network (RNN). Simulated data are used in two ways: simplified data and raw data. Simplified data are used to develop the DNN model, while raw data are used to develop the CNN and RNN model. For simplified data, features are extracted from raw data, which are the weight of rolling stock, the speed of rolling stock, and three peak and bottom accelerations from two wheels of rolling stock. In total, there are 14 features used as simplified data for developing the DNN model. For raw data, time-domain accelerations are used directly to develop the CNN and RNN models without processing and data extraction. Hyperparameter tuning is performed to ensure that the performance of each model is optimized. Grid search is used for performing hyperparameter tuning. To detect the combined defects, the study proposes two approaches. The first approach uses one model to detect settlement and dipped joint, and the second approach uses two models to detect settlement and dipped joint separately. The results show that the CNN models of both approaches provide the same accuracy of 99%, so one model is good enough to detect settlement and dipped joint. To evaluate the severity of the combined defects, the study applies classification and regression concepts. Classification is used to evaluate the severity by categorizing defects into light, medium, and severe classes, and regression is used to estimate the size of defects. From the study, the CNN model is suitable for evaluating dipped joint severity with an accuracy of 84% and mean absolute error (MAE) of 1.25 mm, and the RNN model is suitable for evaluating settlement severity with an accuracy of 99% and mean absolute error (MAE) of 1.58 mm.


2021 ◽  
pp. 1-55
Author(s):  
Emma A. H. Michie ◽  
Behzad Alaei ◽  
Alvar Braathen

Generating an accurate model of the subsurface for the purpose of assessing the feasibility of a CO2 storage site is crucial. In particular, how faults are interpreted is likely to influence the predicted capacity and integrity of the reservoir; whether this is through identifying high risk areas along the fault, where fluid is likely to flow across the fault, or by assessing the reactivation potential of the fault with increased pressure, causing fluid to flow up the fault. New technologies allow users to interpret faults effortlessly, and in much quicker time, utilizing methods such as Deep Learning. These Deep Learning techniques use knowledge from Neural Networks to allow end-users to compute areas where faults are likely to occur. Although these new technologies may be attractive due to reduced interpretation time, it is important to understand the inherent uncertainties in their ability to predict accurate fault geometries. Here, we compare Deep Learning fault interpretation versus manual fault interpretation, and can see distinct differences to those faults where significant ambiguity exists due to poor seismic resolution at the fault; we observe an increased irregularity when Deep Learning methods are used over conventional manual interpretation. This can result in significant differences between the resulting analyses, such as fault reactivation potential. Conversely, we observe that well-imaged faults show a close similarity between the resulting fault surfaces when both Deep Learning and manual fault interpretation methods are employed, and hence we also observe a close similarity between any attributes and fault analyses made.


2021 ◽  
Author(s):  
Rogini Runghen ◽  
Daniel B Stouffer ◽  
Giulio Valentino Dalla Riva

Collecting network interaction data is difficult. Non-exhaustive sampling and complex hidden processes often result in an incomplete data set. Thus, identifying potentially present but unobserved interactions is crucial both in understanding the structure of large scale data, and in predicting how previously unseen elements will interact. Recent studies in network analysis have shown that accounting for metadata (such as node attributes) can improve both our understanding of how nodes interact with one another, and the accuracy of link prediction. However, the dimension of the object we need to learn to predict interactions in a network grows quickly with the number of nodes. Therefore, it becomes computationally and conceptually challenging for large networks. Here, we present a new predictive procedure combining a graph embedding method with machine learning techniques to predict interactions on the base of nodes' metadata. Graph embedding methods project the nodes of a network onto a---low dimensional---latent feature space. The position of the nodes in the latent feature space can then be used to predict interactions between nodes. Learning a mapping of the nodes' metadata to their position in a latent feature space corresponds to a classic---and low dimensional---machine learning problem. In our current study we used the Random Dot Product Graph model to estimate the embedding of an observed network, and we tested different neural networks architectures to predict the position of nodes in the latent feature space. Flexible machine learning techniques to map the nodes onto their latent positions allow to account for multivariate and possibly complex nodes' metadata. To illustrate the utility of the proposed procedure, we apply it to a large dataset of tourist visits to destinations across New Zealand. We found that our procedure accurately predicts interactions for both existing nodes and nodes newly added to the network, while being computationally feasible even for very large networks. Overall, our study highlights that by exploiting the properties of a well understood statistical model for complex networks and combining it with standard machine learning techniques, we can simplify the link prediction problem when incorporating multivariate node metadata. Our procedure can be immediately applied to different types of networks, and to a wide variety of data from different systems. As such, both from a network science and data science perspective, our work offers a flexible and generalisable procedure for link prediction.


Sign in / Sign up

Export Citation Format

Share Document