Spam Detection on Social Media Using Semantic Convolutional Neural Network

Author(s):  
Gauri Jain ◽  
Manisha Sharma ◽  
Basant Agarwal

This article describes how spam detection in the social media text is becoming increasing important because of the exponential increase in the spam volume over the network. It is challenging, especially in case of text within the limited number of characters. Effective spam detection requires more number of efficient features to be learned. In the current article, the use of a deep learning technology known as a convolutional neural network (CNN) is proposed for spam detection with an added semantic layer on the top of it. The resultant model is known as a semantic convolutional neural network (SCNN). A semantic layer is composed of training the random word vectors with the help of Word2vec to get the semantically enriched word embedding. WordNet and ConceptNet are used to find the word similar to a given word, in case it is missing in the word2vec. The architecture is evaluated on two corpora: SMS Spam dataset (UCI repository) and Twitter dataset (Tweets scrapped from public live tweets). The authors' approach outperforms the-state-of-the-art results with 98.65% accuracy on SMS spam dataset and 94.40% accuracy on Twitter dataset.

Author(s):  
Gauri Jain ◽  
Manisha Sharma ◽  
Basant Agarwal

This article describes how spam detection in the social media text is becoming increasing important because of the exponential increase in the spam volume over the network. It is challenging, especially in case of text within the limited number of characters. Effective spam detection requires more number of efficient features to be learned. In the current article, the use of a deep learning technology known as a convolutional neural network (CNN) is proposed for spam detection with an added semantic layer on the top of it. The resultant model is known as a semantic convolutional neural network (SCNN). A semantic layer is composed of training the random word vectors with the help of Word2vec to get the semantically enriched word embedding. WordNet and ConceptNet are used to find the word similar to a given word, in case it is missing in the word2vec. The architecture is evaluated on two corpora: SMS Spam dataset (UCI repository) and Twitter dataset (Tweets scrapped from public live tweets). The authors' approach outperforms the-state-of-the-art results with 98.65% accuracy on SMS spam dataset and 94.40% accuracy on Twitter dataset.


2021 ◽  
Author(s):  
Mayank Mishra ◽  
Tanupriya Choudhury ◽  
Tanmay Sarkar

Abstract In our work, we look to classify images that make their way into our smartphone devices through various social-media text-messaging platforms. We aim at classifying images into three broad categories: document-based images, quote-based images, and photographs. People, especially students, share many document-based images that include snapshots of essential emails, handwritten notes, articles, etc. Quote based images, consisting of birthday wishes, motivational messages, festival greetings, etc., are among the highly shared images on social media platforms. A significant share of images constitutes photographs of people, including group photographs, selfies, portraits, etc. We train various convolutional neural network (CNN) based models on our self-made dataset and compare their results to find our task’s optimum model.


Author(s):  
Santosh Kumar Bharti ◽  
Sathya Babu Korra

Posting sarcastic messages on social media like Twitter, Facebook, WhatsApp, etc., became a new trend to avoid direct negativity. Detecting this indirect negativity in the social media text has become an important task as they influence every business organization. In the presence of sarcasm, detection of actual sentiment on these texts has become the most challenging task. An automated system is required that will be capable of identifying actual sentiment of a given text in the presence of sarcasm. In this chapter, we proposed an automated system for sarcasm detection in social media text using six algorithms that are capable to analyze the various types of sarcasm occurs in Twitter data. These algorithms use lexical, pragmatic, hyperbolic and contextual features of text to identify sarcasm. In the contextual feature, we mainly focus on situation, topical, temporal, and historical context of the text. The experimental results of proposed approach were compared with state-of-the-art techniques.


2019 ◽  
Vol 8 (2S11) ◽  
pp. 3464-3468

Psychological stress which is a mental illness also causes physical problems to the human. Nowadays social media plays an important role in the world for communication to share their thoughts with their friends and family. The social media analysis is the process of detecting and predicting the user's thoughts and opinions which also one of the important perspective in the developing business environment. The overwhelming stress and long term stress sometimes lead to suicidal ideation. By analyzing the social media content to predict the overwhelming stress state of the users in the earlier stage will reduce the psychological stress and suicidal rate too. In this paper, we address the problem of stress prediction by using social media. The machine learning and deep learning methods to perform the classification of stress analysis. Here both image and text- tweet data are used and the images are processed with the Optical Character Recognition and the text data are processed by using the Natural Language Processing and Convolutional Neural Network for classifying the tweet content of the user as stressed or non-stressed. Furthermore, with the advancement of the machine learning and deep learning method of classification gives a better result in terms of performance and accuracy of the prediction.


2017 ◽  
Vol 2017 ◽  
pp. 1-11 ◽  
Author(s):  
Siqi Tang ◽  
Zhisong Pan ◽  
Xingyu Zhou

This paper proposes an accurate crowd counting method based on convolutional neural network and low-rank and sparse structure. To this end, we firstly propose an effective deep-fusion convolutional neural network to promote the density map regression accuracy. Furthermore, we figure out that most of the existing CNN based crowd counting methods obtain overall counting by direct integral of estimated density map, which limits the accuracy of counting. Instead of direct integral, we adopt a regression method based on low-rank and sparse penalty to promote accuracy of the projection from density map to global counting. Experiments demonstrate the importance of such regression process on promoting the crowd counting performance. The proposed low-rank and sparse based deep-fusion convolutional neural network (LFCNN) outperforms existing crowd counting methods and achieves the state-of-the-art performance.


Author(s):  
Yuheng Hu ◽  
Yili Hong

Residents often rely on newspapers and television to gather hyperlocal news for community awareness and engagement. More recently, social media have emerged as an increasingly important source of hyperlocal news. Thus far, the literature on using social media to create desirable societal benefits, such as civic awareness and engagement, is still in its infancy. One key challenge in this research stream is to timely and accurately distill information from noisy social media data streams to community members. In this work, we develop SHEDR (social media–based hyperlocal event detection and recommendation), an end-to-end neural event detection and recommendation framework with a particular use case for Twitter to facilitate residents’ information seeking of hyperlocal events. The key model innovation in SHEDR lies in the design of the hyperlocal event detector and the event recommender. First, we harness the power of two popular deep neural network models, the convolutional neural network (CNN) and long short-term memory (LSTM), in a novel joint CNN-LSTM model to characterize spatiotemporal dependencies for capturing unusualness in a region of interest, which is classified as a hyperlocal event. Next, we develop a neural pairwise ranking algorithm for recommending detected hyperlocal events to residents based on their interests. To alleviate the sparsity issue and improve personalization, our algorithm incorporates several types of contextual information covering topic, social, and geographical proximities. We perform comprehensive evaluations based on two large-scale data sets comprising geotagged tweets covering Seattle and Chicago. We demonstrate the effectiveness of our framework in comparison with several state-of-the-art approaches. We show that our hyperlocal event detection and recommendation models consistently and significantly outperform other approaches in terms of precision, recall, and F-1 scores. Summary of Contribution: In this paper, we focus on a novel and important, yet largely underexplored application of computing—how to improve civic engagement in local neighborhoods via local news sharing and consumption based on social media feeds. To address this question, we propose two new computational and data-driven methods: (1) a deep learning–based hyperlocal event detection algorithm that scans spatially and temporally to detect hyperlocal events from geotagged Twitter feeds; and (2) A personalized deep learning–based hyperlocal event recommender system that systematically integrates several contextual cues such as topical, geographical, and social proximity to recommend the detected hyperlocal events to potential users. We conduct a series of experiments to examine our proposed models. The outcomes demonstrate that our algorithms are significantly better than the state-of-the-art models and can provide users with more relevant information about the local neighborhoods that they live in, which in turn may boost their community engagement.


Sign in / Sign up

Export Citation Format

Share Document