Deep learning based affective computing

2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Saurabh Kumar

PurposeDecision-making in human beings is affected by emotions and sentiments. The affective computing takes this into account, intending to tailor decision support to the emotional states of people. However, the representation and classification of emotions is a very challenging task. The study used customized methods of deep learning models to aid in the accurate classification of emotions and sentiments.Design/methodology/approachThe present study presents affective computing model using both text and image data. The text-based affective computing was conducted on four standard datasets using three deep learning customized models, namely LSTM, GRU and CNN. The study used four variants of deep learning including the LSTM model, LSTM model with GloVe embeddings, Bi-directional LSTM model and LSTM model with attention layer.FindingsThe result suggests that the proposed method outperforms the earlier methods. For image-based affective computing, the data was extracted from Instagram, and Facial emotion recognition was carried out using three deep learning models, namely CNN, transfer learning with VGG-19 model and transfer learning with ResNet-18 model. The results suggest that the proposed methods for both text and image can be used for affective computing and aid in decision-making.Originality/valueThe study used deep learning for affective computing. Earlier studies have used machine learning algorithms for affective computing. However, the present study uses deep learning for affective computing.

2019 ◽  
Author(s):  
Ismael Araujo ◽  
Juan Gamboa ◽  
Adenilton Silva

To recognize patterns that are usually imperceptible by human beings has been one of the main advantages of using machine learning algorithms The use of Deep Learning techniques has been promising to the classification problems, especially the ones related to image classification. The classification of gases detected by an artificial nose is one other area where Deep Learning techniques can be used to seek classification improvements. Succeeding in a classification task can result in many advantages to quality control, as well as to preventing accidents. In this work, it is presented some Deep Learning models specifically created to the task of gas classification.


Electronics ◽  
2020 ◽  
Vol 9 (3) ◽  
pp. 427 ◽  
Author(s):  
Laith Alzubaidi ◽  
Mohammed A. Fadhel ◽  
Omran Al-Shamma ◽  
Jinglan Zhang ◽  
Ye Duan

Sickle cell anemia, which is also called sickle cell disease (SCD), is a hematological disorder that causes occlusion in blood vessels, leading to hurtful episodes and even death. The key function of red blood cells (erythrocytes) is to supply all the parts of the human body with oxygen. Red blood cells (RBCs) form a crescent or sickle shape when sickle cell anemia affects them. This abnormal shape makes it difficult for sickle cells to move through the bloodstream, hence decreasing the oxygen flow. The precise classification of RBCs is the first step toward accurate diagnosis, which aids in evaluating the danger level of sickle cell anemia. The manual classification methods of erythrocytes require immense time, and it is possible that errors may be made throughout the classification stage. Traditional computer-aided techniques, which have been employed for erythrocyte classification, are based on handcrafted features techniques, and their performance relies on the selected features. They also are very sensitive to different sizes, colors, and complex shapes. However, microscopy images of erythrocytes are very complex in shape with different sizes. To this end, this research proposes lightweight deep learning models that classify the erythrocytes into three classes: circular (normal), elongated (sickle cells), and other blood content. These models are different in the number of layers and learnable filters. The available datasets of red blood cells with sickle cell disease are very small for training deep learning models. Therefore, addressing the lack of training data is the main aim of this paper. To tackle this issue and optimize the performance, the transfer learning technique is utilized. Transfer learning does not significantly affect performance on medical image tasks when the source domain is completely different from the target domain. In some cases, it can degrade the performance. Hence, we have applied the same domain transfer learning, unlike other methods that used the ImageNet dataset for transfer learning. To minimize the overfitting effect, we have utilized several data augmentation techniques. Our model obtained state-of-the-art performance and outperformed the latest methods by achieving an accuracy of 99.54% with our model and 99.98% with our model plus a multiclass SVM classifier on the erythrocytesIDB dataset and 98.87% on the collected dataset.


Medical imaging plays an important role in the diagnosis of some critical diseases and further treatment process of patients. Brain is a central and most complex structure in the human body that works with billions of cells, which controls all other organ functioning. Brain tumours observed as uncontrolled abnormal cell growth in brain tissues. Classification of such cells in a early stage will increase the survival rate of the patient. Machine learning algorithms have contributed much in automation of such tasks. Further improvement in prediction rate is possible through deep learning models. In this paper presents experiments by deep transfer learning models on publicly available dataset for Brain tumour classification. Pre-trained plain and residual feed forward models such as Alexnet, VGG19, ResNet50, ResNet101 and GoogleNet are used for the purpose of feature extraction, Fully connected layers and softmax layer for classification is used commonly. The evaluation metrics Accuracy, Sensitivity, Specificity and F1-Score were computed.


2021 ◽  
Vol 11 (16) ◽  
pp. 7561
Author(s):  
Umair Iqbal ◽  
Johan Barthelemy ◽  
Wanqing Li ◽  
Pascal Perez

Blockage of culverts by transported debris materials is reported as the salient contributor in originating urban flash floods. Conventional hydraulic modeling approaches had no success in addressing the problem primarily because of the unavailability of peak floods hydraulic data and the highly non-linear behavior of debris at the culvert. This article explores a new dimension to investigate the issue by proposing the use of intelligent video analytics (IVA) algorithms for extracting blockage related information. The presented research aims to automate the process of manual visual blockage classification of culverts from a maintenance perspective by remotely applying deep learning models. The potential of using existing convolutional neural network (CNN) algorithms (i.e., DarkNet53, DenseNet121, InceptionResNetV2, InceptionV3, MobileNet, ResNet50, VGG16, EfficientNetB3, NASNet) is investigated over a dataset from three different sources (i.e., images of culvert openings and blockage (ICOB), visual hydrology-lab dataset (VHD), synthetic images of culverts (SIC)) to predict the blockage in a given image. Models were evaluated based on their performance on the test dataset (i.e., accuracy, loss, precision, recall, F1 score, Jaccard Index, region of convergence (ROC) curve), floating point operations per second (FLOPs) and response times to process a single test instance. Furthermore, the performance of deep learning models was benchmarked against conventional machine learning algorithms (i.e., SVM, RF, xgboost). In addition, the idea of classifying deep visual features extracted by CNN models (i.e., ResNet50, MobileNet) using conventional machine learning approaches was also implemented in this article. From the results, NASNet was reported most efficient in classifying the blockage images with the 5-fold accuracy of 85%; however, MobileNet was recommended for the hardware implementation because of its improved response time with 5-fold accuracy comparable to NASNet (i.e., 78%). Comparable performance to standard CNN models was achieved for the case where deep visual features were classified using conventional machine learning approaches. False negative (FN) instances, false positive (FP) instances and CNN layers activation suggested that background noise and oversimplified labelling criteria were two contributing factors in the degraded performance of existing CNN algorithms. A framework for partial automation of the visual blockage classification process was proposed, given that none of the existing models was able to achieve high enough accuracy to completely automate the manual process. In addition, a detection-classification pipeline with higher blockage classification accuracy (i.e., 94%) has been proposed as a potential future direction for practical implementation.


2021 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Priyanka Yadlapalli ◽  
D. Bhavana ◽  
Suryanarayana Gunnam

PurposeComputed tomography (CT) scan can provide valuable information in the diagnosis of lung diseases. To detect the location of the cancerous lung nodules, this work uses novel deep learning methods. The majority of the early investigations used CT, magnetic resonance and mammography imaging. Using appropriate procedures, the professional doctor in this sector analyses these images to discover and diagnose the various degrees of lung cancer. All of the methods used to discover and detect cancer illnesses are time-consuming, expensive and stressful for the patients. To address all of these issues, appropriate deep learning approaches for analyzing these medical images, which included CT scan images, were utilized.Design/methodology/approachRadiologists currently employ chest CT scans to detect lung cancer at an early stage. In certain situations, radiologists' perception plays a critical role in identifying lung melanoma which is incorrectly detected. Deep learning is a new, capable and influential approach for predicting medical images. In this paper, the authors employed deep transfer learning algorithms for intelligent classification of lung nodules. Convolutional neural networks (VGG16, VGG19, MobileNet and DenseNet169) are used to constrain the input and output layers of a chest CT scan image dataset.FindingsThe collection includes normal chest CT scan pictures as well as images from two kinds of lung cancer, squamous and adenocarcinoma impacted chest CT scan images. According to the confusion matrix results, the VGG16 transfer learning technique has the highest accuracy in lung cancer classification with 91.28% accuracy, followed by VGG19 with 89.39%, MobileNet with 85.60% and DenseNet169 with 83.71% accuracy, which is analyzed using Google Collaborator.Originality/valueThe proposed approach using VGG16 maximizes the classification accuracy when compared to VGG19, MobileNet and DenseNet169. The results are validated by computing the confusion matrix for each network type.


2020 ◽  
Vol 34 (5) ◽  
pp. 617-622
Author(s):  
Sai Sudha Sonali Palakodati ◽  
Venkata RamiReddy Chirra ◽  
Yakobu Dasari ◽  
Suneetha Bulla

Detecting the rotten fruits become significant in the agricultural industry. Usually, the classification of fresh and rotten fruits is carried by humans is not effectual for the fruit farmers. Human beings will become tired after doing the same task multiple times, but machines do not. Thus, the project proposes an approach to reduce human efforts, reduce the cost and time for production by identifying the defects in the fruits in the agricultural industry. If we do not detect those defects, those defected fruits may contaminate good fruits. Hence, we proposed a model to avoid the spread of rottenness. The proposed model classifies the fresh fruits and rotten fruits from the input fruit images. In this work, we have used three types of fruits, such as apple, banana, and oranges. A Convolutional Neural Network (CNN) is used for extracting the features from input fruit images, and Softmax is used to classify the images into fresh and rotten fruits. The performance of the proposed model is evaluated on a dataset that is downloaded from Kaggle and produces an accuracy of 97.82%. The results showed that the proposed CNN model can effectively classify the fresh fruits and rotten fruits. In the proposed work, we inspected the transfer learning methods in the classification of fresh and rotten fruits. The performance of the proposed CNN model outperforms the transfer learning models and the state of art methods.


Author(s):  
Matt Ervin Mital ◽  
Rogelio Ruzcko Tobias ◽  
Herbert Villaruel ◽  
Jose Martin Maningo ◽  
Robert Kerwin Billones ◽  
...  

Author(s):  
Yuejun Liu ◽  
Yifei Xu ◽  
Xiangzheng Meng ◽  
Xuguang Wang ◽  
Tianxu Bai

Background: Medical imaging plays an important role in the diagnosis of thyroid diseases. In the field of machine learning, multiple dimensional deep learning algorithms are widely used in image classification and recognition, and have achieved great success. Objective: The method based on multiple dimensional deep learning is employed for the auxiliary diagnosis of thyroid diseases based on SPECT images. The performances of different deep learning models are evaluated and compared. Methods: Thyroid SPECT images are collected with three types, they are hyperthyroidism, normal and hypothyroidism. In the pre-processing, the region of interest of thyroid is segmented and the amount of data sample is expanded. Four CNN models, including CNN, Inception, VGG16 and RNN, are used to evaluate deep learning methods. Results: Deep learning based methods have good classification performance, the accuracy is 92.9%-96.2%, AUC is 97.8%-99.6%. VGG16 model has the best performance, the accuracy is 96.2% and AUC is 99.6%. Especially, the VGG16 model with a changing learning rate works best. Conclusion: The standard CNN, Inception, VGG16, and RNN four deep learning models are efficient for the classification of thyroid diseases with SPECT images. The accuracy of the assisted diagnostic method based on deep learning is higher than that of other methods reported in the literature.


Sign in / Sign up

Export Citation Format

Share Document