scholarly journals Semantic Scene Graph Generation Using RDF Model and Deep Learning

2021 ◽  
Vol 11 (2) ◽  
pp. 826
Author(s):  
Seongyong Kim ◽  
Tae Hyeon Jeon ◽  
Ilsun Rhiu ◽  
Jinhyun Ahn ◽  
Dong-Hyuk Im

Over the last several years, in parallel with the general global advancement in mobile technology and a rise in social media network content consumption, multimedia content production and reproduction has increased exponentially. Therefore, enabled by the rapid recent advancements in deep learning technology, research on scene graph generation is being actively conducted to more efficiently search for and classify images desired by users within a large amount of content. This approach lets users accurately find images they are searching for by expressing meaningful information on image content as nodes and edges of a graph. In this study, we propose a scene graph generation method based on using the Resource Description Framework (RDF) model to clarify semantic relations. Furthermore, we also use convolutional neural network (CNN) and recurrent neural network (RNN) deep learning models to generate a scene graph expressed in a controlled vocabulary of the RDF model to understand the relations between image object tags. Finally, we experimentally demonstrate through testing that our proposed technique can express semantic content more effectively than existing approaches.

2021 ◽  
Vol 13 (11) ◽  
pp. 168781402110622
Author(s):  
Yi-Ren Wang ◽  
Yi-Jyun Wang

Deep learning technology has been widely used in various field in recent years. This study intends to use deep learning algorithms to analyze the aeroelastic phenomenon and compare the differences between Deep Neural Network (DNN) and Long Short-term Memory (LSTM) applied on the flutter speed prediction. In this present work, DNN and LSTM are used to address complex aeroelastic systems by superimposing multi-layer Artificial Neural Network. Under such an architecture, the neurons in neural network can extract features from various flight data. Instead of time-consuming high-fidelity computational fluid dynamics (CFD) method, this study uses the K method to build the aeroelastic flutter speed big data for different flight conditions. The flutter speeds for various flight conditions are predicted by the deep learning methods and verified by the K method. The detailed physical meaning of aerodynamics and aeroelasticity of the prediction results are studied. The LSTM model has a cyclic architecture, which enables it to store information and update it with the latest information at the same time. Although the training of the model is more time-consuming than DNN, this method can increase the memory space. The results of this work show that the LSTM model established in this study can provide more accurate flutter speed prediction than the DNN algorithm.


2020 ◽  
Vol 29 (07n08) ◽  
pp. 2040005
Author(s):  
Zhen Li ◽  
Dan Qu ◽  
Yanxia Li ◽  
Chaojie Xie ◽  
Qi Chen

Deep learning technology promotes the development of neural network machine translation (NMT). End-to-End (E2E) has become the mainstream in NMT. It uses word vectors as the initial value of the input layer. The effect of word vector model directly affects the accuracy of E2E-NMT. Researchers have proposed many approaches to learn word representations and have achieved significant results. However, the drawbacks of these methods still limit the performance of E2E-NMT systems. This paper focuses on the word embedding technology and proposes the PW-CBOW word vector model which can present better semantic information. We apply these word vector models on IWSLT14 German-English, WMT14 English-German, WMT14 English-French corporas. The results evaluate the performance of the PW-CBOW model. In the latest E2E-NMT systems, the PW-CBOW word vector model can improve the performance.


Diagnostics ◽  
2019 ◽  
Vol 9 (4) ◽  
pp. 207 ◽  
Author(s):  
Dana Li ◽  
Bolette Mikela Vilmun ◽  
Jonathan Frederik Carlsen ◽  
Elisabeth Albrecht-Beste ◽  
Carsten Ammitzbøl Lauridsen ◽  
...  

The aim of this study was to systematically review the performance of deep learning technology in detecting and classifying pulmonary nodules on computed tomography (CT) scans that were not from the Lung Image Database Consortium and Image Database Resource Initiative (LIDC-IDRI) database. Furthermore, we explored the difference in performance when the deep learning technology was applied to test datasets different from the training datasets. Only peer-reviewed, original research articles utilizing deep learning technology were included in this study, and only results from testing on datasets other than the LIDC-IDRI were included. We searched a total of six databases: EMBASE, PubMed, Cochrane Library, the Institute of Electrical and Electronics Engineers, Inc. (IEEE), Scopus, and Web of Science. This resulted in 1782 studies after duplicates were removed, and a total of 26 studies were included in this systematic review. Three studies explored the performance of pulmonary nodule detection only, 16 studies explored the performance of pulmonary nodule classification only, and 7 studies had reports of both pulmonary nodule detection and classification. Three different deep learning architectures were mentioned amongst the included studies: convolutional neural network (CNN), massive training artificial neural network (MTANN), and deep stacked denoising autoencoder extreme learning machine (SDAE-ELM). The studies reached a classification accuracy between 68–99.6% and a detection accuracy between 80.6–94%. Performance of deep learning technology in studies using different test and training datasets was comparable to studies using same type of test and training datasets. In conclusion, deep learning was able to achieve high levels of accuracy, sensitivity, and/or specificity in detecting and/or classifying nodules when applied to pulmonary CT scans not from the LIDC-IDRI database.


2019 ◽  
Vol 8 (3) ◽  
pp. 8619-8622

People, due to their complexity and volatile actions, are constantly faced with challenges in understanding the situation in the market share and the forecast for the future. For any financial investment, the stock market is a very important aspect. It is necessary to study while understanding the price fluctuations of the stock market. In this paper, the stock market prediction model using the Recurrent Digital natural Network (RDNN) is described. The model is designed using two important machine learning concepts: the recurrent neural network (RNN), multilayer perceptron (MLP) and reinforcement learning (RL). Deep learning is used to automatically extract important functions of the stock market; reinforcement learning of these functions will be useful for future prediction of the stock market, the system uses historical stock market data to understand the dynamic market behavior when you make decisions in an unknown environment. In this paper, the understanding of the dynamic stock market and the deep learning technology for predicting the price of the future stock market are described.


Author(s):  
Shravani Kharat ◽  
Pooja Shinde ◽  
Preeti Malwadkar ◽  
Dipti Chaudhari

Globally, skin diseases are among the most common health problems in all humans irrespective of age. Prevention and early detection of these diseases can improve the chance of surviving. This model illustrates the identification of skin diseases providing more objective and reliable solutions using deep learning technology and convolutional neural network approach. In this model, the system design, implementation and identification of common skin diseases such as acne, blister, eczema, warts etc. are explained. The system applies deep learning technology to train itself with various images of skin diseases from the Kaggle platform. The accuracy obtained by using deep learning is 83.23%. The main objective of this system is to achieve maximum accuracy of skin disease prediction. Moreover, if the disease is identified the system provides detailed information about the diseases along with home remedies.


Energies ◽  
2020 ◽  
Vol 13 (24) ◽  
pp. 6536
Author(s):  
Chuanhui Li ◽  
Xuewei Liu

Gas hydrate saturation is an important index for evaluating gas hydrate reservoirs, and well logs are an effective method for estimating gas hydrate saturation. To use well logs better to estimate gas hydrate saturation, and to establish the deep internal connections and laws of the data, we propose a method of using deep learning technology to estimate gas hydrate saturation from well logs. Considering that well logs have sequential characteristics, we used the long short-term memory (LSTM) recurrent neural network to predict the gas hydrate saturation from the well logs of two sites in the Shenhu area, South China Sea. By constructing an LSTM recurrent layer and two fully connected layers at one site, we used resistivity and acoustic velocity logs that were sensitive to gas hydrate as input. We used the gas hydrate saturation calculated by the chloride concentration of the pore water as output to train the LSTM network. We achieved a good training result. Applying the trained LSTM recurrent neural network to another site in the same area achieved good prediction of gas hydrate saturation, showing the unique advantages of deep learning technology in gas hydrate saturation estimation.


Author(s):  
Hoseok Choi ◽  
Seokbeen Lim ◽  
Kyeongran Min ◽  
Kyoung-ha Ahn ◽  
Kyoung-Min Lee ◽  
...  

Abstract Objective: With the development in the field of neural networks, Explainable AI (XAI), is being studied to ensure that artificial intelligence models can be explained. There are some attempts to apply neural networks to neuroscientific studies to explain neurophysiological information with high machine learning performances. However, most of those studies have simply visualized features extracted from XAI and seem to lack an active neuroscientific interpretation of those features. In this study, we have tried to actively explain the high-dimensional learning features contained in the neurophysiological information extracted from XAI, compared with the previously reported neuroscientific results. Approach: We designed a deep neural network classifier using 3D information (3D DNN) and a 3D class activation map (3D CAM) to visualize high-dimensional classification features. We used those tools to classify monkey electrocorticogram (ECoG) data obtained from the unimanual and bimanual movement experiment. Main results: The 3D DNN showed better classification accuracy than other machine learning techniques, such as 2D DNN. Unexpectedly, the activation weight in the 3D CAM analysis was high in the ipsilateral motor and somatosensory cortex regions, whereas the gamma-band power was activated in the contralateral areas during unimanual movement, which suggests that the brain signal acquired from the motor cortex contains information about both contralateral movement and ipsilateral movement. Moreover, the hand-movement classification system used critical temporal information at movement onset and offset when classifying bimanual movements. Significance: As far as we know, this is the first study to use high-dimensional neurophysiological information (spatial, spectral, and temporal) with the deep learning method, reconstruct those features, and explain how the neural network works. We expect that our methods can be widely applied and used in neuroscience and electrophysiology research from the point of view of the explainability of XAI as well as its performance.


2019 ◽  
Vol 131 ◽  
pp. 01118
Author(s):  
Fan Tongke

Aiming at the problem of disease diagnosis of large-scale crops, this paper combines machine vision and deep learning technology to propose an algorithm for constructing disease recognition by LM_BP neural network. The images of multiple crop leaves are collected, and the collected pictures are cut by image cutting technology, and the data are obtained by the color distance feature extraction method. The data are input into the disease recognition model, the feature weights are set, and the model is repeatedly trained to obtain accurate results. In this model, the research on corn disease shows that the model is simple and easy to implement, and the data are highly reliable.


Sensors ◽  
2021 ◽  
Vol 21 (9) ◽  
pp. 3164
Author(s):  
Gayoung Jung ◽  
Jonghun Lee ◽  
Incheol Kim

Video scene graph generation (ViDSGG), the creation of video scene graphs that helps in deeper and better visual scene understanding, is a challenging task. Segment-based and sliding-window based methods have been proposed to perform this task. However, they all have certain limitations. This study proposes a novel deep neural network model called VSGG-Net for video scene graph generation. The model uses a sliding window scheme to detect object tracklets of various lengths throughout the entire video. In particular, the proposed model presents a new tracklet pair proposal method that evaluates the relatedness of object tracklet pairs using a pretrained neural network and statistical information. To effectively utilize the spatio-temporal context, low-level visual context reasoning is performed using a spatio-temporal context graph and a graph neural network as well as high-level semantic context reasoning. To improve the detection performance for sparse relationships, the proposed model applies a class weighting technique that adjusts the weight of sparse relationships to a higher level. This study demonstrates the positive effect and high performance of the proposed model through experiments using the benchmark dataset VidOR and VidVRD.


Sign in / Sign up

Export Citation Format

Share Document