Data Compression of Ship Performance and Navigation Information Under Deep Learning

Author(s):  
Lokukaluge P. Perera ◽  
Brage Mo

Various emission control regulations enforce vessels to collect performance and navigation data and evaluate ship energy efficiency by implementing onboard sensors and data acquisition (DAQ) systems. These DAQ systems are designed to collect, store and communicate large amounts of performance and navigation information through complex data handling processes. It is suggested that this information should eventually be transferred to shore based data analysis centers for further processing and storage. However, the associated data transfer costs introduce additional challenges in this process and enforce to investigate cost effective data handling approaches in shipping. That mainly relates to the amount of data that are transferring through various communication networks (i.e. satellites & wireless networks) between vessels and shore based data centers. Hence, this study proposes to use a deep learning approach (i.e. autoencoder system architecture) to compress ship performance and navigation information, which can be transferred through the respective communication networks as a reduced data set. The compressed data set can be expanded in the respective data center requiring further analysis. Therefore, a data set of ship performance and navigation information is analyzed (i.e. compression and expansion) through an autoencoder system architecture in this study. The compressed data set represents a subset of ship performance and navigation information can also be used to evaluate energy efficiency type applications in shipping. Furthermore, the respective input and output data sets of the autoencoder are also compared as statistical distributions to evaluate the network performance.

2021 ◽  
Vol 15 ◽  
pp. 117793222110303
Author(s):  
Asad Ahmed ◽  
Bhavika Mam ◽  
Ramanathan Sowdhamini

Protein-ligand binding prediction has extensive biological significance. Binding affinity helps in understanding the degree of protein-ligand interactions and is a useful measure in drug design. Protein-ligand docking using virtual screening and molecular dynamic simulations are required to predict the binding affinity of a ligand to its cognate receptor. Performing such analyses to cover the entire chemical space of small molecules requires intense computational power. Recent developments using deep learning have enabled us to make sense of massive amounts of complex data sets where the ability of the model to “learn” intrinsic patterns in a complex plane of data is the strength of the approach. Here, we have incorporated convolutional neural networks to find spatial relationships among data to help us predict affinity of binding of proteins in whole superfamilies toward a diverse set of ligands without the need of a docked pose or complex as user input. The models were trained and validated using a stringent methodology for feature extraction. Our model performs better in comparison to some existing methods used widely and is suitable for predictions on high-resolution protein crystal (⩽2.5 Å) and nonpeptide ligand as individual inputs. Our approach to network construction and training on protein-ligand data set prepared in-house has yielded significant insights. We have also tested DEELIG on few COVID-19 main protease-inhibitor complexes relevant to the current public health scenario. DEELIG-based predictions can be incorporated in existing databases including RSCB PDB, PDBMoad, and PDBbind in filling missing binding affinity data for protein-ligand complexes.


Author(s):  
Lokukaluge P. Perera ◽  
Brage Mo

Ocean internet of things (IoT - onboard and onshore) collects big data sets of ship performance and navigation information under various data handling processes. That extract vessel performance and navigation information that are used for ship energy efficiency and emission control applications. However, the quality of ship performance and navigation data can play an important role in such applications, where sensor faults may introduce various erroneous data regions and that may degrade to the outcome. This study proposes visual analytics, where hidden data patterns, clusters, correlations and other useful information are visually from the respective data set extracted, to identify such erroneous data regions. The domain knowledge (i.e. ship performance and navigation conditions) has also been used to interpret such erroneous data regions and identify the respective sensors that relate to the same situations. Finally, a ship performance and navigation data set of a selected vessel is analyzed to identify erroneous data regions for three selected sensor fault situations (i.e. wind, log speed and draft sensors) under the proposed visual analytics. Hence, this approach can be categorized as a sensor specific fault detection methodology by considering the same results.


2021 ◽  
Vol 11 (24) ◽  
pp. 12116
Author(s):  
Shanza Abbas ◽  
Muhammad Umair Khan ◽  
Scott Uk-Jin Lee ◽  
Asad Abbas

Natural language interfaces to databases (NLIDB) has been a research topic for a decade. Significant data collections are available in the form of databases. To utilize them for research purposes, a system that can translate a natural language query into a structured one can make a huge difference. Efforts toward such systems have been made with pipelining methods for more than a decade. Natural language processing techniques integrated with data science methods are researched as pipelining NLIDB systems. With significant advancements in machine learning and natural language processing, NLIDB with deep learning has emerged as a new research trend in this area. Deep learning has shown potential for rapid growth and improvement in text-to-SQL tasks. In deep learning NLIDB, closing the semantic gap in predicting users’ intended columns has arisen as one of the critical and fundamental problems in this research field. Contributions toward this issue have consisted of preprocessed feature inputs and encoding schema elements afore of and more impactful to the targeted model. Various significant work contributed towards this problem notwithstanding, this has been shown to be one of the critical issues for the task of developing NLIDB. Working towards closing the semantic gap between user intention and predicted columns, we present an approach for deep learning text-to-SQL tasks that includes previous columns’ occurrences scores as an additional input feature. Overall exact match accuracy can also be improved by emphasizing the improvement of columns’ prediction accuracy, which depends significantly on column prediction itself. For this purpose, we extract the query fragments from previous queries’ data and obtain the columns’ occurrences and co-occurrences scores. Column occurrences and co-occurrences scores are processed as input features for the encoder–decoder-based text to the SQL model. These scores contribute, as a factor, the probability of having already used columns and tables together in the query history. We experimented with our approach on the currently popular text-to-SQL dataset Spider. Spider is a complex data set containing multiple databases. This dataset includes query–question pairs along with schema information. We compared our exact match accuracy performance with a base model using their test and training data splits. It outperformed the base model’s accuracy, and accuracy was further boosted in experiments with the pretrained language model BERT.


Author(s):  
Lokukaluge P. Perera ◽  
Brage Mo

Modern ships are supported by internet of things (IoT) to collect ship performance and navigation information. That should be utilized towards digitalization of the shipping industry. However, such information collection systems are always associated with large-scale data sets, so called Big Data, where various industrial challenges are encountered during the respective data handling processes. This study proposes a data handling framework with data driven models (i.e. digital models) to cope with the shipping industrial challenges as the main contribution, where conventional mathematical models may fail. The proposed data driven models are developed in a high dimensional space, where the respective ship performance and navigation parameters of a selected vessel are separated as several data clusters. Hence, this study identifies the distribution of the respective data clusters and the structure of each data cluster in relation to ship performance and navigation conditions. An appropriate structure into the data set of ship performance and navigation parameters is assigned by this method as the main contribution. However, the domain knowledge (i.e. vessel operational and navigation conditions) is also included in this situation to derive a meaningful data structure.


Author(s):  
Lokukaluge P. Perera ◽  
Brage Mo

Modern vessels are monitored by Onboard Internet of Things (IoT), sensors and data acquisition (DAQ), to observe ship performance and navigation conditions. Such IoT may create various shipping industrial challenges under large-scale data handling situations. These large-scale data handling issues are often categorized as “Big Data” challenges and this chapter discusses various solutions to overcome such challenges. That consists of a data-handling framework with various data analytics under onboard IoT. The basis for such data analytics is under data driven models presented and developed with engine-propeller combinator diagrams of vessels. The respective results on data analytics of data classification, sensor faults detection, data compression and expansion, integrity verification and regression, and visualization and decision support, are presented along the proposed data handling framework of a selected vessel. Finally, the results are useful for energy efficiency and system reliability applications of shipping discussed.


2021 ◽  
Vol 2083 (2) ◽  
pp. 022041
Author(s):  
Caiyu Liu ◽  
Zuofeng Zhou ◽  
Qingquan Wu

Abstract As an important part of road maintenance, the detection of road sprinkles has attracted extensive attention from scholars. However, after years of research, there are still some problems in the detection of road sprinkles. First of all, the detection accuracy of traditional detection algorithm is deficient. Second, deep learning approaches have great limitations for there are various kinds of sprinkles which makes it difficult to build a data set. In view of the above problems, this paper proposes a road sprinkling detection method based on multi-feature fusion. The characteristics of color, gradient, luminance and neighborhood information were considered in our method. Compared with other traditional methods, our method has higher detection accuracy. In addition, compared with deep learning-based methods, our approach doesn’t involve creating a complex data set and reduces costs. The main contributions of this paper are as follows: I. For the first time, the density clustering algorithm is combined with the detection of sprinkles, which provides a new idea for this field. II. The use of multi-feature fusion improves the accuracy and robustness of the traditional method which makes the algorithm usable in many real-world scenarios.


2019 ◽  
Vol 2019 (1) ◽  
pp. 360-368
Author(s):  
Mekides Assefa Abebe ◽  
Jon Yngve Hardeberg

Different whiteboard image degradations highly reduce the legibility of pen-stroke content as well as the overall quality of the images. Consequently, different researchers addressed the problem through different image enhancement techniques. Most of the state-of-the-art approaches applied common image processing techniques such as background foreground segmentation, text extraction, contrast and color enhancements and white balancing. However, such types of conventional enhancement methods are incapable of recovering severely degraded pen-stroke contents and produce artifacts in the presence of complex pen-stroke illustrations. In order to surmount such problems, the authors have proposed a deep learning based solution. They have contributed a new whiteboard image data set and adopted two deep convolutional neural network architectures for whiteboard image quality enhancement applications. Their different evaluations of the trained models demonstrated their superior performances over the conventional methods.


2020 ◽  
Vol 17 (3) ◽  
pp. 299-305 ◽  
Author(s):  
Riaz Ahmad ◽  
Saeeda Naz ◽  
Muhammad Afzal ◽  
Sheikh Rashid ◽  
Marcus Liwicki ◽  
...  

This paper presents a deep learning benchmark on a complex dataset known as KFUPM Handwritten Arabic TexT (KHATT). The KHATT data-set consists of complex patterns of handwritten Arabic text-lines. This paper contributes mainly in three aspects i.e., (1) pre-processing, (2) deep learning based approach, and (3) data-augmentation. The pre-processing step includes pruning of white extra spaces plus de-skewing the skewed text-lines. We deploy a deep learning approach based on Multi-Dimensional Long Short-Term Memory (MDLSTM) networks and Connectionist Temporal Classification (CTC). The MDLSTM has the advantage of scanning the Arabic text-lines in all directions (horizontal and vertical) to cover dots, diacritics, strokes and fine inflammation. The data-augmentation with a deep learning approach proves to achieve better and promising improvement in results by gaining 80.02% Character Recognition (CR) over 75.08% as baseline.


2020 ◽  
Vol 27 (5) ◽  
pp. 359-369 ◽  
Author(s):  
Cheng Shi ◽  
Jiaxing Chen ◽  
Xinyue Kang ◽  
Guiling Zhao ◽  
Xingzhen Lao ◽  
...  

: Protein-related interaction prediction is critical to understanding life processes, biological functions, and mechanisms of drug action. Experimental methods used to determine proteinrelated interactions have always been costly and inefficient. In recent years, advances in biological and medical technology have provided us with explosive biological and physiological data, and deep learning-based algorithms have shown great promise in extracting features and learning patterns from complex data. At present, deep learning in protein research has emerged. In this review, we provide an introductory overview of the deep neural network theory and its unique properties. Mainly focused on the application of this technology in protein-related interactions prediction over the past five years, including protein-protein interactions prediction, protein-RNA\DNA, Protein– drug interactions prediction, and others. Finally, we discuss some of the challenges that deep learning currently faces.


Author(s):  
Kyungkoo Jun

Background & Objective: This paper proposes a Fourier transform inspired method to classify human activities from time series sensor data. Methods: Our method begins by decomposing 1D input signal into 2D patterns, which is motivated by the Fourier conversion. The decomposition is helped by Long Short-Term Memory (LSTM) which captures the temporal dependency from the signal and then produces encoded sequences. The sequences, once arranged into the 2D array, can represent the fingerprints of the signals. The benefit of such transformation is that we can exploit the recent advances of the deep learning models for the image classification such as Convolutional Neural Network (CNN). Results: The proposed model, as a result, is the combination of LSTM and CNN. We evaluate the model over two data sets. For the first data set, which is more standardized than the other, our model outperforms previous works or at least equal. In the case of the second data set, we devise the schemes to generate training and testing data by changing the parameters of the window size, the sliding size, and the labeling scheme. Conclusion: The evaluation results show that the accuracy is over 95% for some cases. We also analyze the effect of the parameters on the performance.


Sign in / Sign up

Export Citation Format

Share Document