scholarly journals Filtering Reordering Table Using a Novel Recursive Autoencoder Model for Statistical Machine Translation

2017 ◽  
Vol 2017 ◽  
pp. 1-9
Author(s):  
Jinying Kong ◽  
Yating Yang ◽  
Lei Wang ◽  
Xi Zhou ◽  
Tonghai Jiang ◽  
...  

In phrase-based machine translation (PBMT) systems, the reordering table and phrase table are very large and redundant. Unlike most previous works which aim to filter phrase table, this paper proposes a novel deep neural network model to prune reordering table. We cast the task as a deep learning problem where we jointly train two models: a generative model to implement rule embedding and a discriminative model to classify rules. The main contribution of this paper is that we optimize the reordering model in PBMT by filtering reordering table using a recursive autoencoder model. To evaluate the performance of the proposed model, we performed it on public corpus to measure its reordering ability. The experimental results show that our approach obtains high improvement in BLEU score with less scale of reordering table on two language pairs: English-Chinese (+0.28) and Uyghur-Chinese (+0.33) MT.

Electronics ◽  
2021 ◽  
Vol 10 (13) ◽  
pp. 1514
Author(s):  
Seung-Ho Lim ◽  
WoonSik William Suh ◽  
Jin-Young Kim ◽  
Sang-Young Cho

The optimization for hardware processor and system for performing deep learning operations such as Convolutional Neural Networks (CNN) in resource limited embedded devices are recent active research area. In order to perform an optimized deep neural network model using the limited computational unit and memory of an embedded device, it is necessary to quickly apply various configurations of hardware modules to various deep neural network models and find the optimal combination. The Electronic System Level (ESL) Simulator based on SystemC is very useful for rapid hardware modeling and verification. In this paper, we designed and implemented a Deep Learning Accelerator (DLA) that performs Deep Neural Network (DNN) operation based on the RISC-V Virtual Platform implemented in SystemC in order to enable rapid and diverse analysis of deep learning operations in an embedded device based on the RISC-V processor, which is a recently emerging embedded processor. The developed RISC-V based DLA prototype can analyze the hardware requirements according to the CNN data set through the configuration of the CNN DLA architecture, and it is possible to run RISC-V compiled software on the platform, can perform a real neural network model like Darknet. We performed the Darknet CNN model on the developed DLA prototype, and confirmed that computational overhead and inference errors can be analyzed with the DLA prototype developed by analyzing the DLA architecture for various data sets.


2021 ◽  
Vol 10 (9) ◽  
pp. 25394-25398
Author(s):  
Chitra Desai

Deep learning models have demonstrated improved efficacy in image classification since the ImageNet Large Scale Visual Recognition Challenge started since 2010. Classification of images has further augmented in the field of computer vision with the dawn of transfer learning. To train a model on huge dataset demands huge computational resources and add a lot of cost to learning. Transfer learning allows to reduce on cost of learning and also help avoid reinventing the wheel. There are several pretrained models like VGG16, VGG19, ResNet50, Inceptionv3, EfficientNet etc which are widely used.   This paper demonstrates image classification using pretrained deep neural network model VGG16 which is trained on images from ImageNet dataset. After obtaining the convolutional base model, a new deep neural network model is built on top of it for image classification based on fully connected network. This classifier will use features extracted from the convolutional base model.


2019 ◽  
Vol 15 (11) ◽  
pp. 155014771988816 ◽  
Author(s):  
Bing Han ◽  
Xiaohui Yang ◽  
Yafeng Ren ◽  
Wanggui Lan

The running state of a geared transmission system affects the stability and reliability of the whole mechanical system. It will greatly reduce the maintenance cost of a mechanical system to identify the faulty state of the geared transmission system. Based on the measured gear fault vibration signals and the deep learning theory, four fault diagnosis neural network models including fast Fourier transform–deep belief network model, wavelet transform–convolutional neural network model, Hilbert-Huang transform–convolutional neural network model, and comprehensive deep neural network model are developed and trained respectively. The results show that the gear fault diagnosis method based on deep learning theory can effectively identify various gear faults under real test conditions. The comprehensive deep neural network model is the most effective one in gear fault recognition.


Author(s):  
Weimeng Chu ◽  
Shunan Wu ◽  
Xiao He ◽  
Yufei Liu ◽  
Zhigang Wu

The identification accuracy of inertia tensor of combined spacecraft, which is composed by a servicing spacecraft and a captured target, could be easily affected by the measurement noise of angular rate. Due to frequently changing operating environments of combined spacecraft in space, the measurement noise of angular rate can be very complex. In this paper, an inertia tensor identification approach based on deep learning method is proposed to improve the ability of identifying inertia tensor of combined spacecraft in the presence of complex measurement noise. A deep neural network model for identification is constructed and trained by enough training data and a designed learning strategy. To verify the identification performance of the proposed deep neural network model, two testing set with different ranks of measure noises are used for simulation tests. Comparison tests are also delivered among the proposed deep neural network model, recursive least squares identification method, and tradition deep neural network model. The comparison results show that the proposed deep neural network model yields a more accurate and stable identification performance for inertia tensor of combined spacecraft in changeable and complex operating environments.


Circulation ◽  
2020 ◽  
Vol 142 (Suppl_3) ◽  
Author(s):  
Chayakrit Krittanawong ◽  
Kipp W Johnson ◽  
Usman Baber ◽  
Mehmet Aydar ◽  
Zhen Wang ◽  
...  

Introduction: Heart failure (HF) is a leading cause of hospitalization, morbidity and mortality. Deep learning (DL) techniques appear to show promising results in risk stratification and prognosis in several conditions in medicine. However, few methods using DL exist to help quantitatively estimate prognosis of HF. We hypothesized that deep learning (DL) techniques could prognosis of HF using simple variables. We propose application of a custom-built deep-neural-network model to identify mortality in HF patients. Methods: Custom-built deep-neural-networks were assessed using survey data from 42,147 participants from the National Health and Nutrition Examination Survey 1999-2016 (NHANES). Variables were selected using clinical judgment and stepwise backward regressions to develop prediction models. We partitioned the data into training and testing sets and repetitive experiments. We then evaluated model performance based on discrimination and calibration including the area under the receiver-operator characteristics curve (C-statistics), balanced accuracy, probability calibration with sigmoid, and the Brier score, respectively. As sensitivity analyses, we examined results limited to cases with complete clinical information available. We validated models’ performance using Mount Sinai database. Results: Of 42,147 participants with 4,060 variables, 1,491 (3.5%) had HF and HF mortality was 51.8%. In validation cohort, of 26,333 HF patients, the mortality in HF patients was 405 (1.5%). Final model using only 20 variables (age, race, gender, BMI, smoking, alcohol consumption, HTN, COPD, SBP, DBP, HR, HDL, LDL, CRP, A1C, BUN, creatinine, hemoglobin, sodium level, on statin) was tested. A state-of-the-art deep learning models achieved high accuracy for predicting mortality in HF patients with an AUC of 0.96 (95% CI: 0.95-0.99) in the first cohort and AUC of 0.93 (95% CI: 0.91-0.96) in validation cohort. Conclusions: A deep neural network model has shown to have high predictive accuracy and discriminative and calibrative power for prediction of HF mortality. Further research can delineate the clinical implications of DL in predicting HF mortality.


Sensors ◽  
2020 ◽  
Vol 20 (19) ◽  
pp. 5668
Author(s):  
Yan-Cheng Hsu ◽  
Yung-Hui Li ◽  
Ching-Chun Chang ◽  
Latifa Nabila Harfiya

Due to the growing public awareness of cardiovascular disease (CVD), blood pressure (BP) estimation models have been developed based on physiological parameters extracted from both electrocardiograms (ECGs) and photoplethysmograms (PPGs). Still, in order to enhance the usability as well as reduce the sensor cost, researchers endeavor to establish a generalized BP estimation model using only PPG signals. In this paper, we propose a deep neural network model capable of extracting 32 features exclusively from PPG signals for BP estimation. The effectiveness and accuracy of our proposed model was evaluated by the root mean square error (RMSE), mean absolute error (MAE), the Association for the Advancement of Medical Instrumentation (AAMI) standard and the British Hypertension Society (BHS) standard. Experimental results showed that the RMSEs in systolic blood pressure (SBP) and diastolic blood pressure (DBP) are 4.643 mmHg and 3.307 mmHg, respectively, across 9000 subjects, with 80.63% of absolute errors among estimated SBP records lower than 5 mmHg and 90.19% of absolute errors among estimated DBP records lower than 5 mmHg. We demonstrated that our proposed model has remarkably high accuracy on the largest BP database found in the literature, which shows its effectiveness compared to some prior works.


Sensors ◽  
2019 ◽  
Vol 19 (5) ◽  
pp. 1085 ◽  
Author(s):  
Yeongtaek Song ◽  
Incheol Kim

This paper proposes a novel deep neural network model for solving the spatio-temporal-action-detection problem, by localizing all multiple-action regions and classifying the corresponding actions in an untrimmed video. The proposed model uses a spatio-temporal region proposal method to effectively detect multiple-action regions. First, in the temporal region proposal, anchor boxes were generated by targeting regions expected to potentially contain actions. Unlike the conventional temporal region proposal methods, the proposed method uses a complementary two-stage method to effectively detect the temporal regions of the respective actions occurring asynchronously. In addition, to detect a principal agent performing an action among the people appearing in a video, the spatial region proposal process was used. Further, coarse-level features contain comprehensive information of the whole video and have been frequently used in conventional action-detection studies. However, they cannot provide detailed information of each person performing an action in a video. In order to overcome the limitation of coarse-level features, the proposed model additionally learns fine-level features from the proposed action tubes in the video. Various experiments conducted using the LIRIS-HARL and UCF-10 datasets confirm the high performance and effectiveness of the proposed deep neural network model.


2021 ◽  
Vol 2021 ◽  
pp. 1-13
Author(s):  
Di Mu ◽  
Shuning Wang

It is important to accurately estimate the SOC to ensure that the lithium-ion battery is within a safe working range, prevent over-charging and over-discharging, and ultimately improve battery life. However, SOC is an internal state of the battery and cannot be directly measured. This paper proposes a SOC estimation method based on the wide and deep neural network model, which combines the linear regression (LR) model and the backpropagation neural network (BPNN) model. This article uses the dataset provided by the Advanced Energy Storage and Applications (AESA) group to verify the performance of the model. The performance of the proposed model is compared with the common BPNN model in terms of root mean square error (RMSE), average absolute proportional error (MAPE), and SOC estimation error. The validation results prove that the effect of the proposed model in estimating SOC is better than that of the ordinary BPNN model. Compared with the BPNN model, the RMSE values of the SOC predicted value of the wide and deep model in the charging and discharging stages were reduced by 10.2% and 15.4%, respectively. Experimental results show that the maximum SOC estimation error of the model in predicting the SOC during charging and discharging is 0.42% and 0.86%, respectively.


Sign in / Sign up

Export Citation Format

Share Document