scholarly journals Automatic Normalisation of Historical Text

2021 ◽  
Author(s):  
Alexander Robertson

Spelling variation in historical text negatively impacts the performance of naturallanguage processing techniques, so normalisation is an important pre-processingstep. Current methods fall some way short of perfect accuracy, often requiringlarge amounts of training data to be effective, and are rarely evaluated againsta wide range of historical sources. This thesis evaluates three models: a HiddenMarkov Model, which has not been previously used for historical text normalisation; a soft attention Neural Network model, which has previously only been evaluated on a single German dataset; and a hard attention Neural Network model,which is adapted from work on morphological inflection and applied here to historical text normalisation for the first time. Each is evaluated against multipledatasets taken from prior work on historical text normalisation. This facilitatesdirect comparison of this work to that existing work. The hard attention NeuralNetwork model achieves state-of-the-art normalisation accuracy in all datasets,even when the volume of training data is restricted. This work will be of particular interest to researchers working with noisy historical data which they wouldlike to explore using modern computational techniques.

2013 ◽  
Vol 6 (12) ◽  
pp. 41 ◽  
Author(s):  
César Pinzón ◽  
Carlos Plazaola ◽  
Ilka Banfield ◽  
Amaly Fong ◽  
Adán Vega

In order to achieve automation of the plate forming process by line heating, it is necessary to know in advance the deformation to be obtained under specific heating conditions. Currently, different methods exist to predict deformation, but these are limited to specific applications and most of them depend on the computational capacity so that only simple structures can be analyzed. In this paper, a neural network model that can accurately predict distortions produced during the plate forming process by line heating, for a wide range of initial conditions including large structures, is presented. Results were compared with data existing in the literature showing excellent performance. Excellent results were obtained for those cases out of the range of the training data.


Author(s):  
Brian K. Kestner ◽  
Jimmy C.M. Tai ◽  
Dimitri N. Mavris

This paper presents a computationally efficient methodology for generating training data for a transient neural network model of a tip-jet reaction drive system for potential use as an onboard model in a model based control application. This methodology significantly reduces the number of training points required to capture the transient performance of the system. The challenge in developing an onboard model for a tip-jet reaction drive system is that the model has to operate over the whole flight envelope, to account for the different dynamics present in the system, and to adjust to system degradation or potential faults. In addition, the onboard model must execute in less time than the update interval of the controller. To address these issues, a computationally efficient training methodology and neural network surrogate model have been developed that captures the transient performance of the tip-jet reaction system. As the number of inputs to a neural network becomes large, the computational time needed to generate the number of training points required to accurately represent the range of operating conditions of the system may become quite large also. A challenge for the tip-jet reaction drive system is to minimize the number of neural network training points, while maintaining the high accuracy. To address this issue, a novel training methodology is presented which first trains a steady-state neural network model and uses deviations from steady-state operating conditions to define the transient portion of the training data. The combined results from both the transient and the steady-state training data can then be used to create a single transient neural network of the system. The results in this paper demonstrate that a transient neural network using this new computationally efficient training methodology has the potential to be a feasible option for use as an onboard real-time model for model based control of a tip-jet reaction drive system.


2021 ◽  
Author(s):  
Shubham Pandey ◽  
Jiaxing Qu ◽  
Vladan Stevanovic ◽  
Peter St. John ◽  
Prashun Gorai

The discovery of new inorganic materials in unexplored chemical spaces necessitates calculating total energy quickly and with sufficient accuracy. Machine learning models that provide such a capability for both ground-state (GS) and higher-energy structures would be instrumental in accelerating the screening for new materials over vast chemical spaces. Here, we develop a unique graph neural network model to accurately predict the total energy of both GS and higher-energy hypothetical structures. We use ~16,500 density functional theory calculated total energy from the NREL Materials Database and ~11,000 in-house generated hypothetical structures to train our model, thus making sure that the model is not biased towards either GS or higher-energy structures. We also demonstrate that our model satisfactorily ranks the structures in the correct order of their energies for a given composition. Furthermore, we present a thorough error analysis to explain several failure modes of the model, which highlights both prediction outliers and occasional inconsistencies in the training data. By peeling back layers of the neural network model, we are able to derive chemical trends by analyzing how the model represents learned structures and properties.


Author(s):  
Orhun Soydan ◽  
Ahmet Benliay

In this study, it is aimed to understand the effects of structural and vegetative elements that can be used in landscape designs on the temperature factor, which will greatly affect the climatic comfort, by using artificial neural networks. In this context, measurements were carried out in the morning (08:00-09:00), noon (13:00-14:00) and evening (17:00-18:00) of a total of 100 days, 50 days in each of the winter and summer seasons, at 7 randomly selected points in the Akdeniz University Campus. In these measurements, the temperature difference values of 11 cover elements on 7 different floor covering types were measured, and the ambient air temperature, humidity and wind values were also determined. The temperature differences between the areas where the flooring elements are exposed to direct sun and the shadow effect of different plant and cover elements were determined using an infrared laser thermometer. These values were processed with Neural Designer software and possible temperature difference prediction values were created for 57.750 different alternatives with the help of artificial neural network model from 837 sets of data. Evaluation shows that the maximum temperature difference is 15.6°C at noon in the summer months in the red tartan flooring material and Callistemon viminalis cover material. While the artificial neural network model predicts that there will be a high 2-3° C temperature difference for the alternatives, it has made predictions for temperature differences between 0-10°C in winter and 0-16°C in summer months. Although the temperature differences that will occur in the noon hours are distributed over a wide range of values, it seems that the morning and evening forecasts are concentrated between 0-7°C values. Also, it has been determined that the wind and humidity in the environment are more important factors than the ambient temperature in terms of temperature differences.


Symmetry ◽  
2018 ◽  
Vol 10 (11) ◽  
pp. 648 ◽  
Author(s):  
Ismoilov Nusrat ◽  
Sung-Bong Jang

Artificial neural networks (ANN) have attracted significant attention from researchers because many complex problems can be solved by training them. If enough data are provided during the training process, ANNs are capable of achieving good performance results. However, if training data are not enough, the predefined neural network model suffers from overfitting and underfitting problems. To solve these problems, several regularization techniques have been devised and widely applied to applications and data analysis. However, it is difficult for developers to choose the most suitable scheme for a developing application because there is no information regarding the performance of each scheme. This paper describes comparative research on regularization techniques by evaluating the training and validation errors in a deep neural network model, using a weather dataset. For comparisons, each algorithm was implemented using a recent neural network library of TensorFlow. The experiment results showed that an autoencoder had the worst performance among schemes. When the prediction accuracy was compared, data augmentation and the batch normalization scheme showed better performance than the others.


2014 ◽  
Vol 574 ◽  
pp. 342-346
Author(s):  
Hong Yan Duan ◽  
Huan Rong Zhang ◽  
Ming Zheng ◽  
Xiao Hong Wang

The fracture problems of medium carbon steel under extra-low cycle bend torsion fatigue loading were studied using artificial neural networks (ANN) in this paper. The ANN model exhibited excellent comparison with the experimental results. It was concluded that predicted fracture design parameters by the trained neural network model seem more reasonable compared to approximate methods. It is possible to claim that, ANN is fairly promising prediction technique if properly used. Training ANN model was introduced at first. And then the Training data for the development of the neural network model was obtained from the experiments. The input parameters, the presetting deflection and notch open angle, and the output, the cycle times of fracture were used during the network training. The neural network architecture is designed. The ANN model was developed using back propagation architecture with three layers jump connections, where every layer was connected or linked to every previous layer. The number of hidden neurons was determined according to special formula. The performance of system is summarized at last. In order to facilitate the comparisons of predicted values, the error evaluation and mean relative error are obtained. The result show that the training model has good performance, and the experimental data and predicted data from ANN are in good coherence.


2013 ◽  
Vol 345 ◽  
pp. 272-276 ◽  
Author(s):  
Hong Yan Duan ◽  
You Tang Li ◽  
Zhi Jia Sun ◽  
Yang Yang Zhang

The fracture problems of medium carbon steel (MCS) under extra-low cycle bend torsion loading were studied using artificial neural networks (ANN) in this paper. The training data were used in the formation of training set of ANN. The ANN model exhibited excellent comparison with the experimental results. It was concluded that predicted fracture design parameters by the trained neural network model seem more reasonable compared to approximate methods. It is possible to claim that, ANN is fairly promising prediction technique if properly used. Training ANN model was introduced at first. And then the Training data for the development of the neural network model was obtained from the experiments. The input parameters, notch depth and tip radius of the notch, and the output, the cycle times of fracture were used during the network training. The neural network architecture is designed. The ANN model was developed using back propagation architecture with three layers jump connections, where every layer was connected or linked to every previous layer. The number of hidden neurons was determined according to special formula. The performance of system is summarized at last. In order to facilitate the comparisons of predicted values, the error evaluation and mean relative error are obtained. The result show that the training model has good performance, and the experimental data and predicted data from ANN are in good coherence.


2021 ◽  
Vol 21 (5) ◽  
pp. 221-228
Author(s):  
Byungsik Lee

Neural network models based on deep learning algorithms are increasingly used for estimating pile load capacities as supplements of bearing capacity equations and field load tests. A series of hyperparameter tuning is required to improve the performance and reliability of developing a neural network model. In this study, the number of hidden layers and neurons, the activation functions, the optimizing algorithms of the gradient descent method, and the learning rates were tuned. The grid search method was applied for the tuning, which is a hyperpameter optimizer supplied by the developing platform. The cross-validation method was applied to enhance reliability for model validation. An appropriate number of epochs was determined using the early stopping method to prevent the overfitting of the model to the training data. The performance of the tuned optimum model evaluated for the test data set revealed that the model could estimate pile load capacities approximately with an average absolute error of 3,000 kN and a coefficient of determinant of 0.5.


Author(s):  
Yang Li ◽  
Wenming Zheng ◽  
Zhen Cui ◽  
Tong Zhang ◽  
Yuan Zong

In this paper, we propose a novel neural network model, called bi-hemispheres domain adversarial neural network (BiDANN), for EEG emotion recognition. BiDANN is motivated by the neuroscience findings, i.e., the emotional brain's asymmetries between left and right hemispheres. The basic idea of BiDANN is to map the EEG feature data of both left and right hemispheres into discriminative feature spaces separately, in which the data representations can be classified easily. For further precisely predicting the class labels of testing data, we narrow the distribution shift between training and testing data by using a global and two local domain discriminators, which work adversarially to the classifier to encourage domain-invariant data representations to emerge. After that, the learned classifier from labeled training data can be applied to unlabeled testing data naturally. We conduct two experiments to verify the performance of our BiDANN model on SEED database. The experimental results show that the proposed model achieves the state-of-the-art performance.


2021 ◽  
Author(s):  
Shubham Pandey ◽  
Jiaxing Qu ◽  
Vladan Stevanovic ◽  
Peter St. John ◽  
Prashun Gorai

The discovery of new inorganic materials in unexplored chemical spaces necessitates calculating total energy quickly and with sufficient accuracy. Machine learning models that provide such a capability for both ground-state (GS) and higher-energy structures would be instrumental in accelerating the screening for new materials over vast chemical spaces. Here, we develop a unique graph neural network model to accurately predict the total energy of both GS and higher-energy hypothetical structures. We use ~16,500 density functional theory calculated total energy from the NREL Materials Database and ~11,000 in-house generated hypothetical structures to train our model, thus making sure that the model is not biased towards either GS or higher-energy structures. We also demonstrate that our model satisfactorily ranks the structures in the correct order of their energies for a given composition. Furthermore, we present a thorough error analysis to explain several failure modes of the model, which highlights both prediction outliers and occasional inconsistencies in the training data. By peeling back layers of the neural network model, we are able to derive chemical trends by analyzing how the model represents learned structures and properties.


Sign in / Sign up

Export Citation Format

Share Document