scholarly journals On the descriptive power of Neural-Networks as constrained Tensor Networks with exponentially large bond dimension

2021 ◽  
Vol 4 (1) ◽  
Author(s):  
Mario Collura ◽  
Luca Dell'Anna ◽  
Timo Felser ◽  
Simone Montangero

In many cases, neural networks can be mapped into tensor networks with an exponentially large bond dimension. Here, we compare different sub-classes of neural network states, with their mapped tensor network counterpart for studying the ground state of short-range Hamiltonians. We show that when mapping a neural network, the resulting tensor network is highly constrained and thus the neural network states do in general not deliver the naive expected drastic improvement against the state-of-the-art tensor network methods. We explicitly show this result in two paradigmatic examples, the 1D ferromagnetic Ising model and the 2D antiferromagnetic Heisenberg model, addressing the lack of a detailed comparison of the expressiveness of these increasingly popular, variational ans"atze.

2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Ali Ibrahim Khaleel ◽  
Nik Adilah Hanin Zahri ◽  
Muhammad Imran Ahmad

The number of medical images being stored and communicated daily is rapidly increasing, according to the need for these images in medical diagnoses. Hence, the storage space and bandwidths required to store and communicate these images are exponentially increasing, which has brought attention toward compressing these images. In this study, a new compression method is proposed for medical images based on convolutional neural networks. The proposed neural network consists of two main stages: a segmentation stage and an autoencoder. The segmentation stage is used to recognize the Region of Interest (ROI) in the image and provide it to the autoencoder stage, so more emphasis on the information of the ROI is applied. The autoencoder part of the neural network contains a bottleneck layer that has one-eighth of the dimensions of the input image. The values in this layer are used to represent the image, while the following layers are used to decompress the images, after training the neural network. The proposed method is evaluated using the CLEF MED 2009 dataset, where the evaluation results show that the method has significantly better performance, compared to the existing state-of-the-art methods, by providing more visually similar images using less data.


2021 ◽  
Vol 8 ◽  
Author(s):  
Andrey Kardashin ◽  
Alexey Uvarov ◽  
Jacob Biamonte

Tensor network algorithms seek to minimize correlations to compress the classical data representing quantum states. Tensor network algorithms and similar tools—called tensor network methods—form the backbone of modern numerical methods used to simulate many-body physics and have a further range of applications in machine learning. Finding and contracting tensor network states is a computational task, which may be accelerated by quantum computing. We present a quantum algorithm that returns a classical description of a rank-r tensor network state satisfying an area law and approximating an eigenvector given black-box access to a unitary matrix. Our work creates a bridge between several contemporary approaches, including tensor networks, the variational quantum eigensolver (VQE), quantum approximate optimization algorithm (QAOA), and quantum computation.


2020 ◽  
Vol 2020 (10) ◽  
pp. 54-62
Author(s):  
Oleksii VASYLIEV ◽  

The problem of applying neural networks to calculate ratings used in banking in the decision-making process on granting or not granting loans to borrowers is considered. The task is to determine the rating function of the borrower based on a set of statistical data on the effectiveness of loans provided by the bank. When constructing a regression model to calculate the rating function, it is necessary to know its general form. If so, the task is to calculate the parameters that are included in the expression for the rating function. In contrast to this approach, in the case of using neural networks, there is no need to specify the general form for the rating function. Instead, certain neural network architecture is chosen and parameters are calculated for it on the basis of statistical data. Importantly, the same neural network architecture can be used to process different sets of statistical data. The disadvantages of using neural networks include the need to calculate a large number of parameters. There is also no universal algorithm that would determine the optimal neural network architecture. As an example of the use of neural networks to determine the borrower's rating, a model system is considered, in which the borrower's rating is determined by a known non-analytical rating function. A neural network with two inner layers, which contain, respectively, three and two neurons and have a sigmoid activation function, is used for modeling. It is shown that the use of the neural network allows restoring the borrower's rating function with quite acceptable accuracy.


2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Idris Kharroubi ◽  
Thomas Lim ◽  
Xavier Warin

AbstractWe study the approximation of backward stochastic differential equations (BSDEs for short) with a constraint on the gains process. We first discretize the constraint by applying a so-called facelift operator at times of a grid. We show that this discretely constrained BSDE converges to the continuously constrained one as the mesh grid converges to zero. We then focus on the approximation of the discretely constrained BSDE. For that we adopt a machine learning approach. We show that the facelift can be approximated by an optimization problem over a class of neural networks under constraints on the neural network and its derivative. We then derive an algorithm converging to the discretely constrained BSDE as the number of neurons goes to infinity. We end by numerical experiments.


Author(s):  
Saša Vasiljević ◽  
Jasna Glišović ◽  
Nadica Stojanović ◽  
Ivan Grujić

According to the World Health Organization, air pollution with PM10 and PM2.5 (PM-particulate matter) is a significant problem that can have serious consequences for human health. Vehicles, as one of the main sources of PM10 and PM2.5 emissions, pollute the air and the environment both by creating particles by burning fuel in the engine, and by wearing of various elements in some vehicle systems. In this paper, the authors conducted the prediction of the formation of PM10 and PM2.5 particles generated by the wear of the braking system using a neural network (Artificial Neural Networks (ANN)). In this case, the neural network model was created based on the generated particles that were measured experimentally, while the validity of the created neural network was checked by means of a comparative analysis of the experimentally measured amount of particles and the prediction results. The experimental results were obtained by testing on an inertial braking dynamometer, where braking was performed in several modes, that is under different braking parameters (simulated vehicle speed, brake system pressure, temperature, braking time, braking torque). During braking, the concentration of PM10 and PM2.5 particles was measured simultaneously. The total of 196 measurements were performed and these data were used for training, validation, and verification of the neural network. When it comes to simulation, a comparison of two types of neural networks was performed with one output and with two outputs. For each type, network training was conducted using three different algorithms of backpropagation methods. For each neural network, a comparison of the obtained experimental and simulation results was performed. More accurate prediction results were obtained by the single-output neural network for both particulate sizes, while the smallest error was found in the case of a trained neural network using the Levenberg-Marquardt backward propagation algorithm. The aim of creating such a prediction model is to prove that by using neural networks it is possible to predict the emission of particles generated by brake wear, which can be further used for modern traffic systems such as traffic control. In addition, this wear algorithm could be applied on other vehicle systems, such as a clutch or tires.


2021 ◽  
Vol 40 (3) ◽  
pp. 1-13
Author(s):  
Lumin Yang ◽  
Jiajie Zhuang ◽  
Hongbo Fu ◽  
Xiangzhi Wei ◽  
Kun Zhou ◽  
...  

We introduce SketchGNN , a convolutional graph neural network for semantic segmentation and labeling of freehand vector sketches. We treat an input stroke-based sketch as a graph with nodes representing the sampled points along input strokes and edges encoding the stroke structure information. To predict the per-node labels, our SketchGNN uses graph convolution and a static-dynamic branching network architecture to extract the features at three levels, i.e., point-level, stroke-level, and sketch-level. SketchGNN significantly improves the accuracy of the state-of-the-art methods for semantic sketch segmentation (by 11.2% in the pixel-based metric and 18.2% in the component-based metric over a large-scale challenging SPG dataset) and has magnitudes fewer parameters than both image-based and sequence-based methods.


Electronics ◽  
2021 ◽  
Vol 10 (14) ◽  
pp. 1614
Author(s):  
Jonghun Jeong ◽  
Jong Sung Park ◽  
Hoeseok Yang

Recently, the necessity to run high-performance neural networks (NN) is increasing even in resource-constrained embedded systems such as wearable devices. However, due to the high computational and memory requirements of the NN applications, it is typically infeasible to execute them on a single device. Instead, it has been proposed to run a single NN application cooperatively on top of multiple devices, a so-called distributed neural network. In the distributed neural network, workloads of a single big NN application are distributed over multiple tiny devices. While the computation overhead could effectively be alleviated by this approach, the existing distributed NN techniques, such as MoDNN, still suffer from large traffics between the devices and vulnerability to communication failures. In order to get rid of such big communication overheads, a knowledge distillation based distributed NN, called Network of Neural Networks (NoNN), was proposed, which partitions the filters in the final convolutional layer of the original NN into multiple independent subsets and derives smaller NNs out of each subset. However, NoNN also has limitations in that the partitioning result may be unbalanced and it considerably compromises the correlation between filters in the original NN, which may result in an unacceptable accuracy degradation in case of communication failure. In this paper, in order to overcome these issues, we propose to enhance the partitioning strategy of NoNN in two aspects. First, we enhance the redundancy of the filters that are used to derive multiple smaller NNs by means of averaging to increase the immunity of the distributed NN to communication failure. Second, we propose a novel partitioning technique, modified from Eigenvector-based partitioning, to preserve the correlation between filters as much as possible while keeping the consistent number of filters distributed to each device. Throughout extensive experiments with the CIFAR-100 (Canadian Institute For Advanced Research-100) dataset, it has been observed that the proposed approach maintains high inference accuracy (over 70%, 1.53× improvement over the state-of-the-art approach), on average, even when a half of eight devices in a distributed NN fail to deliver their partial inference results.


Electronics ◽  
2020 ◽  
Vol 9 (9) ◽  
pp. 1526 ◽  
Author(s):  
Choongmin Kim ◽  
Jacob A. Abraham ◽  
Woochul Kang ◽  
Jaeyong Chung

Crossbar-based neuromorphic computing to accelerate neural networks is a popular alternative to conventional von Neumann computing systems. It is also referred as processing-in-memory and in-situ analog computing. The crossbars have a fixed number of synapses per neuron and it is necessary to decompose neurons to map networks onto the crossbars. This paper proposes the k-spare decomposition algorithm that can trade off the predictive performance against the neuron usage during the mapping. The proposed algorithm performs a two-level hierarchical decomposition. In the first global decomposition, it decomposes the neural network such that each crossbar has k spare neurons. These neurons are used to improve the accuracy of the partially mapped network in the subsequent local decomposition. Our experimental results using modern convolutional neural networks show that the proposed method can improve the accuracy substantially within about 10% extra neurons.


1991 ◽  
Vol 45 (10) ◽  
pp. 1706-1716 ◽  
Author(s):  
Mark Glick ◽  
Gary M. Hieftje

Artificial neural networks were constructed for the classification of metal alloys based on their elemental constituents. Glow discharge-atomic emission spectra obtained with a photodiode array spectrometer were used in multivariate calibrations for 7 elements in 37 Ni-based alloys (different types) and 15 Fe-based alloys. Subsets of the two major classes formed calibration sets for stepwise multiple linear regression. The remaining samples were used to validate the calibration models. Reference data from the calibration sets were then pooled into a single set to train neural networks with different architectures and different training parameters. After the neural networks learned to discriminate correctly among alloy classes in the training set, their ability to classify samples in the testing set was measured. In general, the neural network approach performed slightly better than the K-nearest neighbor method, but it suffered from a hidden classification mechanism and nonunique solutions. The neural network methodology is discussed and compared with conventional sample-classification techniques, and multivariate calibration of glow discharge spectra is compared with conventional univariate calibration.


2016 ◽  
Vol 38 (2) ◽  
pp. 37-46 ◽  
Author(s):  
Mateusz Kaczmarek ◽  
Agnieszka Szymańska

Abstract Nonlinear structural mechanics should be taken into account in the practical design of reinforced concrete structures. Cracking is one of the major sources of nonlinearity. Description of deflection of reinforced concrete elements is a computational problem, mainly because of the difficulties in modelling the nonlinear stress-strain relationship of concrete and steel. In design practise, in accordance with technical rules (e.g., Eurocode 2), a simplified approach for reinforced concrete is used, but the results of simplified calculations differ from the results of experimental studies. Artificial neural network is a versatile modelling tool capable of making predictions of values that are difficult to obtain in numerical analysis. This paper describes the creation and operation of a neural network for making predictions of deflections of reinforced concrete beams at different load levels. In order to obtain a database of results, that is necessary for training and testing the neural network, a research on measurement of deflections in reinforced concrete beams was conducted by the authors in the Certified Research Laboratory of the Building Engineering Institute at Wrocław University of Science and Technology. The use of artificial neural networks is an innovation and an alternative to traditional methods of solving the problem of calculating the deflections of reinforced concrete elements. The results show the effectiveness of using artificial neural network for predicting the deflection of reinforced concrete beams, compared with the results of calculations conducted in accordance with Eurocode 2. The neural network model presented in this paper can acquire new data and be used for further analysis, with availability of more research results.


Sign in / Sign up

Export Citation Format

Share Document