Two Projection Neural Networks With Reduced Model Complexity for Nonlinear Programming

2020 ◽  
Vol 31 (6) ◽  
pp. 2020-2029 ◽  
Author(s):  
Youshen Xia ◽  
Jun Wang ◽  
Wenzhong Guo
1991 ◽  
Vol 02 (04) ◽  
pp. 331-339 ◽  
Author(s):  
Jiahan Chen ◽  
Michael A. Shanblatt ◽  
Chia-Yiu Maa

A method for improving the performance of artificial neural networks for linear and nonlinear programming is presented. By analyzing the behavior of the conventional penalty function, the reason for the inherent degenerating accuracy is discovered. Based on this, a new combination penalty function is proposed which can ensure that the equilibrium point is acceptably close to the optimal point. A known neural network model has been modified by using the new penalty function and the corresponding circuit scheme is given. Simulation results show that the relative error for linear and nonlinear programming is substantially reduced by the new method.


2021 ◽  
pp. 1-17
Author(s):  
Luis Sa-Couto ◽  
Andreas Wichert

Abstract Convolutional neural networks (CNNs) evolved from Fukushima's neocognitron model, which is based on the ideas of Hubel and Wiesel about the early stages of the visual cortex. Unlike other branches of neocognitron-based models, the typical CNN is based on end-to-end supervised learning by backpropagation and removes the focus from built-in invariance mechanisms, using pooling not as a way to tolerate small shifts but as a regularization tool that decreases model complexity. These properties of end-to-end supervision and flexibility of structure allow the typical CNN to become highly tuned to the training data, leading to extremely high accuracies on typical visual pattern recognition data sets. However, in this work, we hypothesize that there is a flip side to this capability, a hidden overfitting. More concretely, a supervised, backpropagation based CNN will outperform a neocognitron/map transformation cascade (MTCCXC) when trained and tested inside the same data set. Yet if we take both models trained and test them on the same task but on another data set (without retraining), the overfitting appears. Other neocognitron descendants like the What-Where model go in a different direction. In these models, learning remains unsupervised, but more structure is added to capture invariance to typical changes. Knowing that, we further hypothesize that if we repeat the same experiments with this model, the lack of supervision may make it worse than the typical CNN inside the same data set, but the added structure will make it generalize even better to another one. To put our hypothesis to the test, we choose the simple task of handwritten digit classification and take two well-known data sets of it: MNIST and ETL-1. To try to make the two data sets as similar as possible, we experiment with several types of preprocessing. However, regardless of the type in question, the results align exactly with expectation.


Author(s):  
Trevor J. Bihl ◽  
William A. Young II ◽  
Gary R. Weckman

Despite the natural advantage humans have for recognizing and interpreting patterns, large and complex datasets, as in Big Data, preclude efficient human analysis. Artificial neural networks (ANNs) provide a family of pattern recognition approaches for prediction, clustering and classification applicable to KDD with ANN model complexity ranging from simple (for small problems) highly complex (for large issues). To provide a starting point for readers, this chapter first describes foundational concepts that relate to ANNs. A listing of commonly used ANN methods, heuristics, and criteria for initializing ANNs is then discussed. Common pre- and post- data processing methods for dimensionality reduction and data quality issues are then described. The authors then provide a tutorial example of ANN analysis. Finally, the authors list and describe applications of ANNs to specific business related endeavors for further reading.


Author(s):  
Trevor J. Bihl ◽  
William A. Young II ◽  
Gary R. Weckman

Despite the natural advantage humans have for recognizing and interpreting patterns, large and complex datasets, as in big data, preclude efficient human analysis. Artificial neural networks (ANNs) provide a family of pattern recognition approaches for prediction, clustering, and classification applicable to KDD with ANN model complexity ranging from simple (for small problems) to highly complex (for large issues). To provide a starting point for readers, this chapter first describes foundational concepts that relate to ANNs. A listing of commonly used ANN methods, heuristics, and criteria for initializing ANNs are then discussed. Common pre- and post-data processing methods for dimensionality reduction and data quality issues are then described. The authors then provide a tutorial example of ANN analysis. Finally, the authors list and describe applications of ANNs to specific business-related endeavors for further reading.


Sign in / Sign up

Export Citation Format

Share Document