Application of matrix filters and braid theory for the procedural generation of neural network architectures

Author(s):  
О.А. Лукьянова ◽  
О.Ю. Никитин ◽  
А.С. Кунин

Представлены результаты исследований, связанных с автоматическим формированием архитектур нейронных сетей, состоящих из набора модулей. Реализован алгоритмический подход, основанный на формировании матриц активных модулей. Предложены способы процедурной генерации архитектур нейронных сетей для решения задач классификации. Дается описание процесса автоматического формирования архитектуры PathNet, реализованной на основе новых подходов, а также рассматриваются примеры генерации трех новых архитектур глубоких нейронных сетей (3DNN, GraphNet и BraidNet). Архитектура BraidNet включает в себя построение графа связей сети на основе теории кос. Исследование на примере задачи классификации изображений MNIST показало применимость всех четырех предложенных нейронных сетей к распознаванию образов. There are various approaches to the algorithmic specification of the network structure in the deep learning problems, which are successfully used in applications. These methods can be generalized by the concept of procedural generation of neural network architectures. Methodology. In the work, we use binary matrix filters. The filters are obtained with the help of the Hadamard product. Such filters define active network modules, thereby changing the way information is transmitted between layers. To build various architectures, the theory of braids is used in the work. The article reproduces the wellknown PathNet architecture. Examples of generating three new deep neural network architectures (3DNN, GraphNet, and BraidNet) are examined. Findings. The paper shows how the procedural generation of neural network architectures allows avoiding manually setting the network structure and automatically forming it. The use of matrix filters simplifies the process of generating network architecture due to a large number of possible combinations of modules and connections between them. Using the MNIST classification problem as an example, it is shown how the architectures presented in the article solve real-world pattern recognition problems. The results of application of neural networks indicate their diminishing tendency to retraining due to the subsequent convergence and the presence of stochastic dynamics in the learning process. Originality/value. Learning methods with dynamic adaptive changes in the network architecture allows achieving satisfactory accuracy faster and should also be less prone to retraining. The BraidNet algorithm presented in the article is applicable for ICT SB RAS, 2019 a convenient brief record of the structure of a neural network in genetic algorithms. Such features make BraidNet a promising algorithm for further application and research in complex problems of pattern recognition, including using neuroevolutionary approaches.

2002 ◽  
Vol 14 (9) ◽  
pp. 2157-2179 ◽  
Author(s):  
M. W. Spratling ◽  
M. H. Johnson

A large and influential class of neural network architectures uses postintegration lateral inhibition as a mechanism for competition. We argue that these algorithms are computationally deficient in that they fail to generate, or learn, appropriate perceptual representations under certain circumstances. An alternative neural network architecture is presented here in which nodes compete for the right to receive inputs rather than for the right to generate outputs. This form of competition, implemented through preintegration lateral inhibition, does provide appropriate coding properties and can be used to learn such representations efficiently. Furthermore, this architecture is consistent with both neuroanatomical and neurophysiological data. We thus argue that preintegration lateral inhibition has computational advantages over conventional neural network architectures while remaining equally biologically plausible.


2000 ◽  
Vol 15 (2) ◽  
pp. 151-170 ◽  
Author(s):  
MIROSLAV KUBAT

An appropriately designed architecture of a neural network is essential to many realistic pattern-recognition tasks. A choice of just the right number of neurons, and their interconnections, can cut learning costs by orders of magnitude, and still warrant high classification accuracy. Surprisingly, textbooks often neglect this issue. A specialist seeking systematic information will soon realize that relevant material is scattered over diverse sources, each with a different perspective, terminology and goals. This brief survey attempts to rectify the situation by explaining the involved aspects, and by describing some of the fundamental techniques.


2021 ◽  
Vol 4 (2) ◽  
pp. 335-342
Author(s):  
Juliansyah Putra Tanjung

There are various types of wheat scattered in the world. Usually it takes a long time to recognize the type of wheat seed by manual method because wheat germ has a physical appearance that looks the same as others. One method that can be used is an Artificial Neural Network. In this study, the data used were secondary data which consisted of data from the variable physical characteristics of wheat germ. The types of wheat seeds that are classified are 3. The Artificial Neural Network architecture used in this study is 5. By comparing the 5 Artificial Neural Network architectures, it is concluded that the architecture consisting of 3 layers and 4 layers is more precise in the classification of wheat germ types. The accuracy obtained by the 2 Artificial Neural Network architectures is 90% and 90%, respectively.


Sign in / Sign up

Export Citation Format

Share Document