scholarly journals DETEKSI GANGGUAN LAMBUNG MELALUI CITRA IRIS MATA MENGGUNAKAN METODE JARINGAN SYARAF TIRUAN HEBB RULE

2016 ◽  
Vol 1 (1) ◽  
Author(s):  
Riezka Yana Simamora ◽  
Huzaeni Huzaeni ◽  
Muhammad Rizka
Keyword(s):  

Iridology merupakan metode menganalisis iris mata untuk mendeteksi kelemahan organ tubuh melalui ciri-ciri maupun tanda-tanda yang muncul pada iris mata. Iris mata memiliki kelebihan spesifik yang dapat merekam semua kondisi organ tubuh, salah satunya adalah organ lambung. Dengan memanfaatkan biometrik dan ilmu iridologi, maka pada Tugas Akhir ini dibangun sebuah perangkat lunak untuk mendeteksi gangguan lambung dengan menggunakan metode hebb rule. Mekanisme yang dilakukan oleh sistem dimulai dengan menginput citra iris mata yang kemudian diubah menjadi citra grayscale. Citra iris grayscale ditransformasikan ke dalam koordinat polar untuk memudahkan proses pengambilan daerah lambung pada lapisan pertama iris. Citra iris daerah lambung selanjutnya dilakukan proses pendeteksian tepi dengan menggunakan operator canny yang akan digunakan sebagai input metode hebb rule. Metode hebb rule yang akan menentukan iris mata yang terdapat gangguan lambung ataupun tidak dengan menghitung bobot dan net dari setiap nilai vektor yang membentuk pada pola iris daerah lambung. Terdapat beberapa faktor yang dapat mempengaruhi proses pendeteksian, seperti noise pada citra masukan dan pencahayaan yang masuk ke iris daerah lambung. Dari 40 citra iris mata yang diuji terdapat 31 citra yang mampu dikenali. Sehingga tingkat akurasi sistem ini adalah 77,50%. Berdasarkan hasil tersebut, maka dapat disimpulkan bahwa sistem ini mampu mendeteksi gangguan lambung melalui citra iris mata.Kata Kunci : Iridology, iris mata, lambung, citra polar, canny, hebb rule.

1994 ◽  
Vol 191 (1-2) ◽  
pp. 127-133 ◽  
Author(s):  
Caren Marzban ◽  
Raju Viswanathan

2010 ◽  
Vol 22 (6) ◽  
pp. 1399-1444 ◽  
Author(s):  
Michael Pfeiffer ◽  
Bernhard Nessler ◽  
Rodney J. Douglas ◽  
Wolfgang Maass

We introduce a framework for decision making in which the learning of decision making is reduced to its simplest and biologically most plausible form: Hebbian learning on a linear neuron. We cast our Bayesian-Hebb learning rule as reinforcement learning in which certain decisions are rewarded and prove that each synaptic weight will on average converge exponentially fast to the log-odd of receiving a reward when its pre- and postsynaptic neurons are active. In our simple architecture, a particular action is selected from the set of candidate actions by a winner-take-all operation. The global reward assigned to this action then modulates the update of each synapse. Apart from this global reward signal, our reward-modulated Bayesian Hebb rule is a pure Hebb update that depends only on the coactivation of the pre- and postsynaptic neurons, not on the weighted sum of all presynaptic inputs to the postsynaptic neuron as in the perceptron learning rule or the Rescorla-Wagner rule. This simple approach to action-selection learning requires that information about sensory inputs be presented to the Bayesian decision stage in a suitably preprocessed form resulting from other adaptive processes (acting on a larger timescale) that detect salient dependencies among input features. Hence our proposed framework for fast learning of decisions also provides interesting new hypotheses regarding neural nodes and computational goals of cortical areas that provide input to the final decision stage.


Author(s):  
R. М. Peleshchak ◽  
V. V. Lytvyn ◽  
О. І. Cherniak ◽  
І. R. Peleshchak ◽  
М. V. Doroshenko

Context. To reduce the computational resource time in the problems of diagnosing and recognizing distorted images based on a fully connected stochastic pseudospin neural network, it becomes necessary to thin out synaptic connections between neurons, which is solved using the method of diagonalizing the matrix of synaptic connections without losing interaction between all neurons in the network. Objective. To create an architecture of a stochastic pseudo-spin neural network with diagonal synaptic connections without loosing the interaction between all the neurons in the layer to reduce its learning time. Method. The paper uses the Hausholder method, the method of compressing input images based on the diagonalization of the matrix of synaptic connections and the computer mathematics system MATLAB for converting a fully connected neural network into a tridiagonal form with hidden synaptic connections between all neurons. Results. We developed a model of a stochastic neural network architecture with sparse renormalized synaptic connections that take into account deleted synaptic connections. Based on the transformation of the synaptic connection matrix of a fully connected neural network into a Hessenberg matrix with tridiagonal synaptic connections, we proposed a renormalized local Hebb rule. Using the computer mathematics system “WolframMathematica 11.3”, we calculated, as a function of the number of neurons N, the relative tuning time of synaptic connections (per iteration) in a stochastic pseudospin neural network with a tridiagonal connection Matrix, relative to the tuning time of synaptic connections (per iteration) in a fully connected synaptic neural network. Conclusions. We found that with an increase in the number of neurons, the tuning time of synaptic connections (per iteration) in a stochastic pseudospin neural network with a tridiagonal connection Matrix, relative to the tuning time of synaptic connections (per iteration) in a fully connected synaptic neural network, decreases according to a hyperbolic law. Depending on the direction of pseudospin neurons, we proposed a classification of a renormalized neural network with a ferromagnetic structure, an antiferromagnetic structure, and a dipole glass.


Sign in / Sign up

Export Citation Format

Share Document