scholarly journals Dual stochasticity in the cortex as a biologically plausible learning with the most efficient coding

2019 ◽  
Author(s):  
Jun-nosuke Teramae

AbstractNeurons and synapses in the cerebral cortex behave stochastically. The advantages of such stochastic properties have been proposed in several works, but the relationship and synergy between the stochasticities of neurons and synapses remain largely unexplored. Here, we show that these stochastic features can be inseparably integrated into a simple framework that provides a practical and biologically plausible learning algorithm that consistently accounts for various experimental results, including the most efficient power-law coding of the cortex. The derived algorithm overcomes many of the limitations of conventional learning algorithms of neural networks. As an experimentally testable prediction, we derived the slow retrograde modulation of the excitability of neurons from this algorithm. Because of the simplicity and flexibility of this algorithm, we anticipate that it will be useful in the development of neuromorphic devices and scalable AI chips, and that it will help bridge the gap between neuroscience and machine learning.

1989 ◽  
Vol 1 (3) ◽  
pp. 201-222 ◽  
Author(s):  
Adam N. Mamelak ◽  
J. Allan Hobson

Bizarreness is a cognitive feature common to REM sleep dreams, which can be easily measured. Because bizarreness is highly specific to dreaming, we propose that it is most likely brought about by changes in neuronal activity that are specific to REM sleep. At the level of the dream plot, bizarreness can be defined as either discontinuity or incongruity. In addition, the dreamer's thoughts about the plot may be logically deficient. We propose that dream bizarreness is the cognitive concomitant of two kinds of changes in neuronal dynamics during REM sleep. One is the disinhibition of forebrain networks caused by the withdrawal of the modulatory influences of norepinephrine (NE) and serotonin (5HT) in REM sleep, secondary to cessation of firing of locus coeruleus and dorsal raphe neurons. This aminergic demodulation can be mathematically modeled as a shift toward increased error at the outputs from neural networks, and these errors might be represented cognitively as incongruities and/or discontinuities. We also consider the possibility that discontinuities are the cognitive concomitant of sudden bifurcations or “jumps” in the responses of forebrain neuronal networks. These bifurcations are caused by phasic discharge of pontogeniculooccipital (PGO) neurons during REM sleep, providing a source of cholinergic modulation to the forebrain which could evoke unpredictable network responses. When phasic PGO activity stops, the resultant activity in the brain may be wholly unrelated to patterns of activity dominant before such phasic stimulation began. Mathematically such sudden shifts from one pattern of activity to a second, unrelated one is called a bifurcation. We propose that the neuronal bifurcations brought about by PGO activity might be represented cognitively as bizarre discontinuities of dream plot. We regard these proposals as preliminary attempts to model the relationship between dream cognition and REM sleep neurophysiology. This neurophysiological model of dream bizarreness may also prove useful in understanding the contributions of REM sleep to the developmental and experiential plasticity of the cerebral cortex.


Author(s):  
Amirata Ghorbani ◽  
Abubakar Abid ◽  
James Zou

In order for machine learning to be trusted in many applications, it is critical to be able to reliably explain why the machine learning algorithm makes certain predictions. For this reason, a variety of methods have been developed recently to interpret neural network predictions by providing, for example, feature importance maps. For both scientific robustness and security reasons, it is important to know to what extent can the interpretations be altered by small systematic perturbations to the input data, which might be generated by adversaries or by measurement biases. In this paper, we demonstrate how to generate adversarial perturbations that produce perceptively indistinguishable inputs that are assigned the same predicted label, yet have very different interpretations. We systematically characterize the robustness of interpretations generated by several widely-used feature importance interpretation methods (feature importance maps, integrated gradients, and DeepLIFT) on ImageNet and CIFAR-10. In all cases, our experiments show that systematic perturbations can lead to dramatically different interpretations without changing the label. We extend these results to show that interpretations based on exemplars (e.g. influence functions) are similarly susceptible to adversarial attack. Our analysis of the geometry of the Hessian matrix gives insight on why robustness is a general challenge to current interpretation approaches.


2020 ◽  
pp. practneurol-2020-002688
Author(s):  
Stephen D Auger ◽  
Benjamin M Jacobs ◽  
Ruth Dobson ◽  
Charles R Marshall ◽  
Alastair J Noyce

Modern clinical practice requires the integration and interpretation of ever-expanding volumes of clinical data. There is, therefore, an imperative to develop efficient ways to process and understand these large amounts of data. Neurologists work to understand the function of biological neural networks, but artificial neural networks and other forms of machine learning algorithm are likely to be increasingly encountered in clinical practice. As their use increases, clinicians will need to understand the basic principles and common types of algorithm. We aim to provide a coherent introduction to this jargon-heavy subject and equip neurologists with the tools to understand, critically appraise and apply insights from this burgeoning field.


Machine learning in recent years has become an integral part of our day to day life and the ease of use has improved a lot in the past decade.There are various ways to make the model to work in smaller devices.A modest method to advance any machine learning algorithm to work in smaller devices is to provide the output of large complex models as input to smaller models which can be easily deployed into mobile phones .We provided a framework where the large models can even learn the domain knowledge which is integrated as first-order logic rules and explicitly includes that knowledge into the smaller model by simultaneously training of both the models.This can be achieved by transfer learning where the knowledge learned by one model can be used to teach the other model.Domain knowledge integration is the most critical part here and it can be done by using some of the constraint principles where the scope of the data is reduced based upon the constraints mentioned. One of the best representation of domain knowledge is logic rules where the knowledge is encoded as predicates.This framework provides a way to integrate human knowledge into deep neural networks that can be easily deployed into any devices.


Author(s):  
Fawziya M. Rammo ◽  
Mohammed N. Al-Hamdani

Many languages identification (LID) systems rely on language models that use machine learning (ML) approaches, LID systems utilize rather long recording periods to achieve satisfactory accuracy. This study aims to extract enough information from short recording intervals in order to successfully classify the spoken languages under test. The classification process is based on frames of (2-18) seconds where most of the previous LID systems were based on much longer time frames (from 3 seconds to 2 minutes). This research defined and implemented many low-level features using MFCC (Mel-frequency cepstral coefficients), containing speech files in five languages (English. French, German, Italian, Spanish), from voxforge.org an open-source corpus that consists of user-submitted audio clips in various languages, is the source of data used in this paper. A CNN (convolutional Neural Networks) algorithm applied in this paper for classification and the result was perfect, binary language classification had an accuracy of 100%, and five languages classification with six languages had an accuracy of 99.8%.


2020 ◽  
Vol ahead-of-print (ahead-of-print) ◽  
Author(s):  
Yejin Lee ◽  
Dae-Young Kim

Purpose Using the decision tree model, this study aims to understand the online travelers booking behaviors on Expedia.com, by examining influential determinants of online hotel booking, especially for longer-stay travelers. The geographical distance is also considered in understanding the booking behaviors trisecting travel destinations (i.e. Americas, Europe and Asia). Design/methodology/approach The data were obtained from American Statistical Association DataFest and Expedia.com. Based on the US travelers who made hotel reservation on the website, the study used a machine learning algorithm, decision tree, to analyze the influential determinants on hotel booking considering the geographical distance between origin and destination. Findings The results of the findings demonstrate that the choice of package product is the prioritized determinant for longer-stay hotel guests. Several similarities and differences were found from the significant determinants of the decision tree, in accordance with the geographic distance among the Americas, Europe and Asia. Research limitations/implications This paper presents the extension to an existing machine learning environment, and especially to the decision tree model. The findings are anticipated to expand the understanding of online hotel booking and apprehend the influential determinants toward consumers’ decision-making process regarding the relationship between geographical distance and traveler’s hotel staying duration. Originality/value This research brings a meaningful understanding of the hospitality and tourism industry, especially to the realm of machine learning adapted to an online booking website. It provides a unique approach to comprehend and forecast consumer behavior with data mining.


2020 ◽  
Vol 10 (3) ◽  
pp. 1042 ◽  
Author(s):  
Juan L. Rastrollo-Guerrero ◽  
Juan A. Gómez-Pulido ◽  
Arturo Durán-Domínguez

Predicting students’ performance is one of the most important topics for learning contexts such as schools and universities, since it helps to design effective mechanisms that improve academic results and avoid dropout, among other things. These are benefited by the automation of many processes involved in usual students’ activities which handle massive volumes of data collected from software tools for technology-enhanced learning. Thus, analyzing and processing these data carefully can give us useful information about the students’ knowledge and the relationship between them and the academic tasks. This information is the source that feeds promising algorithms and methods able to predict students’ performance. In this study, almost 70 papers were analyzed to show different modern techniques widely applied for predicting students’ performance, together with the objectives they must reach in this field. These techniques and methods, which pertain to the area of Artificial Intelligence, are mainly Machine Learning, Collaborative Filtering, Recommender Systems, and Artificial Neural Networks, among others.


2019 ◽  
Vol 40 (Supplement_1) ◽  
Author(s):  
I Korsakov ◽  
A Gusev ◽  
T Kuznetsova ◽  
D Gavrilov ◽  
R Novitskiy

Abstract Abstract Background Advances in precision medicine will require an increasingly individualized prognostic evaluation of patients in order to provide the patient with appropriate therapy. The traditional statistical methods of predictive modeling, such as SCORE, PROCAM, and Framingham, according to the European guidelines for the prevention of cardiovascular disease, not adapted for all patients and require significant human involvement in the selection of predictive variables, transformation and imputation of variables. In ROC-analysis for prediction of significant cardiovascular disease (CVD), the areas under the curve for Framingham: 0.62–0.72, for SCORE: 0.66–0.73 and for PROCAM: 0.60–0.69. To improve it, we apply for approaches to predict a CVD event rely on conventional risk factors by machine learning and deep learning models to 10-year CVD event prediction by using longitudinal electronic health record (EHR). Methods For machine learning, we applied logistic regression (LR) and recurrent neural networks with long short-term memory (LSTM) units as a deep learning algorithm. We extract from longitudinal EHR the following features: demographic, vital signs, diagnoses (ICD-10-cm: I21-I22.9: I61-I63.9) and medication. The problem in this step, that near 80 percent of clinical information in EHR is “unstructured” and contains errors and typos. Missing data are important for the correct training process using by deep learning & machine learning algorithm. The study cohort included patients between the ages of 21 to 75 with a dynamic observation window. In total, we got 31517 individuals in the dataset, but only 3652 individuals have all features or missing features values can be easy to impute. Among these 3652 individuals, 29.4% has a CVD, mean age 49.4 years, 68,2% female. Evaluation We randomly divided the dataset into a training and a test set with an 80/20 split. The LR was implemented with Python Scikit-Learn and the LSTM model was implemented with Keras using Tensorflow as the backend. Results We applied machine learning and deep learning models using the same features as traditional risk scale and longitudinal EHR features for CVD prediction, respectively. Machine learning model (LR) achieved an AUROC of 0.74–0.76 and deep learning (LSTM) 0.75–0.76. By using features from EHR logistic regression and deep learning models improved the AUROC to 0.78–0.79. Conclusion The machine learning models outperformed a traditional clinically-used predictive model for CVD risk prediction (i.e. SCORE, PROCAM, and Framingham equations). This approach was used to create a clinical decision support system (CDSS). It uses both traditional risk scales and models based on neural networks. Especially important is the fact that the system can calculate the risks of cardiovascular disease automatically and recalculate immediately after adding new information to the EHR. The results are delivered to the user's personal account.


Author(s):  
Chandrahas Mishra ◽  
D. L. Gupta

Deep learning is a technique of machine learning in artificial intelligence area. Deep learning in a refined "machine learning" algorithm that far surpasses a considerable lot of its forerunners in its capacities to perceive syllables and picture. Deep learning is as of now a greatly dynamic examination territory in machine learning and example acknowledgment society. It has increased colossal triumphs in an expansive zone of utilizations, for example, speech recognition, computer vision and natural language processing and numerous industry item. Neural network is used to implement the machine learning or to design intelligent machines. In this paper brief introduction to all machine learning paradigm and application area of deep machine learning and different types of neural networks with applications is discussed.


Sign in / Sign up

Export Citation Format

Share Document