scholarly journals Big data, machine learning and artificial intelligence: a neurologist’s guide

2020 ◽  
pp. practneurol-2020-002688
Author(s):  
Stephen D Auger ◽  
Benjamin M Jacobs ◽  
Ruth Dobson ◽  
Charles R Marshall ◽  
Alastair J Noyce

Modern clinical practice requires the integration and interpretation of ever-expanding volumes of clinical data. There is, therefore, an imperative to develop efficient ways to process and understand these large amounts of data. Neurologists work to understand the function of biological neural networks, but artificial neural networks and other forms of machine learning algorithm are likely to be increasingly encountered in clinical practice. As their use increases, clinicians will need to understand the basic principles and common types of algorithm. We aim to provide a coherent introduction to this jargon-heavy subject and equip neurologists with the tools to understand, critically appraise and apply insights from this burgeoning field.

A large volume of datasets is available in various fields that are stored to be somewhere which is called big data. Big Data healthcare has clinical data set of every patient records in huge amount and they are maintained by Electronic Health Records (EHR). More than 80 % of clinical data is the unstructured format and reposit in hundreds of forms. The challenges and demand for data storage, analysis is to handling large datasets in terms of efficiency and scalability. Hadoop Map reduces framework uses big data to store and operate any kinds of data speedily. It is not solely meant for storage system however conjointly a platform for information storage moreover as processing. It is scalable and fault-tolerant to the systems. Also, the prediction of the data sets is handled by machine learning algorithm. This work focuses on the Extreme Machine Learning algorithm (ELM) that can utilize the optimized way of finding a solution to find disease risk prediction by combining ELM with Cuckoo Search optimization-based Support Vector Machine (CS-SVM). The proposed work also considers the scalability and accuracy of big data models, thus the proposed algorithm greatly achieves the computing work and got good results in performance of both veracity and efficiency.


Photoniques ◽  
2020 ◽  
pp. 45-48
Author(s):  
Piotr Antonik ◽  
Serge Massar ◽  
Guy Van Der Sande

The recent progress in artificial intelligence has spurred renewed interest in hardware implementations of neural networks. Reservoir computing is a powerful, highly versatile machine learning algorithm well suited for experimental implementations. The simplest highperformance architecture is based on delay dynamical systems. We illustrate its power through a series of photonic examples, including the first all optical reservoir computer and reservoir computers based on lasers with delayed feedback. We also show how reservoirs can be used to emulate dynamical systems. We discuss the perspectives of photonic reservoir computing.


2021 ◽  
Vol 2083 (4) ◽  
pp. 042086
Author(s):  
Yuqi Qin

Abstract Machine learning algorithm is the core of artificial intelligence, is the fundamental way to make computer intelligent, its application in all fields of artificial intelligence. Aiming at the problems of the existing algorithms in the discrete manufacturing industry, this paper proposes a new 0-1 coding method to optimize the learning algorithm, and finally proposes a learning algorithm of “IG type learning only from the best”.


2022 ◽  
Vol 226 (1) ◽  
pp. S362-S363
Author(s):  
Matthew Hoffman ◽  
Wei Liu ◽  
Jade Tunguhan ◽  
Ghamar Bitar ◽  
Kaveeta Kumar ◽  
...  

2020 ◽  
Vol 25 (2) ◽  
pp. 7-13
Author(s):  
Zhangozha A.R. ◽  

On the example of the online game Akinator, the basic principles on which programs of this type are built are considered. Effective technics have been proposed by which artificial intelligence systems can build logical inferences that allow to identify an unknown subject from its description (predicate). To confirm the considered hypotheses, the terminological analysis of definition of the program "Akinator" offered by the author is carried out. Starting from the assumptions given by the author's definition, the article complements their definitions presented by other researchers and analyzes their constituent theses. Finally, some proposals are made for the next steps in improving the program. The Akinator program, at one time, became one of the most famous online games using artificial intelligence. And although this was not directly stated, it was clear to the experts in the field of artificial intelligence that the program uses the techniques of expert systems and is built on inference rules. At the moment, expert systems have lost their positions in comparison with the direction of neural networks in the field of artificial intelligence, however, in the case considered in the article, we are talking about techniques using both directions – hybrid systems. Games for filling semantics interact with the user, expanding their semantic base (knowledge base) and use certain strategies to achieve the best result. The playful form of such semantics filling programs is beneficial for researchers by involving a large number of players. The article examines the techniques used by the Akinator program, and also suggests possible modifications to it in the future. This study, first of all, focuses on how the knowledge base of the Akinator program is built, it consists of incomplete sets, which can be filled and adjusted as a result of further iterations of the program launches. It is important to note our assumption that the order of questions used by the program during the game plays a key role, because it determines its strategy. It was identified that the program is guided by the principles of nonmonotonic logic – the assumptions constructed by the program are not final and can be rejected by it during the game. The three main approaches to acquisite semantics proposed by Jakub Šimko and Mária Bieliková are considered, namely, expert work, crowdsourcing and machine learning. Paying attention to machine learning, the Akinator program using machine learning to build an effective strategy in the game presents a class of hybrid systems that combine the principles of two main areas in artificial intelligence programs – expert systems and neural networks.


Author(s):  
Ladly Patel ◽  
Kumar Abhishek Gaurav

In today's world, a huge amount of data is available. So, all the available data are analyzed to get information, and later this data is used to train the machine learning algorithm. Machine learning is a subpart of artificial intelligence where machines are given training with data and the machine predicts the results. Machine learning is being used in healthcare, image processing, marketing, etc. The aim of machine learning is to reduce the work of the programmer by doing complex coding and decreasing human interaction with systems. The machine learns itself from past data and then predict the desired output. This chapter describes machine learning in brief with different machine learning algorithms with examples and about machine learning frameworks such as tensor flow and Keras. The limitations of machine learning and various applications of machine learning are discussed. This chapter also describes how to identify features in machine learning data.


2020 ◽  
Vol 41 (7) ◽  
pp. 826-830 ◽  
Author(s):  
Arni S. R. Srinivasa Rao ◽  
Jose A. Vazquez

AbstractWe propose the use of a machine learning algorithm to improve possible COVID-19 case identification more quickly using a mobile phone–based web survey. This method could reduce the spread of the virus in susceptible populations under quarantine.


Author(s):  
Amirata Ghorbani ◽  
Abubakar Abid ◽  
James Zou

In order for machine learning to be trusted in many applications, it is critical to be able to reliably explain why the machine learning algorithm makes certain predictions. For this reason, a variety of methods have been developed recently to interpret neural network predictions by providing, for example, feature importance maps. For both scientific robustness and security reasons, it is important to know to what extent can the interpretations be altered by small systematic perturbations to the input data, which might be generated by adversaries or by measurement biases. In this paper, we demonstrate how to generate adversarial perturbations that produce perceptively indistinguishable inputs that are assigned the same predicted label, yet have very different interpretations. We systematically characterize the robustness of interpretations generated by several widely-used feature importance interpretation methods (feature importance maps, integrated gradients, and DeepLIFT) on ImageNet and CIFAR-10. In all cases, our experiments show that systematic perturbations can lead to dramatically different interpretations without changing the label. We extend these results to show that interpretations based on exemplars (e.g. influence functions) are similarly susceptible to adversarial attack. Our analysis of the geometry of the Hessian matrix gives insight on why robustness is a general challenge to current interpretation approaches.


Sign in / Sign up

Export Citation Format

Share Document