The Machine-Learning Approach of Reinforcement Learning

Author(s):  
Thomas Boraud

This chapter assesses alternative approaches of reinforcement learning that are developed by machine learning. The initial goal of this branch of artificial intelligence, which appeared in the middle of the twentieth century, was to develop and implement algorithms that allow a machine to learn. Originally, they were computers or more or less autonomous robotic automata. As artificial intelligence has developed and cross-fertilized with neuroscience, it has begun to be used to model the learning and decision-making processes for biological agents, broadening the meaning of the word ‘machine’. Theoreticians of this discipline define several categories of learning, but this chapter only deals with those which are related to reinforcement learning. To understand how these algorithms work, it is necessary first of all to explain the Markov chain and the Markov decision-making process. The chapter then goes on to examine model-free reinforcement learning algorithms, the actor-critic model, and finally model-based reinforcement learning algorithms.

2021 ◽  
Author(s):  
Yew Kee Wong

Deep learning is a type of machine learning that trains a computer to perform human-like tasks, such as recognizing speech, identifying images or making predictions. Instead of organizing data to run through predefined equations, deep learning sets up basic parameters about the data and trains the computer to learn on its own by recognizing patterns using many layers of processing. This paper aims to illustrate some of the different deep learning algorithms and methods which can be applied to artificial intelligence analysis, as well as the opportunities provided by the application in various decision making domains.


2020 ◽  
Vol 23 (6) ◽  
pp. 1172-1191
Author(s):  
Artem Aleksandrovich Elizarov ◽  
Evgenii Viktorovich Razinkov

Recently, such a direction of machine learning as reinforcement learning has been actively developing. As a consequence, attempts are being made to use reinforcement learning for solving computer vision problems, in particular for solving the problem of image classification. The tasks of computer vision are currently one of the most urgent tasks of artificial intelligence. The article proposes a method for image classification in the form of a deep neural network using reinforcement learning. The idea of ​​the developed method comes down to solving the problem of a contextual multi-armed bandit using various strategies for achieving a compromise between exploitation and research and reinforcement learning algorithms. Strategies such as -greedy, -softmax, -decay-softmax, and the UCB1 method, and reinforcement learning algorithms such as DQN, REINFORCE, and A2C are considered. The analysis of the influence of various parameters on the efficiency of the method is carried out, and options for further development of the method are proposed.


2012 ◽  
pp. 695-703
Author(s):  
George Tzanis ◽  
Christos Berberidis ◽  
Ioannis Vlahavas

Machine learning is one of the oldest subfields of artificial intelligence and is concerned with the design and development of computational systems that can adapt themselves and learn. The most common machine learning algorithms can be either supervised or unsupervised. Supervised learning algorithms generate a function that maps inputs to desired outputs, based on a set of examples with known output (labeled examples). Unsupervised learning algorithms find patterns and relationships over a given set of inputs (unlabeled examples). Other categories of machine learning are semi-supervised learning, where an algorithm uses both labeled and unlabeled examples, and reinforcement learning, where an algorithm learns a policy of how to act given an observation of the world.


Author(s):  
George Tzanis ◽  
Christos Berberidis ◽  
Ioannis Vlahavas

Machine learning is one of the oldest subfields of artificial intelligence and is concerned with the design and development of computational systems that can adapt themselves and learn. The most common machine learning algorithms can be either supervised or unsupervised. Supervised learning algorithms generate a function that maps inputs to desired outputs, based on a set of examples with known output (labeled examples). Unsupervised learning algorithms find patterns and relationships over a given set of inputs (unlabeled examples). Other categories of machine learning are semi-supervised learning, where an algorithm uses both labeled and unlabeled examples, and reinforcement learning, where an algorithm learns a policy of how to act given an observation of the world.


2021 ◽  
Vol 73 (09) ◽  
pp. 43-43
Author(s):  
Reza Garmeh

The digital transformation that began several years ago continues to grow and evolve. With new advancements in data analytics and machine-learning algorithms, field developers today see more benefits to upgrading their traditional development work flows to automated artificial-intelligence work flows. The transformation has helped develop more-efficient and truly integrated development approaches. Many development scenarios can be automatically generated, examined, and updated very quickly. These approaches become more valuable when coupled with physics-based integrated asset models that are kept close to actual field performance to reduce uncertainty for reactive decision making. In unconventional basins with enormous completion and production databases, data-driven decisions powered by machine-learning techniques are increasing in popularity to solve field development challenges and optimize cube development. Finding a trend within massive amounts of data requires an augmented artificial intelligence where machine learning and human expertise are coupled. With slowed activity and uncertainty in the oil and gas industry from the COVID-19 pandemic and growing pressure for cleaner energy and environmental regulations, operators had to shift economic modeling for environmental considerations, predicting operational hazards and planning mitigations. This has enlightened the value of field development optimization, shifting from traditional workflow iterations on data assimilation and sequential decision making to deep reinforcement learning algorithms to find the best well placement and well type for the next producer or injector. Operators are trying to adapt with the new environment and enhance their capabilities to efficiently plan, execute, and operate field development plans. Collaboration between different disciplines and integrated analyses are key to the success of optimized development strategies. These selected papers and the suggested additional reading provide a good view of what is evolving with field development work flows using data analytics and machine learning in the era of digital transformation. Recommended additional reading at OnePetro: www.onepetro.org. SPE 203073 - Data-Driven and AI Methods To Enhance Collaborative Well Planning and Drilling-Risk Prediction by Richard Mohan, ADNOC, et al. SPE 200895 - Novel Approach To Enhance the Field Development Planning Process and Reservoir Management To Maximize the Recovery Factor of Gas Condensate Reservoirs Through Integrated Asset Modeling by Oswaldo Espinola Gonzalez, Schlumberger, et al. SPE 202373 - Efficient Optimization and Uncertainty Analysis of Field Development Strategies by Incorporating Economic Decisions in Reservoir Simulation Models by James Browning, Texas Tech University, et al.


2018 ◽  
Vol 30 (10) ◽  
pp. 1391-1404 ◽  
Author(s):  
Wouter Kool ◽  
Samuel J. Gershman ◽  
Fiery A. Cushman

Decision-making algorithms face a basic tradeoff between accuracy and effort (i.e., computational demands). It is widely agreed that humans can choose between multiple decision-making processes that embody different solutions to this tradeoff: Some are computationally cheap but inaccurate, whereas others are computationally expensive but accurate. Recent progress in understanding this tradeoff has been catalyzed by formalizing it in terms of model-free (i.e., habitual) versus model-based (i.e., planning) approaches to reinforcement learning. Intuitively, if two tasks offer the same rewards for accuracy but one of them is much more demanding, we might expect people to rely on habit more in the difficult task: Devoting significant computation to achieve slight marginal accuracy gains would not be “worth it.” We test and verify this prediction in a sequential reinforcement learning task. Because our paradigm is amenable to formal analysis, it contributes to the development of a computational model of how people balance the costs and benefits of different decision-making processes in a task-specific manner; in other words, how we decide when hard thinking is worth it.


2021 ◽  
Vol 29 (Supplement_1) ◽  
pp. i18-i18
Author(s):  
N Hassan ◽  
R Slight ◽  
D Weiand ◽  
A Vellinga ◽  
G Morgan ◽  
...  

Abstract Introduction Sepsis is a life-threatening condition that is associated with increased mortality. Artificial intelligence tools can inform clinical decision making by flagging patients who may be at risk of developing infection and subsequent sepsis and assist clinicians with their care management. Aim To identify the optimal set of predictors used to train machine learning algorithms to predict the likelihood of an infection and subsequent sepsis and inform clinical decision making. Methods This systematic review was registered in PROSPERO database (CRD42020158685). We searched 3 large databases: Medline, Cumulative Index of Nursing and Allied Health Literature, and Embase, using appropriate search terms. We included quantitative primary research studies that focused on sepsis prediction associated with bacterial infection in adult population (>18 years) in all care settings, which included data on predictors to develop machine learning algorithms. The timeframe of the search was 1st January 2000 till the 25th November 2019. Data extraction was performed using a data extraction sheet, and a narrative synthesis of eligible studies was undertaken. Narrative analysis was used to arrange the data into key areas, and compare and contrast between the content of included studies. Quality assessment was performed using Newcastle-Ottawa Quality Assessment scale, which was used to evaluate the quality of non-randomized studies. Bias was not assessed due to the non-randomised nature of the included studies. Results Fifteen articles met our inclusion criteria (Figure 1). We identified 194 predictors that were used to train machine learning algorithms to predict infection and subsequent sepsis, with 13 predictors used on average across all included studies. The most significant predictors included age, gender, smoking, alcohol intake, heart rate, blood pressure, lactate level, cardiovascular disease, endocrine disease, cancer, chronic kidney disease (eGFR<60ml/min), white blood cell count, liver dysfunction, surgical approach (open or minimally invasive), and pre-operative haematocrit < 30%. These predictors were used for the development of all the algorithms in the fifteen articles. All included studies used artificial intelligence techniques to predict the likelihood of sepsis, with average sensitivity 77.5±19.27, and average specificity 69.45±21.25. Conclusion The type of predictors used were found to influence the predictive power and predictive timeframe of the developed machine learning algorithm. Two strengths of our review were that we included studies published since the first definition of sepsis was published in 2001, and identified factors that can improve the predictive ability of algorithms. However, we note that the included studies had some limitations, with three studies not validating the models that they developed, and many tools limited by either their reduced specificity or sensitivity or both. This work has important implications for practice, as predicting the likelihood of sepsis can help inform the management of patients and concentrate finite resources to those patients who are most at risk. Producing a set of predictors can also guide future studies in developing more sensitive and specific algorithms with increased predictive time window to allow for preventive clinical measures.


2021 ◽  
Author(s):  
Jorge Crespo Alvarez ◽  
Bryan Ferreira Hernández ◽  
Sandra Sumalla Cano

This work, developed under the NUTRIX Project, has the objective to develop artificial intelligence algorithms based on the open source platform Knime that allows to characterize and predict the adherence of individuals to diet before starting the treatment. The machine learning algorithms developed under this project have significantly increased the confidence (a priory probability) that a patient leaves the treatment (diet) before starting: from 17,6% up to 96,5% which can be used as valuable guidance during the decision-making process of professionals in the area of ​dietetics and nutrition.


Intelligent technology has touched and improved upon almost every aspect of employee life cycle, Human resource is one of the areas, which has greatly benefited. Transformation of work mainly question the way we work, where we work, how we work and mainly care about the environment and surroundings in which we work. The main goal is to support the organizations to break out their traditional way of work and further move towards to an environment, which brings more pleasing atmosphere, flexible, empowering and communicative. Machine learning, algorithms and artificial intelligence are the latest technology buzzing around the HR professional minds. Artificial intelligence designed to take decisions based on data fed into the programs. The key difference between rhythm and balance is of choice vs adjustment. The choice is made easier, only with the help of priority, quick decision-making, time and communication. To maintain the above scenario digitalisation plays a vital role. In this paper, we suggest the artificial assistants focus on improving the rhythm of individual


Author(s):  
Deeksha Kaul ◽  
Harika Raju ◽  
B. K. Tripathy

In this chapter, the authors discuss the use of quantum computing concepts to optimize the decision-making capability of classical machine learning algorithms. Machine learning, a subfield of artificial intelligence, implements various techniques to train a computer to learn and adapt to various real-time tasks. With the volume of data exponentially increasing, solving the same problems using classical algorithms becomes more tedious and time consuming. Quantum computing has varied applications in many areas of computer science. One such area which has been transformed a lot through the introduction of quantum computing is machine learning. Quantum computing, with its ability to perform tasks in logarithmic time, aids in overcoming the limitations of classical machine learning algorithms.


Sign in / Sign up

Export Citation Format

Share Document