learning agents
Recently Published Documents


TOTAL DOCUMENTS

445
(FIVE YEARS 178)

H-INDEX

21
(FIVE YEARS 3)

Author(s):  
Felipe Leno da Silva ◽  
Patrick MacAlpine ◽  
Roxana Rădulescu ◽  
Fernando P. Santos ◽  
Patrick Mannion

PLoS ONE ◽  
2022 ◽  
Vol 17 (1) ◽  
pp. e0262505
Author(s):  
Simon Carrignon ◽  
R. Alexander Bentley ◽  
Matthew Silk ◽  
Nina H. Fefferman

The global pandemic of COVID-19 revealed the dynamic heterogeneity in how individuals respond to infection risks, government orders, and community-specific social norms. Here we demonstrate how both individual observation and social learning are likely to shape behavioral, and therefore epidemiological, dynamics over time. Efforts to delay and reduce infections can compromise their own success, especially when disease risk and social learning interact within sub-populations, as when people observe others who are (a) infected and/or (b) socially distancing to protect themselves from infection. Simulating socially-learning agents who observe effects of a contagious virus, our modelling results are consistent with with 2020 data on mask-wearing in the U.S. and also concur with general observations of cohort induced differences in reactions to public health recommendations. We show how shifting reliance on types of learning affect the course of an outbreak, and could therefore factor into policy-based interventions incorporating age-based cohort differences in response behavior.


2022 ◽  
Vol 119 (1) ◽  
pp. e2107431118
Author(s):  
Gautam Reddy ◽  
Boris I. Shraiman ◽  
Massimo Vergassola

Ants, mice, and dogs often use surface-bound scent trails to establish navigation routes or to find food and mates, yet their tracking strategies remain poorly understood. Chemotaxis-based strategies cannot explain casting, a characteristic sequence of wide oscillations with increasing amplitude performed upon sustained loss of contact with the trail. We propose that tracking animals have an intrinsic, geometric notion of continuity, allowing them to exploit past contacts with the trail to form an estimate of where it is headed. This estimate and its uncertainty form an angular sector, and the emergent search patterns resemble a “sector search.” Reinforcement learning agents trained to execute a sector search recapitulate the various phases of experimentally observed tracking behavior. We use ideas from polymer physics to formulate a statistical description of trails and show that search geometry imposes basic limits on how quickly animals can track trails. By formulating trail tracking as a Bellman-type sequential optimization problem, we quantify the geometric elements of optimal sector search strategy, effectively explaining why and when casting is necessary. We propose a set of experiments to infer how tracking animals acquire, integrate, and respond to past information on the tracked trail. More generally, we define navigational strategies relevant for animals and biomimetic robots and formulate trail tracking as a behavioral paradigm for learning, memory, and planning.


2021 ◽  
Vol 11 (3-4) ◽  
pp. 1-35
Author(s):  
Jonathan Dodge ◽  
Roli Khanna ◽  
Jed Irvine ◽  
Kin-ho Lam ◽  
Theresa Mai ◽  
...  

Explainable AI is growing in importance as AI pervades modern society, but few have studied how explainable AI can directly support people trying to assess an AI agent. Without a rigorous process, people may approach assessment in ad hoc ways—leading to the possibility of wide variations in assessment of the same agent due only to variations in their processes. AAR, or After-Action Review, is a method some military organizations use to assess human agents, and it has been validated in many domains. Drawing upon this strategy, we derived an After-Action Review for AI (AAR/AI), to organize ways people assess reinforcement learning agents in a sequential decision-making environment. We then investigated what AAR/AI brought to human assessors in two qualitative studies. The first investigated AAR/AI to gather formative information, and the second built upon the results, and also varied the type of explanation (model-free vs. model-based) used in the AAR/AI process. Among the results were the following: (1) participants reporting that AAR/AI helped to organize their thoughts and think logically about the agent, (2) AAR/AI encouraged participants to reason about the agent from a wide range of perspectives , and (3) participants were able to leverage AAR/AI with the model-based explanations to falsify the agent’s predictions.


Entropy ◽  
2021 ◽  
Vol 24 (1) ◽  
pp. 58
Author(s):  
Jerzy Balicki

Particle swarm optimization algorithm (PSO) is an effective metaheuristic that can determine Pareto-optimal solutions. We propose an extended PSO by introducing quantum gates in order to ensure the diversity of particle populations that are looking for efficient alternatives. The quality of solutions was verified in the issue of assignment of resources in the computing cloud to improve the live migration of virtual machines. We consider the multi-criteria optimization problem of deep learning-based models embedded into virtual machines. Computing clouds with deep learning agents can support several areas of education, smart city or economy. Because deep learning agents require lots of computer resources, seven criteria are studied such as electric power of hosts, reliability of cloud, CPU workload of the bottleneck host, communication capacity of the critical node, a free RAM capacity of the most loaded memory, a free disc memory capacity of the most busy storage, and overall computer costs. Quantum gates modify an accepted position for the current location of a particle. To verify the above concept, various simulations have been carried out on the laboratory cloud based on the OpenStack platform. Numerical experiments have confirmed that multi-objective quantum-inspired particle swarm optimization algorithm provides better solutions than the other metaheuristics.


Author(s):  
Kai Zhang ◽  
Zhongzheng Wang ◽  
Guodong Chen ◽  
Liming Zhang ◽  
Yongfei Yang ◽  
...  

Author(s):  
Antonio Serrano-Muñoz ◽  
Nestor Arana-Arexolaleiba ◽  
Dimitrios Chrysostomou ◽  
Simon Bøgh

AbstractRemanufacturing automation must be designed to be flexible and robust enough to overcome the uncertainties, conditions of the products, and complexities in the planning and operation of the processes. Machine learning methods, in particular reinforcement learning, are presented as techniques to learn, improve, and generalise the automation of many robotic manipulation tasks (most of them related to grasping, picking, or assembly). However, not much has been exploited in remanufacturing, in particular in disassembly tasks. This work presents the state of the art of contact-rich disassembly using reinforcement learning algorithms and a study about the generalisation of object extraction skills when applied to contact-rich disassembly tasks. The generalisation capabilities of two state-of-the-art reinforcement learning agents (trained in simulation) are tested and evaluated in simulation, and real world while perform a disassembly task. Results show that at least one of the agents can generalise the contact-rich extraction skill. Besides, this work identifies key concepts and gaps for the reinforcement learning algorithms’ research and application on disassembly tasks.


2021 ◽  
Author(s):  
Nikolaos Al. Papadopoulos ◽  
Marti Sanchez-Fibla

Multi-Agent Reinforcement Learning reductionist simulations can provide a spectrum of opportunities towards the modeling and understanding of complex social phenomena such as common-pool appropriation. In this paper, a multiplayer variant of Battle-of-the-Exes is suggested as appropriate for experimentation regarding fair and efficient coordination and turn-taking among selfish agents. Going beyond literature’s fairness and efficiency, a novel measure is proposed for turn-taking coordination evaluation, robust to the number of agents and episodes of a system. Six variants of this measure are defined, entitled Alternation Measures or ALT. ALT measures were found sufficient to capture the desired properties (alternation, fair and efficient distribution) in comparison to state-of-the-art measures, thus they were benchmarked and tested through a series of experiments with Reinforcement Learning agents, aspiring to contribute novel tools for a deeper understanding of emergent social outcomes.


2021 ◽  
Author(s):  
Romulo Freire Ferrer Filho ◽  
Yuri Lenon Barbosa Nogueira ◽  
Creto Augusto Vidal ◽  
Joaquim Bento Cavalcante-Neto ◽  
Paulo Bruno de Sousa Serafim

Sign in / Sign up

Export Citation Format

Share Document