scholarly journals SurRoL: An Open-source Reinforcement Learning Centered and dVRK Compatible Platform for Surgical Robot Learning

Author(s):  
Jiaqi Xu ◽  
Bin Li ◽  
Bo Lu ◽  
Yun-Hui Liu ◽  
Qi Dou ◽  
...  
AI Magazine ◽  
2013 ◽  
Vol 34 (3) ◽  
pp. 89-92 ◽  
Author(s):  
Julian Togelius ◽  
Noor Shaker ◽  
Sergey Karakovskiy ◽  
Georgios N. Yannakakis

We give a brief overview of the Mario AI Championship, a series of competitions based on an open source clone of the seminal platform game Super Mario Bros. The competition has four tracks. The gameplay and learning tracks resemble traditional reinforcement learning competitions, the Level generation track focuses on the generation of entertaining game levels, and the Turing Test track focuses on humanlike game-playing behavior. We also outline some lessons learned from the competition and its future. The article is written by the four organizers of the competition.


2012 ◽  
Vol 151 ◽  
pp. 498-502
Author(s):  
Jin Xue Zhang ◽  
Hai Zhu Pan

This paper is concerned with Q-learning , a very popular algorithm for reinforcement learning ,for obstacle avoidance through neural networks. The principle tells that the focus always must be on both ecological nice tasks and behaviours when designing on robot. Many robot systems have used behavior-based systems since the 1980’s.In this paper, the Khepera robot is trained through the proposed algorithm of Q-learning using the neural networks for the task of obstacle avoidance. In experiments with real and simulated robots, the neural networks approach can be used to make it possible for Q-learning to handle changes in the environment.


atp magazin ◽  
2020 ◽  
Vol 62 (11-12) ◽  
pp. 50-57
Author(s):  
Arne Wahrburg ◽  
Kim Listmann ◽  
Nima Enayati ◽  
René Kirsten

Das Thema „Robot Learning“ erfährt in der akademischen Welt zurzeit große Aufmerksamkeit, insbesondere die Anwendungvon Reinforcement Learning in der Robotik. Aus industrieller Sicht versprechen die rasanten Fortschritte im Bereich Robot Learning verkürzte Inbetriebnahmezeiten, vereinfachte  Programmierung, höhere Produktivität und Kostenreduktionen. In diesem Beitragwird das Potenzial der Technologie hinsichtlich industrieller Anwendungen beleuchtet. Es werden wesentliche Herausforderungen herausgestellt und mögliche Ansätze diskutiert, wie Ergebnisse aus dem Bereich Robot Learning in Richtung industrieller Anwendbarkeit getrieben werden können.


Author(s):  
John Aslanides ◽  
Jan Leike ◽  
Marcus Hutter

Many state-of-the-art reinforcement learning (RL) algorithms typically assume that the environment is an ergodic Markov Decision Process (MDP). In contrast, the field of universal reinforcement learning (URL) is concerned with algorithms that make as few assumptions as possible about the environment. The universal Bayesian agent AIXI and a family of related URL algorithms have been developed in this setting. While numerous theoretical optimality results have been proven for these agents, there has been no empirical investigation of their behavior to date. We present a short and accessible survey of these URL algorithms under a unified notation and framework, along with results of some experiments that qualitatively illustrate some properties of the resulting policies, and their relative performance on partially-observable gridworld environments. We also present an open- source reference implementation of the algorithms which we hope will facilitate further understanding of, and experimentation with, these ideas.


2002 ◽  
Vol 1 (1) ◽  
pp. 93-100
Author(s):  
Zhou Changjiu ◽  
Meng Qingchun ◽  
Guo Zhongwen ◽  
Qu Wiefen ◽  
Yin Bo

2020 ◽  
Vol 17 (3) ◽  
pp. 172988142091374
Author(s):  
Alexander Fabisch ◽  
Malte Langosz ◽  
Frank Kirchner

Reinforcement learning and behavior optimization are becoming more and more popular in the field of robotics because algorithms are mature enough to tackle real problems in this domain. Robust implementations of state-of-the-art algorithms are often not publicly available though, and experiments are hardly reproducible because open-source implementations are often not available or are still in a stage of research code. Consequently, often it is infeasible to deploy these algorithms on robotic systems. BOLeRo closes this gap for policy search and evolutionary algorithms by delivering open-source implementations of behavior learning algorithms for robots. It is easy to integrate in robotic middlewares and it can be used to compare methods and develop prototypes in simulation.


2019 ◽  
Vol 150 ◽  
pp. 162-170 ◽  
Author(s):  
Armando Plasencia ◽  
Yulia Shichkina ◽  
Ileana Suárez ◽  
Zoila Ruiz

Sign in / Sign up

Export Citation Format

Share Document