scholarly journals The Archimedean trap: Why traditional reinforcement learning will probably not yield AGI

2020 ◽  
Vol 11 (1) ◽  
pp. 70-85
Author(s):  
Samuel Allen Alexander

AbstractAfter generalizing the Archimedean property of real numbers in such a way as to make it adaptable to non-numeric structures, we demonstrate that the real numbers cannot be used to accurately measure non-Archimedean structures. We argue that, since an agent with Artificial General Intelligence (AGI) should have no problem engaging in tasks that inherently involve non-Archimedean rewards, and since traditional reinforcement learning rewards are real numbers, therefore traditional reinforcement learning probably will not lead to AGI. We indicate two possible ways traditional reinforcement learning could be altered to remove this roadblock.

2021 ◽  
Author(s):  
Pamul Yadav ◽  
Taewoo Kim ◽  
Ho Suk ◽  
Junyong Lee ◽  
Hyeonseong Jeong ◽  
...  

<p>Faster adaptability to open-world novelties by intelligent agents is a necessary factor in achieving the goal of creating Artificial General Intelligence (AGI). Current RL framework does not considers the unseen changes (novelties) in the environment. Therefore, in this paper, we have proposed OODA-RL, a Reinforcement Learning based framework that can be used to develop robust RL algorithms capable of handling both the known environments as well as adaptation to the unseen environments. OODA-RL expands the definition of internal composition of the agent as compared to the abstract definition in the classical RL framework, allowing the RL researchers to incorporate novelty adaptation techniques as an add-on feature to the existing SoTA as well as yet-to-be-developed RL algorithms.</p>


2021 ◽  
Author(s):  
Pamul Yadav ◽  
Taewoo Kim ◽  
Ho Suk ◽  
Junyong Lee ◽  
Hyeonseong Jeong ◽  
...  

<p>Faster adaptability to open-world novelties by intelligent agents is a necessary factor in achieving the goal of creating Artificial General Intelligence (AGI). Current RL framework does not considers the unseen changes (novelties) in the environment. Therefore, in this paper, we have proposed OODA-RL, a Reinforcement Learning based framework that can be used to develop robust RL algorithms capable of handling both the known environments as well as adaptation to the unseen environments. OODA-RL expands the definition of internal composition of the agent as compared to the abstract definition in the classical RL framework, allowing the RL researchers to incorporate novelty adaptation techniques as an add-on feature to the existing SoTA as well as yet-to-be-developed RL algorithms.</p>


2021 ◽  
Author(s):  
Pamul Yadav ◽  
Taewoo Kim ◽  
Ho Suk ◽  
Junyong Lee ◽  
Hyeonseong Jeong ◽  
...  

<p>Faster adaptability to open-world novelties by intelligent agents is a necessary factor in achieving the goal of creating Artificial General Intelligence (AGI). Current RL framework does not considers the unseen changes (novelties) in the environment. Therefore, in this paper, we have proposed OODA-RL, a Reinforcement Learning based framework that can be used to develop robust RL algorithms capable of handling both the known environments as well as adaptation to the unseen environments. OODA-RL expands the definition of internal composition of the agent as compared to the abstract definition in the classical RL framework, allowing the RL researchers to incorporate novelty adaptation techniques as an add-on feature to the existing SoTA as well as yet-to-be-developed RL algorithms.</p>


Order ◽  
2021 ◽  
Author(s):  
Péter Vrana

AbstractGiven a commutative semiring with a compatible preorder satisfying a version of the Archimedean property, the asymptotic spectrum, as introduced by Strassen (J. reine angew. Math. 1988), is an essentially unique compact Hausdorff space together with a map from the semiring to the ring of continuous functions. Strassen’s theorem characterizes an asymptotic relaxation of the preorder that asymptotically compares large powers of the elements up to a subexponential factor as the pointwise partial order of the corresponding functions, realizing the asymptotic spectrum as the space of monotone semiring homomorphisms to the nonnegative real numbers. Such preordered semirings have found applications in complexity theory and information theory. We prove a generalization of this theorem to preordered semirings that satisfy a weaker polynomial growth condition. This weaker hypothesis does not ensure in itself that nonnegative real-valued monotone homomorphisms characterize the (appropriate modification of the) asymptotic preorder. We find a sufficient condition as well as an equivalent condition for this to hold. Under these conditions the asymptotic spectrum is a locally compact Hausdorff space satisfying a similar universal property as in Strassen’s work.


2021 ◽  
pp. 027836492098785
Author(s):  
Julian Ibarz ◽  
Jie Tan ◽  
Chelsea Finn ◽  
Mrinal Kalakrishnan ◽  
Peter Pastor ◽  
...  

Deep reinforcement learning (RL) has emerged as a promising approach for autonomously acquiring complex behaviors from low-level sensor observations. Although a large portion of deep RL research has focused on applications in video games and simulated control, which does not connect with the constraints of learning in real environments, deep RL has also demonstrated promise in enabling physical robots to learn complex skills in the real world. At the same time, real-world robotics provides an appealing domain for evaluating such algorithms, as it connects directly to how humans learn: as an embodied agent in the real world. Learning to perceive and move in the real world presents numerous challenges, some of which are easier to address than others, and some of which are often not considered in RL research that focuses only on simulated domains. In this review article, we present a number of case studies involving robotic deep RL. Building off of these case studies, we discuss commonly perceived challenges in deep RL and how they have been addressed in these works. We also provide an overview of other outstanding challenges, many of which are unique to the real-world robotics setting and are not often the focus of mainstream RL research. Our goal is to provide a resource both for roboticists and machine learning researchers who are interested in furthering the progress of deep RL in the real world.


2021 ◽  
pp. 1-6
Author(s):  
Scott McLean ◽  
Gemma J. M. Read ◽  
Jason Thompson ◽  
P. A. Hancock ◽  
Paul M. Salmon

1995 ◽  
Vol 38 (2) ◽  
pp. 223-229
Author(s):  
John Lindsay Orr

AbstractA linearly ordered set A is said to shuffle into another linearly ordered set B if there is an order preserving surjection A —> B such that the preimage of each member of a cofinite subset of B has an arbitrary pre-defined finite cardinality. We show that every countable linearly ordered set shuffles into itself. This leads to consequences on transformations of subsets of the real numbers by order preserving maps.


Sign in / Sign up

Export Citation Format

Share Document