scholarly journals Truss assembly by space robot and task error recovery via reinforcement learning

Author(s):  
K. Senda ◽  
T. Matsumoto
2018 ◽  
Author(s):  
Minryung R. Song ◽  
Sang Wan Lee

AbstractDopamine activity may transition between two patterns: phasic responses to reward-predicting cues and ramping activity arising when an agent approaches the reward. However, when and why dopamine activity transitions between these modes is not understood. We hypothesize that the transition between ramping and phasic patterns reflects resource allocation which addresses the task dimensionality problem during reinforcement learning (RL). By parsimoniously modifying a standard temporal difference (TD) learning model to accommodate a mixed presentation of both experimental and environmental stimuli, we simulated dopamine transitions and compared it with experimental data from four different studies. The results suggested that dopamine transitions from ramping to phasic patterns as the agent narrows down candidate stimuli for the task; the opposite occurs when the agent needs to re-learn candidate stimuli due to a value change. These results lend insight into how dopamine deals with the tradeoff between cognitive resource and task dimensionality during RL.


2011 ◽  
Vol 23 (6) ◽  
pp. 939-950 ◽  
Author(s):  
Rintaro Haraguchi ◽  
◽  
Yukiyasu Domae ◽  
Koji Shiratsuchi ◽  
Yasuo Kitaaki ◽  
...  

To realize automatic robot-based electrical and electronic product assembly, we developed the handling of cables with connectors - flexible goods which are an obstacle to automation. The element technologies we developed include 3D vision sensing for cable extraction, force control for connector insertion, error recovery for improving system stability, and task-level programming for quick system start-up. We assembled FA control equipment to verify the feasibility of our developments.


2009 ◽  
Vol 3 (6) ◽  
pp. 671-680 ◽  
Author(s):  
Tetsuya Morizono ◽  
◽  
Yoji Yamada ◽  
Masatake Higashi ◽  
◽  
...  

Controlling “feel” when operating a power-assist robot is important for improving robot operability, user satisfaction, and task performance efficiency. Autonomous adjustment of “feel” is considered with robots under impedance control, and reinforcement learning in adjustment when a task includes repetitive positioning is discussed. Experimental results demonstrate that an operational “feel” pattern appropriate for positioning at a goal is developed by adjustment. Adjustment assuming a single fixed goal is expanded to cases including multiple goals, in which it is assumed that one goal is chosen by a user in real time. To adjust operational “feel” to individual goals, an algorithm infers the goal. The same result as that for a single fixed goal is obtained in experiments, but experimental results suggest that design must be improved to where the accuracy of inference to the goal is taken into account by the adjustment learning algorithm.


Sign in / Sign up

Export Citation Format

Share Document