Poster #S253 SHOW ME THE MONEY REVISITED: MONETARY REINFORCEMENT VERSUS INTRINSIC REWARD FOR LEARNING IN SCHIZOPHRENIA

2014 ◽  
Vol 153 ◽  
pp. S181
Author(s):  
Jimmy Choi ◽  
Joanna Fiszdon ◽  
Steven Silverstein ◽  
Deanna M. Barch
2020 ◽  
Vol 3 (1) ◽  

The aim of this study is to investigate the relationship between extrinsic and intrinsic reward on retention among Gen Y employees in Malaysian manufacturing companies. The data was collected from 113 respondents worked in manufacturing companies located in Seri Kembangan, Selangor using questionnaires. Multiple regression analysis was conducted to test the hypotheses. The results showed both extrinsic and intrinsic reward are the factors influencing retaining Gen Y in manufacturing companies. The discussion on the analysis, limitation of the study, recommendation for future research and conclusion were discussed at the end of this study. In a nutshell, it was proven extrinsic reward and intrinsic reward has contributed to the retention of Gen Y employees.


1983 ◽  
Vol 57 (3_suppl) ◽  
pp. 1255-1262 ◽  
Author(s):  
Michael A. Persinger

Mystical and religious experiences are hypothesized to be evoked by transient, electrical microseizures within deep structures of the temporal lobe. Although experiential details are affected by context and reinforcement history, basic themes reflect the inclusion of different amygdaloid-hippocampal structures and adjacent cortices. Whereas the unusual electrical coherence allows access to infantile memories of parents, a source of god expectations, specific stimulation evokes out-of-body experiences, space-time distortions, intense meaningfulness, and dreamy scenes. The species-specific similarities in temporal lobe properties enhance the homogeneity of cross-cultural experiences. They exist along a continuum that ranges from “early morning highs” to recurrent bouts of conversion and dominating religiosity. Predisposing factors include any biochemical or genetic factors that produce temporal lobe lability. A variety of precipitating stimuli provoke these experiences, but personal (life) crises and death bed conditions are optimal. These temporal lobe microseizures can be learned as responses to existential trauma because stimulation is of powerful intrinsic reward regions and reduction of death anxiety occurs. The implications of these transients as potent modifiers of human behavior are considered.


Author(s):  
Nicolas Bougie ◽  
Ryutaro Ichise

Deep reinforcement learning (DRL) methods traditionally struggle with tasks where environment rewards are sparse or delayed, which entails that exploration remains one of the key challenges of DRL. Instead of solely relying on extrinsic rewards, many state-of-the-art methods use intrinsic curiosity as exploration signal. While they hold promise of better local exploration, discovering global exploration strategies is beyond the reach of current methods. We propose a novel end-to-end intrinsic reward formulation that introduces high-level exploration in reinforcement learning. Our curiosity signal is driven by a fast reward that deals with local exploration and a slow reward that incentivizes long-time horizon exploration strategies. We formulate curiosity as the error in an agent’s ability to reconstruct the observations given their contexts. Experimental results show that this high-level exploration enables our agents to outperform prior work in several Atari games.


Complexity ◽  
2021 ◽  
Vol 2021 ◽  
pp. 1-9
Author(s):  
Huale Li ◽  
Rui Cao ◽  
Xuan Wang ◽  
Xiaohan Hou ◽  
Tao Qian ◽  
...  

In recent years, deep reinforcement learning (DRL) achieves great success in many fields, especially in the field of games, such as AlphaGo, AlphaZero, and AlphaStar. However, due to the reward sparsity problem, the traditional DRL-based method shows limited performance in 3D games, which contain much higher dimension of state space. To solve this problem, in this paper, we propose an intrinsic-based policy optimization (IBPO) algorithm for reward sparsity. In the IBPO, a novel intrinsic reward is integrated into the value network, which provides an additional reward in the environment with sparse reward, so as to accelerate the training. Besides, to deal with the problem of value estimation bias, we further design three types of auxiliary tasks, which can evaluate the state value and the action more accurately in 3D scenes. Finally, a framework of auxiliary intrinsic-based policy optimization (AIBPO) is proposed, which improves the performance of the IBPO. The experimental results show that the method is able to deal with the reward sparsity problem effectively. Therefore, the proposed method may be applied to real-world scenarios, such as 3-dimensional navigation and automatic driving, which can improve the sample utilization to reduce the cost of interactive sample collected by the real equipment.


1989 ◽  
Vol 65 (2) ◽  
pp. 515-520 ◽  
Author(s):  
K. Shah ◽  
C. M. Bradshaw ◽  
E. Szabadi

Four women pressed a button in five two-component concurrent variable-ratio variable-ratio ( conc VR VR) schedules of monetary reinforcement. There was no consistent tendency towards “probability matching” (distribution of responses between the two components in proportion to the relative probabilities of reinforcement); three of the four subjects showed exclusive preference for the schedule associated with the higher probability of reinforcement. These results are similar to results previously obtained with pigeons and rats in concurrent VR VR schedules.


2017 ◽  
Vol 35 (2) ◽  
pp. 163-180
Author(s):  
N. Pontus Leander ◽  
Aaron C. Kay ◽  
Tanya L. Chartrand ◽  
B. Keith Payne
Keyword(s):  

Author(s):  
Michael Dann ◽  
Fabio Zambetta ◽  
John Thangarajah

Sparse reward games, such as the infamous Montezuma’s Revenge, pose a significant challenge for Reinforcement Learning (RL) agents. Hierarchical RL, which promotes efficient exploration via subgoals, has shown promise in these games. However, existing agents rely either on human domain knowledge or slow autonomous methods to derive suitable subgoals. In this work, we describe a new, autonomous approach for deriving subgoals from raw pixels that is more efficient than competing methods. We propose a novel intrinsic reward scheme for exploiting the derived subgoals, applying it to three Atari games with sparse rewards. Our agent’s performance is comparable to that of state-of-the-art methods, demonstrating the usefulness of the subgoals found.


Sign in / Sign up

Export Citation Format

Share Document