Context-Aware Ranking by Constructing a Virtual Environment for Reinforcement Learning

Author(s):  
Junqi Zhang ◽  
Jiaxin Mao ◽  
Yiqun Liu ◽  
Ruizhe Zhang ◽  
Min Zhang ◽  
...  
2021 ◽  
Vol 47 ◽  
pp. 101229
Author(s):  
Dan E. Kröhling ◽  
Omar J.A. Chiotti ◽  
Ernesto C. Martínez

2021 ◽  
Vol 9 ◽  
Author(s):  
Peiran Xie ◽  
Guangming Zhang ◽  
Yuguang Niu ◽  
Tianshu Sun

The control of flue gas emission in thermal power plants has been a topic of concern. Selective catalytic reduction technology has been widely used as an effective flue gas treatment technology. However, precisely controlling the amount of ammonia injected remains a challenge. Too much ammonia not only causes secondary pollution but also corrodes the reactor equipment, while too little ammonia does not effectively reduce the NOx content. In recent years, deep reinforcement learning has achieved better results than traditional methods in decision making and control, which provides new methods for better control of selective catalytic reduction systems. The purpose of this research is to design an intelligent controller using reinforcement learning technology, which can accurately control ammonia injection, and achieve higher denitrification effect and less secondary pollution. To train the deep reinforcement learning controller, a high-precision virtual denitration environment is first constructed. In order to make the virtual environment more realistic, this virtual environment was designed as a special structure with two decoders and a unique approach was used in fitting the virtual environment. A deep deterministic policy agent is used as an intelligent controller to control the amount of injected ammonia. To make the intelligent controller more stable, the actor-critic framework and the experience pool approach were adopted. The results show that the intelligent controller can control the emissions of nitrogen oxides and ammonia at the outlet of the reactor after training in virtual environment.


2019 ◽  
Vol 19 (6) ◽  
pp. 657-678 ◽  
Author(s):  
Kamalakanta Sethi ◽  
E. Sai Rupesh ◽  
Rahul Kumar ◽  
Padmalochan Bera ◽  
Y. Venu Madhav

Author(s):  
Jacquelyne Forgette ◽  
Michael Katchabaw

A key challenge in programming virtual environments is to produce virtual characters that are autonomous and capable of action selections that appear believable. In this chapter, motivations are used as a basis for learning using reinforcements. With motives driving the decisions of characters, their actions will appear less structured and repetitious, and more human in nature. This will also allow developers to easily create virtual characters with specific motivations, based mostly on their narrative purposes or roles in the virtual world. With minimum and maximum desirable motive values, the characters use reinforcement learning to drive action selection to maximize their rewards across all motives. Experimental results show that a character can learn to satisfy as many as four motives, even with significantly delayed rewards, and motive changes that are caused by other characters in the world. While the actions tested are simple in nature, they show the potential of a more complicated motivation driven reinforcement learning system. The developer need only define a character's motivations, and the character will learn to act realistically over time in the virtual environment.


2021 ◽  
Author(s):  
Zhijun Tu ◽  
Jian Ma ◽  
Tian Xia ◽  
Wenzhe Zhao ◽  
Pengju Ren ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document