Towards Cooperative Caching for Vehicular Networks with Multi-level Federated Reinforcement Learning

Author(s):  
Lei Zhao ◽  
Yongyi Ran ◽  
Hao Wang ◽  
Junxia Wang ◽  
Jiangtao Luo
Author(s):  
Ning Zhao ◽  
Hao Wu ◽  
F. Richard Yu ◽  
Lifu Wang ◽  
Weiting Zhang ◽  
...  

2019 ◽  
Vol 9 (3) ◽  
pp. 502 ◽  
Author(s):  
Cristyan Gil ◽  
Hiram Calvo ◽  
Humberto Sossa

Programming robots for performing different activities requires calculating sequences of values of their joints by taking into account many factors, such as stability and efficiency, at the same time. Particularly for walking, state of the art techniques to approximate these sequences are based on reinforcement learning (RL). In this work we propose a multi-level system, where the same RL method is used first to learn the configuration of robot joints (poses) that allow it to stand with stability, and then in the second level, we find the sequence of poses that let it reach the furthest distance in the shortest time, while avoiding falling down and keeping a straight path. In order to evaluate this, we focus on measuring the time it takes for the robot to travel a certain distance. To our knowledge, this is the first work focusing both on speed and precision of the trajectory at the same time. We implement our model in a simulated environment using q-learning. We compare with the built-in walking modes of an NAO robot by improving normal-speed and enhancing robustness in fast-speed. The proposed model can be extended to other tasks and is independent of a particular robot model.


IEEE Network ◽  
2020 ◽  
Vol 34 (3) ◽  
pp. 57-63 ◽  
Author(s):  
Muzhou Xiong ◽  
Yuepeng Li ◽  
Lin Gu ◽  
Shengli Pan ◽  
Deze Zeng ◽  
...  

Algorithms ◽  
2018 ◽  
Vol 11 (9) ◽  
pp. 134 ◽  
Author(s):  
Gabriele Russo Russo ◽  
Matteo Nardelli ◽  
Valeria Cardellini ◽  
Francesco Lo Presti

The capability of efficiently processing the data streams emitted by nowadays ubiquitous sensing devices enables the development of new intelligent services. Data Stream Processing (DSP) applications allow for processing huge volumes of data in near real-time. To keep up with the high volume and velocity of data, these applications can elastically scale their execution on multiple computing resources to process the incoming data flow in parallel. Being that data sources and consumers are usually located at the network edges, nowadays the presence of geo-distributed computing resources represents an attractive environment for DSP. However, controlling the applications and the processing infrastructure in such wide-area environments represents a significant challenge. In this paper, we present a hierarchical solution for the autonomous control of elastic DSP applications and infrastructures. It consists of a two-layered hierarchical solution, where centralized components coordinate subordinated distributed managers, which, in turn, locally control the elastic adaptation of the application components and deployment regions. Exploiting this framework, we design several self-adaptation policies, including reinforcement learning based solutions. We show the benefits of the presented self-adaptation policies with respect to static provisioning solutions, and discuss the strengths of reinforcement learning based approaches, which learn from experience how to optimize the application performance and resource allocation.


Sign in / Sign up

Export Citation Format

Share Document