A novel approach of dynamic base station switching strategy based on Markov decision process for interference alignment in VANETs

2020 ◽  
Vol 26 (8) ◽  
pp. 5561-5578
Author(s):  
Chong Zhao ◽  
Jianghong Han ◽  
Xu Ding ◽  
Fan Yang
Electronics ◽  
2021 ◽  
Vol 10 (2) ◽  
pp. 190
Author(s):  
Wu Ouyang ◽  
Zhigang Chen ◽  
Jia Wu ◽  
Genghua Yu ◽  
Heng Zhang

As transportation becomes more convenient and efficient, users move faster and faster. When a user leaves the service range of the original edge server, the original edge server needs to migrate the tasks offloaded by the user to other edge servers. An effective task migration strategy needs to fully consider the location of users, the load status of edge servers, and energy consumption, which make designing an effective task migration strategy a challenge. In this paper, we innovatively proposed a mobile edge computing (MEC) system architecture consisting of multiple smart mobile devices (SMDs), multiple unmanned aerial vehicle (UAV), and a base station (BS). Moreover, we establish the model of the Markov decision process with unknown rewards (MDPUR) based on the traditional Markov decision process (MDP), which comprehensively considers the three aspects of the migration distance, the residual energy status of the UAVs, and the load status of the UAVs. Based on the MDPUR model, we propose a advantage-based value iteration (ABVI) algorithm to obtain the effective task migration strategy, which can help the UAV group to achieve load balancing and reduce the total energy consumption of the UAV group under the premise of ensuring user service quality. Finally, the results of simulation experiments show that the ABVI algorithm is effective. In particular, the ABVI algorithm has better performance than the traditional value iterative algorithm. And in a dynamic environment, the ABVI algorithm is also very robust.


Author(s):  
Abayomi Ajofoyinbo ◽  
◽  
David O. Olowokere ◽  
Oye Ibidapo-Obe

The paper presents N-element switchable beam antennas (BAs) system design for Wireless Sensor Node (WSN), in which the operation of the BAs was characterized by semi-Markov Decision Process (SMDP) with variable sojourn time. A matrix-based switching methodology was introduced for selecting an operational BA, based on the received signal power by each BA. Optimal analysis was carried-out to obtain optimal results in terms of the maximum total of sum of discounted reward in current states. Also developed in the study is the methodology for switching a BA from non-operational to operational state and vice-versa. The effectiveness of this switchable BAs system design was tested via numerical analysis implemented in MATLAB software. Numerical results show that this novel approach enables the WSN equipped with BAs to select and maintain an operational BA in receive (or transmit) mode for the entire duration of packets reception (or transmission). The authors found no paper in the existing literature that provides this capability.


Author(s):  
Kazuteru Miyazaki ◽  
◽  
Shigenobu Kobayashi ◽  

Exploitation-oriented learning (XoL) is a novel approach to goal-directed learning from interaction. Reinforcement learning is much more focused on learning and ensures optimality in Markov decision process (MDP) environments, XoL involves learning a rational policy that obtains rewards continuously and very quickly. PS-r*, a form of XoL, involves learning a useful rational policy not inferior to the random walk in the partially observed Markov decision process (POMDP) where reward types number one. PS-r*, however, requires O(MN2) memory where N is the number of sensory input types and M is an action. We propose PS-r#for learning a useful rational policy in the POMDP using O(MN) memory. PS-r#effectiveness is confirmed in numerical examples.


Mathematics ◽  
2021 ◽  
Vol 9 (12) ◽  
pp. 1385
Author(s):  
Irais Mora-Ochomogo ◽  
Marco Serrato ◽  
Jaime Mora-Vargas ◽  
Raha Akhavan-Tabatabaei

Natural disasters represent a latent threat for every country in the world. Due to climate change and other factors, statistics show that they continue to be on the rise. This situation presents a challenge for the communities and the humanitarian organizations to be better prepared and react faster to natural disasters. In some countries, in-kind donations represent a high percentage of the supply for the operations, which presents additional challenges. This research proposes a Markov Decision Process (MDP) model to resemble operations in collection centers, where in-kind donations are received, sorted, packed, and sent to the affected areas. The decision addressed is when to send a shipment considering the uncertainty of the donations’ supply and the demand, as well as the logistics costs and the penalty of unsatisfied demand. As a result of the MDP a Monotone Optimal Non-Decreasing Policy (MONDP) is proposed, which provides valuable insights for decision-makers within this field. Moreover, the necessary conditions to prove the existence of such MONDP are presented.


Sign in / Sign up

Export Citation Format

Share Document