scholarly journals Reconfigurable Intelligent Surface-assisted Edge Computing to Minimize Delay in Task Offloading

Author(s):  
vikas kumar ◽  
Mithun Mukherjee

The advantage of computational resources in edge computing near the data source has kindled growing interest in delay-sensitive Internet of Things (IoT) applications. However, the benefit of the edge server is limited by the uploading and downloading links between end-users and edge servers when these end-users seek computational resources from edge servers. The scenario becomes more severe when the user-end's devices are in the shaded region resulting in low uplink/downlink quality. In this paper, we consider a reconfigurable intelligent surface (RIS)-assisted edge computing system, where the benefits of RIS are exploited to improve the uploading transmission rate. We further aim to minimize the delay of worst-case in the network when the end-users either compute task data in their local CPU or offload task data to the edge server. Next, we optimize the uploading bandwidth allocation for every end-user's task data to minimize the maximum delay in the network. The above optimization problem is formulated as quadratically constrained quadratic programming. Afterward, we solve this problem by semidefinite relaxation. Finally, the simulation results demonstrate that the proposed strategy is scalable under various network settings.

2021 ◽  
Author(s):  
vikas kumar ◽  
Mithun Mukherjee

The advantage of computational resources in edge computing near the data source has kindled growing interest in delay-sensitive Internet of Things (IoT) applications. However, the benefit of the edge server is limited by the uploading and downloading links between end-users and edge servers when these end-users seek computational resources from edge servers. The scenario becomes more severe when the user-end's devices are in the shaded region resulting in low uplink/downlink quality. In this paper, we consider a reconfigurable intelligent surface (RIS)-assisted edge computing system, where the benefits of RIS are exploited to improve the uploading transmission rate. We further aim to minimize the delay of worst-case in the network when the end-users either compute task data in their local CPU or offload task data to the edge server. Next, we optimize the uploading bandwidth allocation for every end-user's task data to minimize the maximum delay in the network. The above optimization problem is formulated as quadratically constrained quadratic programming. Afterward, we solve this problem by semidefinite relaxation. Finally, the simulation results demonstrate that the proposed strategy is scalable under various network settings.


Author(s):  
Yong Xiao ◽  
Ling Wei ◽  
Junhao Feng ◽  
Wang En

Edge computing has emerged for meeting the ever-increasing computation demands from delay-sensitive Internet of Things (IoT) applications. However, the computing capability of an edge device, including a computing-enabled end user and an edge server, is insufficient to support massive amounts of tasks generated from IoT applications. In this paper, we aim to propose a two-tier end-edge collaborative computation offloading policy to support as much as possible computation-intensive tasks while making the edge computing system strongly stable. We formulate the two-tier end-edge collaborative offloading problem with the objective of minimizing the task processing and offloading cost constrained to the stability of queue lengths of end users and edge servers. We perform analysis of the Lyapunov drift-plus-penalty properties of the problem. Then, a cost-aware computation offloading (CACO) algorithm is proposed to find out optimal two-tier offloading decisions so as to minimize the cost while making the edge computing system stable. Our simulation results show that the proposed CACO outperforms the benchmarked algorithms, especially under various number of end users and edge servers.


Author(s):  
Jaber Almutairi ◽  
Mohammad Aldossary

AbstractRecently, the number of Internet of Things (IoT) devices connected to the Internet has increased dramatically as well as the data produced by these devices. This would require offloading IoT tasks to release heavy computation and storage to the resource-rich nodes such as Edge Computing and Cloud Computing. Although Edge Computing is a promising enabler for latency-sensitive related issues, its deployment produces new challenges. Besides, different service architectures and offloading strategies have a different impact on the service time performance of IoT applications. Therefore, this paper presents a novel approach for task offloading in an Edge-Cloud system in order to minimize the overall service time for latency-sensitive applications. This approach adopts fuzzy logic algorithms, considering application characteristics (e.g., CPU demand, network demand and delay sensitivity) as well as resource utilization and resource heterogeneity. A number of simulation experiments are conducted to evaluate the proposed approach with other related approaches, where it was found to improve the overall service time for latency-sensitive applications and utilize the edge-cloud resources effectively. Also, the results show that different offloading decisions within the Edge-Cloud system can lead to various service time due to the computational resources and communications types.


2021 ◽  
Vol 561 ◽  
pp. 70-80
Author(s):  
Guangshun Li ◽  
Xinrong Ren ◽  
Junhua Wu ◽  
Wanting Ji ◽  
Haili Yu ◽  
...  

Sensors ◽  
2021 ◽  
Vol 21 (14) ◽  
pp. 4798
Author(s):  
Fangni Chen ◽  
Anding Wang ◽  
Yu Zhang ◽  
Zhengwei Ni ◽  
Jingyu Hua

With the increasing deployment of IoT devices and applications, a large number of devices that can sense and monitor the environment in IoT network are needed. This trend also brings great challenges, such as data explosion and energy insufficiency. This paper proposes a system that integrates mobile edge computing (MEC) technology and simultaneous wireless information and power transfer (SWIPT) technology to improve the service supply capability of WSN-assisted IoT applications. A novel optimization problem is formulated to minimize the total system energy consumption under the constraints of data transmission rate and transmitting power requirements by jointly considering power allocation, CPU frequency, offloading weight factor and energy harvest weight factor. Since the problem is non-convex, we propose a novel alternate group iteration optimization (AGIO) algorithm, which decomposes the original problem into three subproblems, and alternately optimizes each subproblem using the group interior point iterative algorithm. Numerical simulations validate that the energy consumption of our proposed design is much lower than the two benchmark algorithms. The relationship between system variables and energy consumption of the system is also discussed.


Sign in / Sign up

Export Citation Format

Share Document