scholarly journals Privacy-Aware Online Task Offloading for Mobile-Edge Computing

2021 ◽  
Vol 2021 ◽  
pp. 1-16
Author(s):  
Dali Zhu ◽  
Ting Li ◽  
Haitao Liu ◽  
Jiyan Sun ◽  
Liru Geng ◽  
...  

Mobile edge computing (MEC) has been envisaged as one of the most promising technologies in the fifth generation (5G) mobile networks. It allows mobile devices to offload their computation-demanding and latency-critical tasks to the resource-rich MEC servers. Accordingly, MEC can significantly improve the latency performance and reduce energy consumption for mobile devices. Nonetheless, privacy leakage may occur during the task offloading process. Most existing works ignored these issues or just investigated the system-level solution for MEC. Privacy-aware and user-level task offloading optimization problems receive much less attention. In order to tackle these challenges, a privacy-preserving and device-managed task offloading scheme is proposed in this paper for MEC. This scheme can achieve near-optimal latency and energy performance while protecting the location privacy and usage pattern privacy of users. Firstly, we formulate the joint optimization problem of task offloading and privacy preservation as a semiparametric contextual multi-armed bandit (MAB) problem, which has a relaxed reward model. Then, we propose a privacy-aware online task offloading (PAOTO) algorithm based on the transformed Thompson sampling (TS) architecture, through which we can (1) receive the best possible delay and energy consumption performance, (2) achieve the goal of preserving privacy, and (3) obtain an online device-managed task offloading policy without requiring any system-level information. Simulation results demonstrate that the proposed scheme outperforms the existing methods in terms of minimizing the system cost and preserving the privacy of users.

Algorithms ◽  
2019 ◽  
Vol 12 (2) ◽  
pp. 48 ◽  
Author(s):  
Ming Zhao ◽  
Ke Zhou

Mobile Edge Computing (MEC) is an innovative technique, which can provide cloud-computing near mobile devices on the edge of networks. Based on the MEC architecture, this paper proposes an ARIMA-BP-based Selective Offloading (ABSO) strategy, which minimizes the energy consumption of mobile devices while meeting the delay requirements. In ABSO, we exploit an ARIMA-BP model for estimating computation capacity of the edge cloud, and then design a Selective Offloading Algorithm for obtaining offloading strategy. Simulation results reveal that the ABSO can apparently decrease the energy consumption of mobile devices in comparison with other offloading methods.


2019 ◽  
Vol 62 (10) ◽  
pp. 1450-1462 ◽  
Author(s):  
Zikai Zhang ◽  
Jigang Wu ◽  
Long Chen ◽  
Guiyuan Jiang ◽  
Siew-Kei Lam

AbstractThe task offloading problem, which aims to balance the energy consumption and latency for Mobile Edge Computing (MEC), is still a challenging problem due to the dynamic changing system environment. To reduce energy while guaranteeing delay constraint for mobile applications, we propose an access control management architecture for 5G heterogeneous network by making full use of Base Station’s storage capability and reusing repetitive computational resource for tasks. For applications that rely on real-time information, we propose two algorithms to offload tasks with consideration of both energy efficiency and computation time constraint. For the first scenario, i.e. the rarely changing system environment, an optimal static algorithm is proposed based on dynamic programming technique to get the exact solution. For the second scenario, i.e. the frequently changing system environment, a two-stage online algorithm is proposed to adaptively obtain the current optimal solution in real time. Simulation results demonstrate that the exact algorithm in the first scenario runs 4 times faster than the enumeration method. In the second scenario, the proposed online algorithm can reduce the energy consumption and computation time violation rate by 16.3% and 25% in comparison with existing methods.


2022 ◽  
Author(s):  
Bin Xu ◽  
Tao Deng ◽  
Yichuan Liu ◽  
Yunkai Zhao ◽  
Zipeng Xu ◽  
...  

Abstract The combination of idle computing resources in mobile devices and the computing capacity of mobile edge servers enables all available devices in an edge network to complete all computing tasks in coordination to effectively improve the computing capacity of the edge network. This is a research hotspot for 5G technology applications. Previous research has focused on the minimum energy consumption and/or delay to determine the formulation of the computational offloading strategy but neglected the cost required for the computation of collaborative devices (mobile devices, mobile edge servers, etc.); therefore, we proposed a cost-based collaborative computation offloading model. In this model, when a task requests these devices' assistance in computing, it needs to pay the corresponding calculation cost; and on this basis, the task is offloaded and computed. In addition, for the model, we propose an adaptive neighborhood search based on simulated annealing algorithm (ANSSA) to jointly optimize the offloading decision and resource allocation with the goal of minimizing the sum of both the energy consumption and calculation cost. The adaptive mechanism enables different operators to update the probability of selection according to historical experience and environmental perception, which makes the individual evolution have certain autonomy. A large number of experiments conducted on different scales of mobile user instances show that the ANSSA can obtain satisfactory time performance with guaranteed solution quality. The experimental results demonstrate the superiority of the mobile edge computing (MEC) offloading system. It is of great significance to strike a balance between maintaining the life cycle of smart mobile devices and breaking the performance bottleneck of MEC servers.


2022 ◽  
Vol 2022 ◽  
pp. 1-13
Author(s):  
Ping Qi

Traditional intent recognition algorithms of intelligent prosthesis often use deep learning technology. However, deep learning’s high accuracy comes at the expense of high computational and energy consumption requirements. Mobile edge computing is a viable solution to meet the high computation and real-time execution requirements of deep learning algorithm on mobile device. In this paper, we consider the computation offloading problem of multiple heterogeneous edge servers in intelligent prosthesis scenario. Firstly, we present the problem definition and the detail design of MEC-based task offloading model for deep neural network. Then, considering the mobility of amputees, the mobility-aware energy consumption model and latency model are proposed. By deploying the deep learning-based motion intent recognition algorithm on intelligent prosthesis in a real-world MEC environment, the effectiveness of the task offloading and scheduling strategy is demonstrated. The experimental results show that the proposed algorithms can always find the optimal task offloading and scheduling decision.


Entropy ◽  
2021 ◽  
Vol 23 (9) ◽  
pp. 1146
Author(s):  
Peng Huang ◽  
Minjiang Deng ◽  
Zhiliang Kang ◽  
Qinshan Liu ◽  
Lijia Xu

Mobile edge computing (MEC) focuses on transferring computing resources close to the user's device, and it provides high-performance and low-delay services for mobile devices. It is an effective method to deal with computationally intensive and delay-sensitive tasks. Given the large number of underutilized computing resources for mobile devices in urban areas, leveraging these underutilized resources offers tremendous opportunities and value. Considering the spatiotemporal dynamics of user devices, the uncertainty of rich computing resources and the state of network channels in the MEC system, computing resource allocation in mobile devices with idle computing resources will affect the response time of task requesting. To solve these problems, this paper considers the case in which a mobile device can learn from a neighboring IoT device when offloading a computing request. On this basis, a novel self-adaptive learning of task offloading algorithm (SAda) is designed to minimize the average offloading delay in the MEC system. SAda adopts a distributed working mode and has a perception function to adapt to the dynamic environment in reality; it does not require frequent access to equipment information. Extensive simulations demonstrate that SAda achieves preferable latency performance and low learning error compared to the existing upper bound algorithms.


2021 ◽  
Vol 17 (6) ◽  
pp. 155014772110230
Author(s):  
Aamir Abbas ◽  
Ali Raza ◽  
Farhan Aadil ◽  
Muazzam Maqsood

With the recent advancements in communication technologies, the realization of computation-intensive applications like virtual/augmented reality, face recognition, and real-time video processing becomes possible at mobile devices. These applications require intensive computations for real-time decision-making and better user experience. However, mobile devices and Internet of things have limited energy and computational power. Executing such computationally intensive tasks on edge devices either leads to high computation latency or high energy consumption. Recently, mobile edge computing has been evolved and used for offloading these complex tasks. In mobile edge computing, Internet of things devices send their tasks to edge servers, which in turn perform fast computation. However, many Internet of things devices and edge server put an upper limit on concurrent task execution. Moreover, executing a very small size task (1 KB) over an edge server causes increased energy consumption due to communication. Therefore, it is required to have an optimal selection for tasks offloading such that the response time and energy consumption will become minimum. In this article, we proposed an optimal selection of offloading tasks using well-known metaheuristics, ant colony optimization algorithm, whale optimization algorithm, and Grey wolf optimization algorithm using variant design of these algorithms according to our problem through mathematical modeling. Executing multiple tasks at the server tends to provide high response time that leads to overloading and put additional latency at task computation. We also graphically represent the tradeoff between energy and delay that, how both parameters are inversely proportional to each other, using values from simulation. Results show that Grey wolf optimization outperforms the others in terms of optimizing energy consumption and execution latency while selected optimal set of offloading tasks.


Sign in / Sign up

Export Citation Format

Share Document