Improving the Accuracy-Latency Trade-off of Edge-Cloud Computation Offloading for Deep Learning Services

Author(s):  
Xiaobo Zhao ◽  
Minoo Hosseinzadeh ◽  
Nathaniel Hudson ◽  
Hana Khamfroush ◽  
Daniel E. Lucani
Sensors ◽  
2019 ◽  
Vol 19 (19) ◽  
pp. 4307 ◽  
Author(s):  
Jungchan Cho ◽  
Minsik Lee

As artificial intelligence (AI)- or deep-learning-based technologies become more popular,the main research interest in the field is not only on their accuracy, but also their efficiency, e.g., theability to give immediate results on the users’ inputs. To achieve this, there have been many attemptsto embed deep learning technology on intelligent sensors. However, there are still many obstacles inembedding a deep network in sensors with limited resources. Most importantly, there is an apparenttrade-off between the complexity of a network and its processing time, and finding a structure witha better trade-off curve is vital for successful applications in intelligent sensors. In this paper, wepropose two strategies for designing a compact deep network that maintains the required level ofperformance even after minimizing the computations. The first strategy is to automatically determinethe number of parameters of a network by utilizing group sparsity and knowledge distillation (KD)in the training process. By doing so, KD can compensate for the possible losses in accuracy causedby enforcing sparsity. Nevertheless, a problem in applying the first strategy is the unclarity indetermining the balance between the accuracy improvement due to KD and the parameter reductionby sparse regularization. To handle this balancing problem, we propose a second strategy: a feedbackcontrol mechanism based on the proportional control theory. The feedback control logic determinesthe amount of emphasis to be put on network sparsity during training and is controlled based onthe comparative accuracy losses of the teacher and student models in the training. A surprising facthere is that this control scheme not only determines an appropriate trade-off point, but also improvesthe trade-off curve itself. The results of experiments on CIFAR-10, CIFAR-100, and ImageNet32 X 32datasets show that the proposed method is effective in building a compact network while preventingperformance degradation due to sparsity regularization much better than other baselines.


Sensors ◽  
2019 ◽  
Vol 19 (6) ◽  
pp. 1446 ◽  
Author(s):  
Liang Huang ◽  
Xu Feng ◽  
Luxin Zhang ◽  
Liping Qian ◽  
Yuan Wu

This paper studies mobile edge computing (MEC) networks where multiple wireless devices (WDs) offload their computation tasks to multiple edge servers and one cloud server. Considering different real-time computation tasks at different WDs, every task is decided to be processed locally at its WD or to be offloaded to and processed at one of the edge servers or the cloud server. In this paper, we investigate low-complexity computation offloading policies to guarantee quality of service of the MEC network and to minimize WDs’ energy consumption. Specifically, both a linear programing relaxation-based (LR-based) algorithm and a distributed deep learning-based offloading (DDLO) algorithm are independently studied for MEC networks. We further propose a heterogeneous DDLO to achieve better convergence performance than DDLO. Extensive numerical results show that the DDLO algorithms guarantee better performance than the LR-based algorithm. Furthermore, the DDLO algorithm generates an offloading decision in less than 1 millisecond, which is several orders faster than the LR-based algorithm.


Sensors ◽  
2021 ◽  
Vol 21 (23) ◽  
pp. 8072
Author(s):  
Yu-Bang Chang ◽  
Chieh Tsai ◽  
Chang-Hong Lin ◽  
Poki Chen

As the techniques of autonomous driving become increasingly valued and universal, real-time semantic segmentation has become very popular and challenging in the field of deep learning and computer vision in recent years. However, in order to apply the deep learning model to edge devices accompanying sensors on vehicles, we need to design a structure that has the best trade-off between accuracy and inference time. In previous works, several methods sacrificed accuracy to obtain a faster inference time, while others aimed to find the best accuracy under the condition of real time. Nevertheless, the accuracies of previous real-time semantic segmentation methods still have a large gap compared to general semantic segmentation methods. As a result, we propose a network architecture based on a dual encoder and a self-attention mechanism. Compared with preceding works, we achieved a 78.6% mIoU with a speed of 39.4 FPS with a 1024 × 2048 resolution on a Cityscapes test submission.


Sign in / Sign up

Export Citation Format

Share Document