计算机科学
边缘计算
移动边缘计算
推论
分布式计算
调度(生产过程)
初始化
边缘设备
移动计算
移动设备
计算卸载
GSM演进的增强数据速率
云计算
人工智能
计算机网络
程序设计语言
经济
运营管理
操作系统
作者
Zhicheng Liu,Jinduo Song,Chao Qiu,Xiaofei Wang,Xu Chen,Qiang He,Hao Sheng
标识
DOI:10.1109/tmc.2022.3218724
摘要
As the primary driver of intelligent mobile applications, deep neural networks (DNNs) have gradually deployed to millions of mobile devices, producing massive latency-sensitive and computation-intensive tasks daily. Mobile edge computing facilitates the deployment of computing resources at the edge, which enables fine-grained offloading of DNN inference tasks from mobile devices to edge nodes. However, most existing studies have not systematically considered three crucial performance aspects: scheduling multiple streams of DNN inference tasks, leveraging multi-exit models to hasten task processing, and partitioning inference models for partial offloading. To this end, this paper proposes an adaptive inference framework in mobile edge computing, which can dynamically select the exit point and partition point for multiple inference task streams. We design a dynamic programming algorithm to obtain an efficient solution under the ideal condition that task arrival information is known. Further, we design a learning-based algorithm for online scheduling, whose training efficiency is improved based on historical experience initialization and priority experience replay. Experimental results show that compared with the Greedy algorithm, the online algorithm improves the performance on two environmental parameters by an average of 5.9% and 32%, respectively.
科研通智能强力驱动
Strongly Powered by AbleSci AI