Context-aware Adaptive Surgery: A Fast and E?ective Framework for Adaptative Model Partition

Hongli Wang, Bin Guo, Jiaqi Liu, Sicong Liu, Yungang Wu, Zhiwen Yu

科研成果: 期刊稿件文章同行评审

21 引用 (Scopus)

摘要

Deep Neural Networks (DNNs) have made massive progress in many fields and deploying DNNs on end devices has become an emerging trend to make intelligence closer to users. However, it is challenging to deploy large-scale and computation-intensive DNNs on resource-constrained end devices due to their small size and lightweight. To this end, model partition, which aims to partition DNNs into multiple parts to realize the collaborative computing of multiple devices, has received extensive research attention. To find the optimal partition, most existing approaches need to run from scratch under given resource constraints. However, they ignore that resources of devices (e.g., storage, battery power), and performance requirements (e.g., inference latency), are often continuously changing, making the optimal partition solution change constantly during processing. Therefore, it is very important to reduce the tuning latency of model partition to realize the real-time adaption under the changing processing context. To address these problems, we propose the Context-aware Adaptive Surgery (CAS) framework to actively perceive the changing processing context, and adaptively find the appropriate partition solution in real-time. Specifically, we construct the partition state graph to comprehensively model different partition solutions of DNNs by import context resources. Then "the neighbor effect"is proposed, which provides the heuristic rule for the search process. When the processing context changes, CAS adopts the runtime search algorithm, Graph-based Adaptive DNN Surgery (GADS), to quickly find the appropriate partition that satisfies resource constraints under the guidance of the neighbor effect. The experimental results show that CAS realizes adaptively rapid tuning of the model partition solutions in 10ms scale even for large DNNs (2.25x to 221.7x search time improvement than the state-of-the-art researches), and the total inference latency still keeps the same level with baselines.

源语言英语
文章编号3478073
期刊Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies
5
3
DOI
出版状态已出版 - 9月 2021

指纹

探究 'Context-aware Adaptive Surgery: A Fast and E?ective Framework for Adaptative Model Partition' 的科研主题。它们共同构成独一无二的指纹。

引用此