AdaKnife: Flexible DNN Offloading for Inference Acceleration on Heterogeneous Mobile Devices

Sicong Liu, Hao Luo, Xiao Chen Li, Yao Li, Bin Guo, Zhiwen Yu, Yu Zhan Wang, Ke Ma, Ya San Ding, Yuan Yao

科研成果: 期刊稿件文章同行评审

3 引用 (Scopus)

摘要

The integration of deep neural network (DNN) intelligence into embedded mobile devices is expanding rapidly, supporting a wide range of applications. DNN compression techniques, which adapt models to resource-constrained mobile environments, often force a trade-off between efficiency and accuracy. Distributed DNN inference, leveraging multiple mobile devices, emerges as a promising alternative to enhance inference efficiency without compromising accuracy. However, effectively decoupling DNN models into fine-grained components for optimal parallel acceleration presents significant challenges. Current partitioning methods, including layer-level and operator or channel-level partitioning, provide only partial solutions and struggle with the heterogeneous nature of DNN compilation frameworks, complicating direct model offloading. In response, we introduce AdaKnife, an adaptive framework for accelerated inference across heterogeneous mobile devices. AdaKnife enables on-demand mixed-granularity DNN partitioning via computational graph analysis, facilitates efficient cross-framework model transitions with operator optimization for offloading, and improves the feasibility of parallel partitioning using a greedy operator parallelism algorithm. Our empirical studies show that AdaKnife achieves a 66.5% reduction in latency compared to baselines.

源语言英语
页(从-至)736-748
页数13
期刊IEEE Transactions on Mobile Computing
24
2
DOI
出版状态已出版 - 2025

指纹

探究 'AdaKnife: Flexible DNN Offloading for Inference Acceleration on Heterogeneous Mobile Devices' 的科研主题。它们共同构成独一无二的指纹。

引用此