TY - GEN
T1 - AdaShadow
T2 - 22nd ACM Conference on Embedded Networked Sensor Systems, SenSys 2024
AU - Fang, Cheng
AU - Liu, Sicong
AU - Zhou, Zimu
AU - Guo, Bin
AU - Tang, Jiaqi
AU - Ma, Ke
AU - Yu, Zhiwen
N1 - Publisher Copyright:
© 2024 Copyright is held by the owner/author(s).
PY - 2024/11/4
Y1 - 2024/11/4
N2 - On-device adapting to continual, unpredictable domain shifts is essential for mobile applications like autonomous driving and augmented reality to deliver seamless user experiences in evolving environments. Test-time adaptation (TTA) emerges as a promising solution by tuning model parameters with unlabeled live data immediately before prediction. However, TTA's unique forward-backward-reforward pipeline notably increases the latency over standard inference, undermining the responsiveness in time-sensitive mobile applications. This paper presents AdaShadow, a responsive test-time adaptation framework for non-stationary mobile data distribution and resource dynamics via selective updates of adaptation-critical layers. Although the tactic is recognized in generic on-device training, TTA's unsupervised and online context presents unique challenges in estimating layer importance and latency, as well as scheduling the optimal layer update plan. AdaShadow addresses these challenges with a backpropagation-free assessor to rapidly identify critical layers, a unit-based runtime predictor to account for resource dynamics in latency estimation, and an online scheduler for prompt layer update planning. Also, AdaShadow incorporates a memory I/O-aware computation reuse scheme to further reduce latency in the reforwardpass. Results show that AdaShadow achieves the best accuracy-latency balance under continual shifts. At low memory and energy costs, Adashadow provides a 2x to 3.5x speedup (ms-level) over state-of-the-art TTA methods with comparable accuracy and a 14.8% to 25.4% accuracy boost over efficient supervised methods with similar latency.
AB - On-device adapting to continual, unpredictable domain shifts is essential for mobile applications like autonomous driving and augmented reality to deliver seamless user experiences in evolving environments. Test-time adaptation (TTA) emerges as a promising solution by tuning model parameters with unlabeled live data immediately before prediction. However, TTA's unique forward-backward-reforward pipeline notably increases the latency over standard inference, undermining the responsiveness in time-sensitive mobile applications. This paper presents AdaShadow, a responsive test-time adaptation framework for non-stationary mobile data distribution and resource dynamics via selective updates of adaptation-critical layers. Although the tactic is recognized in generic on-device training, TTA's unsupervised and online context presents unique challenges in estimating layer importance and latency, as well as scheduling the optimal layer update plan. AdaShadow addresses these challenges with a backpropagation-free assessor to rapidly identify critical layers, a unit-based runtime predictor to account for resource dynamics in latency estimation, and an online scheduler for prompt layer update planning. Also, AdaShadow incorporates a memory I/O-aware computation reuse scheme to further reduce latency in the reforwardpass. Results show that AdaShadow achieves the best accuracy-latency balance under continual shifts. At low memory and energy costs, Adashadow provides a 2x to 3.5x speedup (ms-level) over state-of-the-art TTA methods with comparable accuracy and a 14.8% to 25.4% accuracy boost over efficient supervised methods with similar latency.
KW - latency-efficient test-time adaptation
KW - mobile environments
UR - http://www.scopus.com/inward/record.url?scp=85211809871&partnerID=8YFLogxK
U2 - 10.1145/3666025.3699339
DO - 10.1145/3666025.3699339
M3 - 会议稿件
AN - SCOPUS:85211809871
T3 - SenSys 2024 - Proceedings of the 2024 ACM Conference on Embedded Networked Sensor Systems
SP - 295
EP - 308
BT - SenSys 2024 - Proceedings of the 2024 ACM Conference on Embedded Networked Sensor Systems
PB - Association for Computing Machinery, Inc
Y2 - 4 November 2024 through 7 November 2024
ER -