TY - JOUR
T1 - Hierarchical Deep Reinforcement Learning for Computation Offloading in Autonomous Multi-Robot Systems
AU - Gao, Wen
AU - Yu, Zhiwen
AU - Wang, Liang
AU - Cui, Helei
AU - Guo, Bin
AU - Xiong, Hui
N1 - Publisher Copyright:
© 2016 IEEE.
PY - 2024
Y1 - 2024
N2 - To ensure system responsiveness, some compute-intensive tasks are usually offloaded to cloud or edge computing devices. In environments where connection to external computing facilities is unavailable, computation offloading among members within an autonomous multi-robot system (AMRS) becomes a solution. The challenge lies in how to maximize the use of other members' idle resources without disrupting their local computation tasks. Therefore, this study proposes HRL-AMRS, a hierarchical deep reinforcement learning framework designed to distribute computational loads and reduce the processing time of computational tasks within an AMRS. In this framework, the high-level must consider the impact of data loading scales determined by low-level under varying computational device states on the actual processing times. In addition, the low-level employs Long Short-Term Memory (LSTM) networks to enhance the understanding of time-series states of computing devices. Experimental results show that, across various task sizes and numbers of robots, the framework reduces processing times by an average of 4.32% compared to baseline methods.
AB - To ensure system responsiveness, some compute-intensive tasks are usually offloaded to cloud or edge computing devices. In environments where connection to external computing facilities is unavailable, computation offloading among members within an autonomous multi-robot system (AMRS) becomes a solution. The challenge lies in how to maximize the use of other members' idle resources without disrupting their local computation tasks. Therefore, this study proposes HRL-AMRS, a hierarchical deep reinforcement learning framework designed to distribute computational loads and reduce the processing time of computational tasks within an AMRS. In this framework, the high-level must consider the impact of data loading scales determined by low-level under varying computational device states on the actual processing times. In addition, the low-level employs Long Short-Term Memory (LSTM) networks to enhance the understanding of time-series states of computing devices. Experimental results show that, across various task sizes and numbers of robots, the framework reduces processing times by an average of 4.32% compared to baseline methods.
KW - computation offloading
KW - Multi-robot systems
KW - reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85211628067&partnerID=8YFLogxK
U2 - 10.1109/LRA.2024.3511408
DO - 10.1109/LRA.2024.3511408
M3 - 文章
AN - SCOPUS:85211628067
SN - 2377-3766
JO - IEEE Robotics and Automation Letters
JF - IEEE Robotics and Automation Letters
ER -