TY - JOUR
T1 - Asynchronous Localization for Underwater Acoustic Sensor Networks
T2 - A Continuous Control Deep Reinforcement Learning Approach
AU - Zhou, Chengyi
AU - Liu, Meiqin
AU - Zhang, Senlin
AU - Zheng, Ronghao
AU - Dong, Shanling
AU - Liu, Zhunga
N1 - Publisher Copyright:
© 2014 IEEE.
PY - 2024/3/15
Y1 - 2024/3/15
N2 - The localization of underwater acoustic sensor networks (UASNs) has emerged as a critical research area in the marine information fusion field. Generally, the convex optimization method is adopted to solve the localization problem. However, this method has limitations in complex underwater environments, since it is difficult to transform the nonconvex optimization problem into a convex optimization problem under such conditions. Recently, deep reinforcement learning (DRL) has shown great potential and promise in solving intricate optimization tasks. Motivated by this, we propose to adopt DRL for UASNs localization to improve accuracy and robustness. The key challenge is that existing DRL-based methods require discretization of the environment, which leads to a compromise between search time and localization precision. To address this challenge, we first model the localization problem as a Markov decision process (MDP) with continuous state and action spaces and subsequently introduce a continuous control DRL framework to solve the localization problem. Within this framework, we develop three continuous control DRL-based localization estimators to address the localization problem in unsupervised, supervised, and semisupervised scenarios. Comprehensive simulations demonstrate the effectiveness of our approach, as the proposed solutions exhibit several advantageous features compared to traditional methods, such as: 1) compared with the convex optimization-based method, the convex relaxation is not required; 2) compared with the least squares method, the proposed estimators are capable of converging to a global optimal state; and 3) compared with the discrete control DRL method, the proposed estimators reduce localization time and enhance localization accuracy significantly.
AB - The localization of underwater acoustic sensor networks (UASNs) has emerged as a critical research area in the marine information fusion field. Generally, the convex optimization method is adopted to solve the localization problem. However, this method has limitations in complex underwater environments, since it is difficult to transform the nonconvex optimization problem into a convex optimization problem under such conditions. Recently, deep reinforcement learning (DRL) has shown great potential and promise in solving intricate optimization tasks. Motivated by this, we propose to adopt DRL for UASNs localization to improve accuracy and robustness. The key challenge is that existing DRL-based methods require discretization of the environment, which leads to a compromise between search time and localization precision. To address this challenge, we first model the localization problem as a Markov decision process (MDP) with continuous state and action spaces and subsequently introduce a continuous control DRL framework to solve the localization problem. Within this framework, we develop three continuous control DRL-based localization estimators to address the localization problem in unsupervised, supervised, and semisupervised scenarios. Comprehensive simulations demonstrate the effectiveness of our approach, as the proposed solutions exhibit several advantageous features compared to traditional methods, such as: 1) compared with the convex optimization-based method, the convex relaxation is not required; 2) compared with the least squares method, the proposed estimators are capable of converging to a global optimal state; and 3) compared with the discrete control DRL method, the proposed estimators reduce localization time and enhance localization accuracy significantly.
KW - Continuous control
KW - deep reinforcement learning (DRL)
KW - localization
KW - underwater acoustic sensor networks (UASNs)
UR - http://www.scopus.com/inward/record.url?scp=85174839017&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2023.3324392
DO - 10.1109/JIOT.2023.3324392
M3 - 文章
AN - SCOPUS:85174839017
SN - 2327-4662
VL - 11
SP - 9505
EP - 9521
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 6
ER -