TY - JOUR
T1 - Adaptive Image-Based Visual Servoing Using Reinforcement Learning with Fuzzy State Coding
AU - Shi, Haobin
AU - Wu, Haibo
AU - Xu, Chenxi
AU - Zhu, Jinhui
AU - Hwang, Maxwell
AU - Hwang, Kao Shing
N1 - Publisher Copyright:
© 1993-2012 IEEE.
PY - 2020/12
Y1 - 2020/12
N2 - Image-based visual servoing (IBVS) allows precise control of positioning and motion for relatively stationary targets using visual feedback. For IBVS, a mixture parameter \beta allows better approximation of the image Jacobian matrix, which has a significant effect on the performance of IBVS. However, the setting for the mixture parameter depends on the camera's real-time posture; there is no clear way to define the change rules for most IBVS applications. Using simple model-free reinforcement learning, Q-learning, this article proposes a method to adaptively adjust the image Jacobian matrix for IBVS. If the state-space is discretized, traditional Q-learning encounters problems with the resolution that can cause sudden changes in the action, so the visual servoing system performs poorly. Besides, a robot in a real-world environment also cannot learn on as large a scale as virtual agents, so the efficiency with which agents learn must be increased. This article proposes a method that uses fuzzy state coding to accelerate learning during the training phase and to produce a smooth output in the application phase of the learning experience. A method that compensates for delay also allows more accurate extraction of features in a real environment. The results for simulation and experiment demonstrate that the proposed method performs better than other methods, in terms of learning speed, movement trajectory, and convergence time.
AB - Image-based visual servoing (IBVS) allows precise control of positioning and motion for relatively stationary targets using visual feedback. For IBVS, a mixture parameter \beta allows better approximation of the image Jacobian matrix, which has a significant effect on the performance of IBVS. However, the setting for the mixture parameter depends on the camera's real-time posture; there is no clear way to define the change rules for most IBVS applications. Using simple model-free reinforcement learning, Q-learning, this article proposes a method to adaptively adjust the image Jacobian matrix for IBVS. If the state-space is discretized, traditional Q-learning encounters problems with the resolution that can cause sudden changes in the action, so the visual servoing system performs poorly. Besides, a robot in a real-world environment also cannot learn on as large a scale as virtual agents, so the efficiency with which agents learn must be increased. This article proposes a method that uses fuzzy state coding to accelerate learning during the training phase and to produce a smooth output in the application phase of the learning experience. A method that compensates for delay also allows more accurate extraction of features in a real environment. The results for simulation and experiment demonstrate that the proposed method performs better than other methods, in terms of learning speed, movement trajectory, and convergence time.
KW - Delay compensation
KW - fuzzy method
KW - image Jacobian matrix
KW - image-based visual servoing
KW - mobile robot
KW - reinforcement learning (RL)
UR - http://www.scopus.com/inward/record.url?scp=85097345253&partnerID=8YFLogxK
U2 - 10.1109/TFUZZ.2020.2991147
DO - 10.1109/TFUZZ.2020.2991147
M3 - 文章
AN - SCOPUS:85097345253
SN - 1063-6706
VL - 28
SP - 3244
EP - 3255
JO - IEEE Transactions on Fuzzy Systems
JF - IEEE Transactions on Fuzzy Systems
IS - 12
M1 - 9082110
ER -