TY - JOUR
T1 - Optimized control for human-multi-robot collaborative manipulation via multi-player Q-learning
AU - Liu, Xing
AU - Huang, Panfeng
AU - Ge, Shuzhi Sam
N1 - Publisher Copyright:
© 2021 The Franklin Institute
PY - 2021/7
Y1 - 2021/7
N2 - In this paper, optimized interaction control is investigated for human-multi-robot collaboration control problems, which cannot be described by the traditional impedance controller. To realize global optimized interaction performance, the multi-player non-zero sum game theory is employed to obtain the optimized interaction control of each robot agent. Regarding the game strategies, Nash equilibrium strategy is utilized in this paper. In human-multi-robot collaboration problems, the dynamics parameters of the human arm and the manipulated object are usually unknown. To obviate the dependence on these parameters, the multi-player Q-learning method is employed. Moreover, for the human-multi-robot collaboration problem, the optimized solution is difficult to resolve due to the existence of the desired reference position. A multi-player Nash Q-learning algorithm considering the desired reference position is proposed to deal with the problem. The validity of the proposed method is verified through simulation studies.
AB - In this paper, optimized interaction control is investigated for human-multi-robot collaboration control problems, which cannot be described by the traditional impedance controller. To realize global optimized interaction performance, the multi-player non-zero sum game theory is employed to obtain the optimized interaction control of each robot agent. Regarding the game strategies, Nash equilibrium strategy is utilized in this paper. In human-multi-robot collaboration problems, the dynamics parameters of the human arm and the manipulated object are usually unknown. To obviate the dependence on these parameters, the multi-player Q-learning method is employed. Moreover, for the human-multi-robot collaboration problem, the optimized solution is difficult to resolve due to the existence of the desired reference position. A multi-player Nash Q-learning algorithm considering the desired reference position is proposed to deal with the problem. The validity of the proposed method is verified through simulation studies.
UR - http://www.scopus.com/inward/record.url?scp=85107639707&partnerID=8YFLogxK
U2 - 10.1016/j.jfranklin.2021.03.017
DO - 10.1016/j.jfranklin.2021.03.017
M3 - 文章
AN - SCOPUS:85107639707
SN - 0016-0032
VL - 358
SP - 5639
EP - 5658
JO - Journal of the Franklin Institute
JF - Journal of the Franklin Institute
IS - 11
ER -