TY - JOUR
T1 - An adaptive algorithm for consensus improving in group decision making based on reinforcement learning
AU - Hengsheng, Zhang
AU - Rui, Zhu
AU - Quantao, Wang
AU - Haobin, Shi
AU - Hwang, Kao Shing
N1 - Publisher Copyright:
© 2021 The Chinese Institute of Engineers.
PY - 2022
Y1 - 2022
N2 - In group decision-making problems with reciprocal preference relations, the process of improving individual consistency and consensus degree among decision-makers is dynamic and iter ative. Traditional automatic consensus-reaching process has some problems, such as adopting a fixed strategy in a deterministic environment without considering the dynamics of the decision-making environment, the destruction of individual consistency, etc. To solve these problems, an adaptive consensus-reaching model in a dynamic environment is proposed in this paper. Firstly, a Q-learning algorithm is used to build an environment model for different decision-making states of matrix modification. On the premise of modifying the preference matrix with a small matrix deviation, the optimal modification strategy is learned to improve the consensus degree among decision-makers. Second, we propose a method to control individual consistency in the process of consensus reaching by using a reward function. Finally, several numerical examples are used to illustrate the effectiveness and feasibility of the proposed algorithm. The experimental results show that the proposed algorithm significantly improves the consensus degree of the decision-makers in a small matrix deviation and ensures that the decision-maker’s individual consistency is not destroyed.
AB - In group decision-making problems with reciprocal preference relations, the process of improving individual consistency and consensus degree among decision-makers is dynamic and iter ative. Traditional automatic consensus-reaching process has some problems, such as adopting a fixed strategy in a deterministic environment without considering the dynamics of the decision-making environment, the destruction of individual consistency, etc. To solve these problems, an adaptive consensus-reaching model in a dynamic environment is proposed in this paper. Firstly, a Q-learning algorithm is used to build an environment model for different decision-making states of matrix modification. On the premise of modifying the preference matrix with a small matrix deviation, the optimal modification strategy is learned to improve the consensus degree among decision-makers. Second, we propose a method to control individual consistency in the process of consensus reaching by using a reward function. Finally, several numerical examples are used to illustrate the effectiveness and feasibility of the proposed algorithm. The experimental results show that the proposed algorithm significantly improves the consensus degree of the decision-makers in a small matrix deviation and ensures that the decision-maker’s individual consistency is not destroyed.
KW - consensus
KW - Group decision making
KW - reinforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85122138442&partnerID=8YFLogxK
U2 - 10.1080/02533839.2021.2012522
DO - 10.1080/02533839.2021.2012522
M3 - 文章
AN - SCOPUS:85122138442
SN - 0253-3839
VL - 45
SP - 161
EP - 174
JO - Journal of the Chinese Institute of Engineers, Transactions of the Chinese Institute of Engineers,Series A
JF - Journal of the Chinese Institute of Engineers, Transactions of the Chinese Institute of Engineers,Series A
IS - 2
ER -