TY - JOUR
T1 - Optimal Energy Management for Multistack Fuel Cell Vehicles Based on Hybrid Quantum Reinforcement Learning
AU - Shi, Wenzhuo
AU - Sun, Xianzhuo
AU - Zhang, Zelong
AU - Chen, Junyu
AU - Du, Yuhua
AU - Ruan, Jiaqi
AU - Ding, Yibo
AU - Wang, Lei
AU - Huangfu, Yigeng
AU - Xu, Zhao
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2025
Y1 - 2025
N2 - This article proposes a driving condition recognition (DCR)-based hybrid quantum deep deterministic policy gradient (HQDDPG) method for energy management in multistack fuel cell vehicle hybrid power systems (MFCV HPSs) and its quantum simulation setup on digital signal processors (DSPs). Driving conditions are initially segmented into microtrips and clustered into three types. The DCR method, using a learning vector quantization neural network (LVQNN), is then developed, thus accurately and efficiently identifying driving condition types. Subsequently, quantum reinforcement learning (RL) is proposed to achieve optimal energy management of MFCV HPSs, i.e., power allocation among the multiple fuel cells to minimize the economic metrics based on the DCR results. Compared to classical large-scale neural networks, quantum RL reduces parameters by combining a parameterized quantum circuit (PQC) with a single-layer classical neural network. The PQC encodes and processes state information through quantum mechanics for enhanced computational expressiveness, while the classical neural network transforms the quantum measurement expectations into actionable outputs for energy management. The trained hybrid quantum circuits are implemented on DSPs through quantum simulations. The method is validated through controller hardware-in-the-loop (CHIL) experiments, demonstrating superior performance in optimizing economic metrics compared to conventional methods.
AB - This article proposes a driving condition recognition (DCR)-based hybrid quantum deep deterministic policy gradient (HQDDPG) method for energy management in multistack fuel cell vehicle hybrid power systems (MFCV HPSs) and its quantum simulation setup on digital signal processors (DSPs). Driving conditions are initially segmented into microtrips and clustered into three types. The DCR method, using a learning vector quantization neural network (LVQNN), is then developed, thus accurately and efficiently identifying driving condition types. Subsequently, quantum reinforcement learning (RL) is proposed to achieve optimal energy management of MFCV HPSs, i.e., power allocation among the multiple fuel cells to minimize the economic metrics based on the DCR results. Compared to classical large-scale neural networks, quantum RL reduces parameters by combining a parameterized quantum circuit (PQC) with a single-layer classical neural network. The PQC encodes and processes state information through quantum mechanics for enhanced computational expressiveness, while the classical neural network transforms the quantum measurement expectations into actionable outputs for energy management. The trained hybrid quantum circuits are implemented on DSPs through quantum simulations. The method is validated through controller hardware-in-the-loop (CHIL) experiments, demonstrating superior performance in optimizing economic metrics compared to conventional methods.
KW - Digital signal processor (DSP)-based quantum simulation
KW - driving condition recognition (DCR)
KW - energy management
KW - hybrid quantum reinforcement learning (RL)
KW - multistack fuel cell vehicle (MFCV)
UR - http://www.scopus.com/inward/record.url?scp=85218240390&partnerID=8YFLogxK
U2 - 10.1109/TTE.2025.3542021
DO - 10.1109/TTE.2025.3542021
M3 - 文章
AN - SCOPUS:85218240390
SN - 2332-7782
VL - 11
SP - 8500
EP - 8511
JO - IEEE Transactions on Transportation Electrification
JF - IEEE Transactions on Transportation Electrification
IS - 3
ER -