TY - GEN
T1 - Sparse Optimization Mechanism and Application of Recurrent Neural Network via Fractional-Order Gradient Descent Learning
AU - Kang, Qian
AU - Yu, Dengxiu
AU - Philip Chen, C. L.
N1 - Publisher Copyright:
© 2025 IEEE.
PY - 2025
Y1 - 2025
N2 - Fractional calculus, leveraging its hereditary and infinite memory properties, offers promising applications in information processing and control systems. Recurrent neural networks (RNNs) have also attracted widespread attention in areas such as time series prediction and temporal system modeling. This paper proposes a theoretical result of RNNs based on a fractional-order (FO) gradient descent method and a sparsity mechanism, which not only enhances the stability and sparsity of the network, but also improves its generalization capability. Specifically, the Caputo derivative is firstly employed to define the fractional-order gradient of the error function, which is applied in the backpropagation training of RNNs. Secondly, the smoothing group L1/2 regularization (SGL1/2) is introduced to successfully overcome the oscillation problem of the error function caused by the traditional group L1/2 regularization (GL1/2), and the network architecture is optimized in terms of the redundant nodes and redundant weights of remaining nodes tend to zero, which further improves the sparsity of the network. Finally, numerical simulation results verify the correctness and effectiveness of the proposed algorithm.
AB - Fractional calculus, leveraging its hereditary and infinite memory properties, offers promising applications in information processing and control systems. Recurrent neural networks (RNNs) have also attracted widespread attention in areas such as time series prediction and temporal system modeling. This paper proposes a theoretical result of RNNs based on a fractional-order (FO) gradient descent method and a sparsity mechanism, which not only enhances the stability and sparsity of the network, but also improves its generalization capability. Specifically, the Caputo derivative is firstly employed to define the fractional-order gradient of the error function, which is applied in the backpropagation training of RNNs. Secondly, the smoothing group L1/2 regularization (SGL1/2) is introduced to successfully overcome the oscillation problem of the error function caused by the traditional group L1/2 regularization (GL1/2), and the network architecture is optimized in terms of the redundant nodes and redundant weights of remaining nodes tend to zero, which further improves the sparsity of the network. Finally, numerical simulation results verify the correctness and effectiveness of the proposed algorithm.
KW - Caputo derivative
KW - Fractional calculus
KW - Recurrent neural networks
KW - Smoothing group L regularization
UR - https://www.scopus.com/pages/publications/105033352108
U2 - 10.1109/CPSI66656.2025.11343898
DO - 10.1109/CPSI66656.2025.11343898
M3 - 会议稿件
AN - SCOPUS:105033352108
T3 - 2025 International Conference on Cyber-Physical Social Intelligence, CPSI 2025
BT - 2025 International Conference on Cyber-Physical Social Intelligence, CPSI 2025
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 2025 International Conference on Cyber-Physical Social Intelligence, CPSI 2025
Y2 - 7 November 2025 through 10 November 2025
ER -