Conjugate Gradient and Variance Reduction Based Online ADMM for Low-Rank Distributed Networks

Yitong Chen, Danqi Jin, Jie Chen, Cédric Richard, Wen Zhang

Research output: Contribution to journalArticlepeer-review

1 Scopus citations

Abstract

Modeling the relationships that may connect optimal parameter vectors is essential for the performance of parameter estimation methods in distributed networks. In this paper, we consider a low-rank relationship and introduce matrix factorization to promote this low-rank property. To devise a distributed algorithm that does not require any prior knowledge about the low-rank space, we first formulate local optimization problems at each node, which are subsequently addressed using the Alternating Direction Method of Multipliers (ADMM). Three subproblems naturally arise from ADMM, each resolved in an online manner with low computational costs. Specifically, the first one is solved using stochastic gradient descent (SGD), while the other two are handled using the conjugate gradient descent method to avoid matrix inversion operations. To further enhance performance, a variance reduction algorithm is incorporated into the SGD. Simulation results validate the effectiveness of the proposed algorithm.

Original languageEnglish
Pages (from-to)706-710
Number of pages5
JournalIEEE Signal Processing Letters
Volume32
DOIs
StatePublished - 2025

Keywords

  • ADMM
  • conjugate gradient descent
  • distributed optimization
  • low-rank
  • variance reduction

Fingerprint

Dive into the research topics of 'Conjugate Gradient and Variance Reduction Based Online ADMM for Low-Rank Distributed Networks'. Together they form a unique fingerprint.

Cite this