摘要
In this paper, we design and analyze a new zeroth-order online algorithm, namely, the zeroth-order online alternating direction method of multipliers (ZOO-ADMM), which enjoys dual advantages of being gradient-free operation and employing the ADMM to accommodate complex structured regularizers. Compared to the first-order gradient-based online algorithm, we show that ZOO-ADMM requires √m times more iterations, leading to a convergence rate of O(√m/√T), where m is the number of optimization variables, and T is the number of iterations. To accelerate ZOO-ADMM, we propose two minibatch strategies: gradient sample averaging and observation averaging, resulting in an improved convergence rate of O(√1 + q−1m/√T), where q is the minibatch size. In addition to convergence analysis, we also demonstrate ZOO-ADMM to applications in signal processing, statistics, and machine learning.
源语言 | 英语 |
---|---|
页 | 288-297 |
页数 | 10 |
出版状态 | 已出版 - 2018 |
活动 | 21st International Conference on Artificial Intelligence and Statistics, AISTATS 2018 - Playa Blanca, Lanzarote, Canary Islands, 西班牙 期限: 9 4月 2018 → 11 4月 2018 |
会议
会议 | 21st International Conference on Artificial Intelligence and Statistics, AISTATS 2018 |
---|---|
国家/地区 | 西班牙 |
市 | Playa Blanca, Lanzarote, Canary Islands |
时期 | 9/04/18 → 11/04/18 |