Even Faster Accelerated Coordinate Descent Using Non-Uniform Sampling

Zeyuan Allen-Zhu, Zheng Qu, Peter Richtárik, Yang Yuan

Research output: Contribution to conferencePaperpeer-review

Abstract / Description of output

Accelerated coordinate descent is widely used in optimization due to its cheap per-iteration cost and scalability to large-scale problems. Up to a primal-dual transformation, it is also the same as accelerated stochastic gradient descent that is one of the central methods used in machine learning. In this paper, we improve the best known running time of accelerated coordinate descent by a factor up to $\sqrt{n}$. Our improvement is based on a clean, novel non-uniform sampling that selects each coordinate with a probability proportional to the square root of its smoothness parameter. Our proof technique also deviates from the classical estimation sequence technique used in prior work. Our speed-up applies to important problems such as empirical risk minimization and solving linear systems, both in theory and in practice.
Original languageEnglish
Pages1110-1119
Publication statusPublished - 19 Jun 2016
Event33rd International Conference on Machine Learning: ICML 2016 - New York, United States
Duration: 19 Jun 201624 Jun 2016
https://icml.cc/Conferences/2016/

Conference

Conference33rd International Conference on Machine Learning
Abbreviated titleICML 2016
Country/TerritoryUnited States
CityNew York
Period19/06/1624/06/16
Internet address

Keywords / Materials (for Non-textual outputs)

  • math.OC
  • cs.DS
  • math.NA
  • stat.ML

Fingerprint

Dive into the research topics of 'Even Faster Accelerated Coordinate Descent Using Non-Uniform Sampling'. Together they form a unique fingerprint.

Cite this