DOE PAGES title logo U.S. Department of Energy
Office of Scientific and Technical Information

Title: Avoiding Communication in Primal and Dual Block Coordinate Descent Methods

Journal Article · · SIAM Journal on Scientific Computing
DOI: https://doi.org/10.1137/17M1134433 · OSTI ID:1544140

Primal and dual block coordinate descent methods are iterative methods for solving regularized and unregularized optimization problems. Distributed-memory parallel implementations of these methods have become popular in analyzing large machine learning datasets. However, existing implementations communicate at every iteration, which, on modern data center and supercomputing architectures, often dominates the cost of floating-point computation. Recent results on communication-avoiding Krylov subspace methods suggest that large speedups are possible by re-organizing iterative algorithms to avoid communication. We show how applying similar algorithmic transformations can lead to primal and dual block coordinate descent methods that only communicate every $$s$$ iterations---where $$s$$ is a tuning parameter---instead of every iteration for the regularized least-squares problem. We show that the communication-avoiding variants reduce the number of synchronizations by a factor of $$s$$ on distributed-memory parallel machines without altering the convergence rate and attain strong scaling speedups of up to $$6.1\times$$ over the “standard algorithm" on a Cray XC30 supercomputer.

Research Organization:
Lawrence Berkeley National Laboratory (LBNL), Berkeley, CA (United States). National Energy Research Scientific Computing Center (NERSC)
Sponsoring Organization:
USDOE
OSTI ID:
1544140
Journal Information:
SIAM Journal on Scientific Computing, Journal Name: SIAM Journal on Scientific Computing Journal Issue: 1 Vol. 41; ISSN 1064-8275
Publisher:
SIAMCopyright Statement
Country of Publication:
United States
Language:
English

References (15)

Iteration complexity of randomized block-coordinate descent methods for minimizing a composite function journal December 2012
Coordinate descent algorithms journal March 2015
On the efficient implementation of preconditioned s-step conjugate gradient methods on multiprocessors with memory hierarchy journal July 1989
s-step iterative methods for symmetric linear systems journal February 1989
An efficient nonsymmetric Lanczos method on parallel vector computers journal October 1992
Communication lower bounds and optimal algorithms for numerical linear algebra journal May 2014
A survey of direct methods for sparse linear systems journal May 2016
Communication-optimal Parallel and Sequential QR and LU Factorizations journal January 2012
Implementation of the GMRES Method Using Householder Transformations journal January 1988
Efficiency of Coordinate Descent Methods on Huge-Scale Optimization Problems journal January 2012
Avoiding Communication in Nonsymmetric Lanczos-Based Krylov Subspace Methods journal January 2013
A Residual Replacement Strategy for Improving the Maximum Attainable Accuracy of $s$-Step Krylov Subspace Methods journal January 2014
Accuracy of the $s$-Step Lanczos Method for the Symmetric Eigenproblem in Finite Precision journal January 2015
LIBSVM: A library for support vector machines journal April 2011
Expected Length of the Longest Probe Sequence in Hash Code Searching journal April 1981

Cited By (1)

Numerical algorithms for high-performance computational science
  • Dongarra, Jack; Grigori, Laura; Higham, Nicholas J.
  • Philosophical Transactions of the Royal Society A: Mathematical, Physical and Engineering Sciences, Vol. 378, Issue 2166 https://doi.org/10.1098/rsta.2019.0066
journal January 2020

Similar Records

Related Subjects