Abstract

This paper concerns an acceleration method for fixed-point iterations that originated in work of D. G. Anderson [J. Assoc. Comput. Mach., 12 (1965), pp. 547–560], which we accordingly call Anderson acceleration here. This method has enjoyed considerable success and wide usage in electronic structure computations, where it is known as Anderson mixing; however, it seems to have been untried or underexploited in many other important applications. Moreover, while other acceleration methods have been extensively studied by the mathematics and numerical analysis communities, this method has received relatively little attention from these communities over the years. A recent paper by H. Fang and Y. Saad [Numer. Linear Algebra Appl., 16 (2009), pp. 197–221] has clarified a remarkable relationship of Anderson acceleration to quasi-Newton (secant updating) methods and extended it to define a broader Anderson family of acceleration methods. In this paper, our goals are to shed additional light on Anderson acceleration and to draw further attention to its usefulness as a general tool. We first show that, on linear problems, Anderson acceleration without truncation is “essentially equivalent” in a certain sense to the generalized minimal residual (GMRES) method. We also show that the Type 1 variant in the Fang–Saad Anderson family is similarly essentially equivalent to the Arnoldi (full orthogonalization) method. We then discuss practical considerations for implementing Anderson acceleration and illustrate its performance through numerical experiments involving a variety of applications.

MSC codes

  1. 65H10
  2. 65F10

Keywords

  1. acceleration methods
  2. fixed-point iterations
  3. generalized minimal residual method
  4. Arnoldi (full orthogonalization) method
  5. iterative methods
  6. expectation-maximization algorithm
  7. mixture densities
  8. alternating least-squares
  9. nonnegative matrix factorization
  10. domain decomposition

Get full access to this article

View all available purchase options and get full access to this article.

References

1.
D. G. Anderson, Iterative procedures for nonlinear integral equations, J. Assoc. Comput. Mach., 12 (1965), pp. 547–560.
2.
C. Boutsidis and E. Gallopoulos, SVD based initialization: A head start for nonnegative matrix factorization, Pattern Recognition, 41 (2008), pp. 1350–1362.
3.
C. Brezinski, Convergence acceleration during the 20th century, J. Comput. Appl. Math., 122 (2000), pp. 1–21.
4.
C. Brezinski and M. Redivo-Zaglia, Extrapolation Methods Theory and Practice, North–Holland, Amsterdam, 1991.
5.
C. Le Bris, Computational chemistry from the perspective of numerical analysis, Acta Numer., 14 (2005), pp. 363–444.
6.
P. N. Brown, A theoretical comparison of the Arnoldi and GMRES algorithms, SIAM J. Sci. Statist. Comput., 12 (1991), pp. 58–78.
7.
C. G. Broyden, A class of methods for solving nonlinear simultaneous equations, Math. Comp., 19 (1965), pp. 577–593.
8.
X.-C. Cai and D. E. Keyes, Nonlinearly preconditioned inexact Newton algorithms, SIAM J. Sci. Comput., 24 (2002), pp. 183–200.
9.
X.-C. Cai, D. E. Keyes, and D. P. Young, A nonlinear additive Schwarz preconditioned inexact Newton method for shocked duct flows, in Domain Decomposition Methods in Science and Engineering (Lyon, France, 2000), N. Débit, M. Garbey, R. Hoppe, J. Périaux, D. Keyes, and Y. Kuznetsov, eds., CIMNE, Barcelona, 2000, pp. 345–352.
10.
X.-C. Cai and M. Sarkis, A restricted additive Schwarz preconditioner for general sparse linear systems, SIAM J. Sci. Comput., 21 (1999), pp. 792–797.
11.
N. N. Carlson and K. Miller, Design and application of a gradient-weighted moving finite element code I: In one dimension, SIAM J. Sci. Comput., 19 (1998), pp. 728–765.
12.
A. P. Dempster, N. M. Laird, and D. B. Rubin, Maximum-likelihood from incomplete data via the EM algorithm, J. Royal Statist. Soc. Ser. B, 39 (1977), pp. 1–38.
13.
J. E. Dennis Jr., and R. B. Schnabel, Least change secant updates for quasi-Newton methods, SIAM Rev., 21 (1979), pp. 443–459.
14.
J. E. Dennis, Jr., and R. B. Schnabel, Numerical Methods for Unconstrained Optimization and Nonlinear Equations, Classics Appl. Math. 16, SIAM, Philadelphia, 1996.
15.
J. E. Dennis Jr., and H. F. Walker, Convergence theorems for least-change secant update methods, SIAM J. Numer. Anal., 18 (1981), pp. 949–987.
16.
T. Eirola and O. Nevanlinna, Accelerating with rank-one updates, Linear Algebra Appl., 121 (1989), pp. 511–520.
17.
L. Eldén, Matrix Methods in Data Mining and Pattern Recognition, Fundam. Algorithms 4, SIAM, Philadelphia, 2007.
18.
V. Eyert, A comparative study on methods for convergence acceleration of iterative vector sequences, J. Comput. Phys., 124 (1996), pp. 271–285.
19.
H. Fang and Y. Saad, Two classes of multisecant methods for nonlinear acceleration, Numer. Linear Algebra Appl., 16 (2009), pp. 197–221.
20.
M. Garbey, Acceleration of the Schwarz method for elliptic problems, SIAM J. Sci. Comput., 26 (2005), pp. 1871–1893.
21.
R. Glowinski, H. B. Keller, and L. Reinhart, Continuation-conjugate gradient methods for the least squares solution of nonlinear boundary value problems, SIAM J. Sci. Statist. Comput., 6 (1985), pp. 793–832.
22.
G. H. Golub and C. Van Loan, Matrix Computations, 3rd ed., The Johns Hopkins University Press, Baltimore, MD, 1996.
23.
K. Jbilou and H. Sadok, Some results about vector extrapolation methods and related fixed-point iterations, J. Comput. Appl. Math., 36 (1991), pp. 385–398.
24.
K. Jbilou and H. Sadok, Analysis of some vector extrapolation methods for solving systems of linear equations, Numer. Math., 70 (1995), pp. 73–89.
25.
K. Jbilou and H. Sadok, Vector extrapolation methods. Applications and numerical comparison, J. Comput. Appl. Math., 122 (2000), pp. 149–165.
26.
C. T. Kelley, Iterative Methods for Linear and Nonlinear Equations, Frontiers Appl. Math. 16, SIAM, Philadelphia, 1995.
27.
D. A. Knoll and D. E. Keyes, Jacobian-free Newton–Krylov methods: A survey of approaches and applications, J. Comput. Phys., 193 (2004), pp. 357–397.
28.
G. Kresse and J. Furthmüller, Efficiency of ab-initio total energy calculations for metals and semiconductors using a plane-wave basis set, Comput. Mater. Sci., 6 (1996), pp. 15–50.
29.
G. Kresse and J. Furthmüller, Efficient iterative schemes for ab initio total-energy calculations using a plane-wave basis set, Phys. Rev. B, 54 (1996), pp. 169–186.
30.
K. Miller, Nonlinear Krylov and moving nodes in the method of lines, J. Comput. Appl. Math., 183 (2005), pp. 275–287.
31.
P. Ni, Anderson Acceleration of Fixed-Point Iteration with Applications to Electronic Structure Computations, Ph.D. thesis, Worcester Polytechnic Institute, Worcester, MA, 2009.
32.
P. Ni and H. F. Walker, A Linearly Constrained Least-Squares Problem in Electronic Structure Computations, Tech. Report MS-1-13-46, Mathematical Sciences Department, Worcester Polytechnic Institute, Worcester, MA, 2010.
33.
C. W. Oosterlee and T. Washio, Krylov subspace acceleration of nonlinear multigrid with application to recirculating flows, SIAM J. Sci. Comput., 21 (2000), pp. 1670–1690.
34.
M. Pernice and H. F. Walker, NITSOL: A Newton iterative solver for nonlinear systems, SIAM J. Sci. Comput., 19 (1998), pp. 302–318.
35.
P. Pulay, Convergence acceleration of iterative sequences. The case of SCF iteration, Chem. Phys. Lett., 73 (1980), pp. 393–398.
36.
P. Pulay, Improved SCF convergence, J. Comput. Chem., 3 (1982), pp. 556–560.
37.
R. A. Redner and H. F. Walker, Mixture densities, maximum likelihood and the EM algorithm, SIAM Rev., 26 (1984), pp. 195–239.
38.
Y. Saad, Krylov subspace methods for solving large unsymmetric linear systems, Math. Comp., 37 (1981), pp. 105–126.
39.
Y. Saad, Iterative Methods for Sparse Linear Systems, 2nd ed., SIAM, Philadelphia, 2003.
40.
Y. Saad and M. H. Schultz, GMRES: A generalized minimal residual algorithm for solving nonsymmetric linear systems, SIAM J. Sci. Statist. Comput., 7 (1986), pp. 856–869.
41.
J. Sherman and W. J. Morrison, Adjustment of an inverse matrix corresponding to changes in the elements of a given column or a given row of the original matrix, Ann. Math. Statistics, 20 (1949), p. 621.
42.
D. A. Smith, W. F. Ford, and A. Sidi, Extrapolation methods for vector sequences, SIAM Rev., 29 (1987), pp. 199–233.
43.
T. Washio and C. W. Oosterlee, Krylov subspace acceleration for nonlinear multigrid schemes, Electron. Trans. Numer. Anal., 6 (1997), pp. 271–290.
44.
M. A. Woodbury, Inverting Modified Matrices, Tech. Report Memorandum Report 42, Statistical Research Group, Princeton University, Princeton, NJ, 1950.
45.
C. Yang, J. C. Meza, B. Lee, and L.-W. Wang, KSSOLV—a MATLAB toolbox for solving the Kohn–Sham equations, ACM Trans. Math. Software, 36 (2009), pp. 1–35.
46.
U. M. Yang, A Family of Preconditioned Iterative Solvers for Sparse Linear Systems, Ph.D. thesis, Department of Computer Science, University of Illinois at Urbana-Champaign, Urbana, IL, 1995.
47.
D. P. Young, W. P. Huffman, R. G. Melvin, C. L. Hilmes, and F. T. Johnson, Nonlinear elimination in aerodynamic analysis and design optimization, in Large-Scale PDE-Constrained Optimization, L. T. Biegler, O. Ghattas, M. Heinkenschloss, and B. van Bloemen Waanders, eds., Springer-Verlag, Berlin, Heidelberg, New York, 2003, pp. 17–43.

Information & Authors

Information

Published In

cover image SIAM Journal on Numerical Analysis
SIAM Journal on Numerical Analysis
Pages: 1715 - 1735
ISSN (online): 1095-7170

History

Submitted: 22 January 2010
Accepted: 17 May 2011
Published online: 16 August 2011

MSC codes

  1. 65H10
  2. 65F10

Keywords

  1. acceleration methods
  2. fixed-point iterations
  3. generalized minimal residual method
  4. Arnoldi (full orthogonalization) method
  5. iterative methods
  6. expectation-maximization algorithm
  7. mixture densities
  8. alternating least-squares
  9. nonnegative matrix factorization
  10. domain decomposition

Authors

Affiliations

Metrics & Citations

Metrics

Citations

If you have the appropriate software installed, you can download article citation data to the citation manager of your choice. Simply select your manager software from the list below and click Download.

Cited By

View Options

View options

PDF

View PDF

Figures

Tables

Media

Share

Share

Copy the content Link

Share with email

Email a colleague

Share on social media