1
Deterministic and stochastic acceleration Deterministic and stochastic acceleration Deterministic and stochastic acceleration Deterministic and stochastic acceleration Massimiliano Lupo Pasini 1 , Michele Benzi 1 , Thomas M. Evans 2 , Steve Massimiliano Lupo Pasini 1 , Michele Benzi 1 , Thomas M. Evans 2 , Steve Massimiliano Lupo Pasini , Michele Benzi , Thomas M. Evans , Steve 1 Emory College, Department of Mathematics and Com 1 Emory College, Department of Mathematics and Com 1 Emory College, Department of Mathematics and Com 2 Oak Ridge National Laboratory, 1 Be 2 Oak Ridge National Laboratory, 1 Be Oak Ridge National Laboratory, 1 Be 3 Georgia Institute of Technology, School of Civil and Environm 3 Georgia Institute of Technology, School of Civil and Environm Introduction Introduction Alternating Anderso Introduction Alternating Anderso Introduction Alternating Anderso AAR uses approximate solutions co Scientific computing is moving to exascale AAR uses approximate solutions co Scientific computing is moving to exascale AAR uses approximate solutions co Richardson’s steps to build the ma Applications require high level of concurrency Richardson’s steps to build the ma Applications require high level of concurrency Richardson’s steps to build the ma Next generation computers will exhibit more hardware failures Next generation computers will exhibit more hardware failures Next generation computers will exhibit more hardware failures applications must be resilient applications must be resilient The Anderson mixing is defined a applications must be resilient The Anderson mixing is defined a Standard Krylov subspace methods struggle to Standard Krylov subspace methods struggle to simultaneously obtain efficiency and concurrency simultaneously obtain efficiency and concurrency Standard Krylov methods struggle to achieve resilience The vector is chosen so as to m Standard Krylov methods struggle to achieve resilience The vector is chosen so as to m Standard Krylov methods struggle to achieve resilience Richardson’s schemes benefit computational and data locality Multiple Richardson’s steps with Richardson’s schemes benefit computational and data locality Multiple Richardson’s steps with Richardson’s schemes benefit computational and data locality computational and data locality computational and data locality Convergence on positive defini Convergence on positive defini Mathematical framework Mathematical framework Mathematical framework Reformulate the sparse linear system of interest Reformulate the sparse linear system of interest Reformulate the sparse linear system of interest as a fixed point scheme as a fixed point scheme as a fixed point scheme Conclusions and fu such that the Neumann series recasts the solution as Conclusions and fu such that the Neumann series recasts the solution as Conclusions and fu Identified classes of matrices for (1) Identified classes of matrices for (1) guaranteed (1) guaranteed Competitive performance compa Competitive performance compa One level fixed point schemes are renown for their deteriorated different choices of preconditione One level fixed point schemes are renown for their deteriorated different choices of preconditione asymptotic convergence rate different choices of preconditione asymptotic convergence rate Convergence analysis on specifi Multilevel schemes can accelerate convergence Convergence analysis on specifi Multilevel schemes can accelerate convergence Performance assessment at extr Multilevel schemes can accelerate convergence Performance assessment at extr Purely deterministic algorithms struggle to deal with inherently Use of GPU accelerations Purely deterministic algorithms struggle to deal with inherently Use of GPU accelerations random faulty phenomena Refere random faulty phenomena Refere Refere Refere “Anderson acceleration of the Jacobi ite One level relaxation scheme “Anderson acceleration of the Jacobi ite One level relaxation scheme alternative to Krylov methods for large, s alternative to Krylov methods for large, s P. P. Pratapa, P. Suryanarayana, and J. P. P. Pratapa, P. Suryanarayana, and J. J. Comput. Phys., 306, pp. 4354, 2016. J. Comput. Phys., 306, pp. 4354, 2016. Converge analysis of Anderson-type a Deterministic accelerations Stochastic accelerations Converge analysis of Anderson-type a Deterministic accelerations improve convergence Stochastic accelerations enhance resilience Converge analysis of Anderson-type a iteration”, M. Lupo Pasini In preparatio improve convergence enhance resilience iteration”, M. Lupo Pasini In preparatio techniques for Richardson-type iterations techniques for Richardson-type iterations techniques for Richardson-type iterations techniques for Richardson-type iterations en P. Hamilton 2 , Stuart R. Slattery 2 , Phanish Suryanarayana 3 en P. Hamilton 2 , Stuart R. Slattery 2 , Phanish Suryanarayana 3 en P. Hamilton , Stuart R. Slattery , Phanish Suryanarayana mputer Science, 400 Dowman Drive, Atlanta, GA 30322 mputer Science, 400 Dowman Drive, Atlanta, GA 30322 mputer Science, 400 Dowman Drive, Atlanta, GA 30322 ethel Valley Rd, Oak Ridge, TN 37830 ethel Valley Rd, Oak Ridge, TN 37830 ethel Valley Rd, Oak Ridge, TN 37830 mental Engineering, 790 Atlantic Drive NW, Atlanta, GA 30332 mental Engineering, 790 Atlantic Drive NW, Atlanta, GA 30332 on-Richardson (AAR) Monte Carlo Linear Solvers (MCLS) on-Richardson (AAR) Monte Carlo Linear Solvers (MCLS) on-Richardson (AAR) Monte Carlo Linear Solvers (MCLS) MCLS use random walks defined on a transition matrix P and a omputed by successive MCLS use random walks defined on a transition matrix P and a omputed by successive MCLS use random walks defined on a transition matrix P and a sequence of weights w = H / P , the statistical estimator omputed by successive atrix sequence of weights w i,j = H i,j / P i,j , the statistical estimator atrix sequence of weights w i,j = H i,j / P i,j , the statistical estimator redefines (1) as atrix redefines (1) as redefines (1) as as as Convergence guaranteed for strictly diagonally dominant, M Convergence guaranteed for strictly diagonally dominant, M matrices, generalized diagonally dominant matrices minimize the residual norm. matrices, generalized diagonally dominant matrices minimize the residual norm. Choice of preconditioners affects convergence hout optimization benefit Choice of preconditioners affects convergence hout optimization benefit Choice of preconditioners affects convergence We use adaptive methods to select the number of histories We use adaptive methods to select the number of histories ite matrices guaranteed Reference solution Tolerance = 0.5 ite matrices guaranteed Reference solution Tolerance = 0.5 Tolerance = 0.1 Tolerance = 0.01 Tolerance = 0.1 Tolerance = 0.01 Conclusions and future developments uture developments Conclusions and future developments uture developments Conclusions and future developments uture developments Identified classes of matrices and preconditioners for which r which convergence is Identified classes of matrices and preconditioners for which r which convergence is MCLS are guaranteed to converge a priori MCLS are guaranteed to converge a priori Difficulty in ensuring a priori convergence of MCLS for a general ared to restarted GMRES for Difficulty in ensuring a priori convergence of MCLS for a general ared to restarted GMRES for problem er problem er problem er Algorithm scalability has not yet been analyzed ic classes of problems Algorithm scalability has not yet been analyzed ic classes of problems Testing MCLS on large parallel architectures and evaluating reme scale Testing MCLS on large parallel architectures and evaluating reme scale their resilience in presence of faults still to be addressed their resilience in presence of faults still to be addressed References ences References ences References ences References ences “A Monte Carlo synthetic-acceleration method for solving the thermal erative method: an efficient “A Monte Carlo synthetic-acceleration method for solving the thermal erative method: an efficient radiation diffusion equation“ . T. M Evans, S. W. Mosher , S. R. Slattery , S. sparse linear systems”, radiation diffusion equation“ . T. M Evans, S. W. Mosher , S. R. Slattery , S. P. Hamilton J. Comp. Phys., Vol. 258, 2014, pp. 338-358. sparse linear systems”, E. Pask P. Hamilton J. Comp. Phys., Vol. 258, 2014, pp. 338-358. E. Pask Analysis of Monte Carlo Accelerated Iterative Methods for Sparse Linear Analysis of Monte Carlo Accelerated Iterative Methods for Sparse Linear Systems”. M. Benzi, T. M. Evans, S. P. Hamilton, M. Lupo Pasini, S. R. acceleration for Richardson’s Systems”. M. Benzi, T. M. Evans, S. P. Hamilton, M. Lupo Pasini, S. R. acceleration for Richardson’s Systems”. M. Benzi, T. M. Evans, S. P. Hamilton, M. Lupo Pasini, S. R. Slattery Numer . Linear Algebra Appl. Vol. 24, Issue 3, 2017. acceleration for Richardson’s on. Slattery Numer . Linear Algebra Appl. Vol. 24, Issue 3, 2017. on.

Deterministic and stochastic acceleration techniques for ...salishan.ahsc-nm.org/uploads/4/9/7/0/49704495/2018-pasini.pdf · Deterministic and stochastic acceleration techniques for

  • Upload
    others

  • View
    14

  • Download
    0

Embed Size (px)

Citation preview

Page 1: Deterministic and stochastic acceleration techniques for ...salishan.ahsc-nm.org/uploads/4/9/7/0/49704495/2018-pasini.pdf · Deterministic and stochastic acceleration techniques for

Deterministic and stochastic acceleration techniques for RichardsonDeterministic and stochastic acceleration techniques for RichardsonDeterministic and stochastic acceleration techniques for RichardsonDeterministic and stochastic acceleration techniques for RichardsonDeterministic and stochastic acceleration techniques for RichardsonMassimiliano Lupo Pasini1, Michele Benzi1, Thomas M. Evans2, Steven P. HamiltonMassimiliano Lupo Pasini1, Michele Benzi1, Thomas M. Evans2, Steven P. HamiltonMassimiliano Lupo Pasini , Michele Benzi , Thomas M. Evans , Steven P. Hamilton

1Emory College, Department of Mathematics and Computer Science, 400 1Emory College, Department of Mathematics and Computer Science, 400 1Emory College, Department of Mathematics and Computer Science, 400 Emory College, Department of Mathematics and Computer Science, 400 2Oak Ridge National Laboratory, 1 Bethel2Oak Ridge National Laboratory, 1 Bethel2Oak Ridge National Laboratory, 1 Bethel

3Georgia Institute of Technology, School of Civil and Environmental Engineering, 790 Atlantic Drive NW, Atlanta, GA 303323Georgia Institute of Technology, School of Civil and Environmental Engineering, 790 Atlantic Drive NW, Atlanta, GA 30332Georgia Institute of Technology, School of Civil and Environmental Engineering, 790 Atlantic Drive NW, Atlanta, GA 30332

IntroductionIntroduction Alternating AndersonIntroduction Alternating AndersonIntroduction Alternating AndersonAAR uses approximate solutions computed• Scientific computing is moving to exascale AAR uses approximate solutions computed• Scientific computing is moving to exascale AAR uses approximate solutions computedRichardson’s steps to build the matrix• Applications require high level of concurrency Richardson’s steps to build the matrix• Applications require high level of concurrency Richardson’s steps to build the matrix• Applications require high level of concurrency

• Next generation computers will exhibit more hardware failures • Next generation computers will exhibit more hardware failures • Next generation computers will exhibit more hardware failures – applications must be resilient– applications must be resilient The Anderson mixing is defined as– applications must be resilient The Anderson mixing is defined as

• Standard Krylov subspace methods struggle to The Anderson mixing is defined as

• Standard Krylov subspace methods struggle to simultaneously obtain efficiency and concurrencysimultaneously obtain efficiency and concurrencysimultaneously obtain efficiency and concurrency

• Standard Krylov methods struggle to achieve resilience The vector is chosen so as to minimize• Standard Krylov methods struggle to achieve resilience The vector is chosen so as to minimize• Standard Krylov methods struggle to achieve resilience • Richardson’s schemes benefit computational and data locality • Multiple Richardson’s steps without optimization benefit • Richardson’s schemes benefit computational and data locality • Multiple Richardson’s steps without optimization benefit • Richardson’s schemes benefit computational and data locality

computational and data localitycomputational and data localitycomputational and data locality• Convergence on positive definite • Convergence on positive definite

Mathematical framework• Convergence on positive definite

Mathematical frameworkMathematical framework• Reformulate the sparse linear system of interest• Reformulate the sparse linear system of interest• Reformulate the sparse linear system of interest

as a fixed point schemeas a fixed point schemeas a fixed point scheme

Conclusions and future such that the Neumann series recasts the solution as Conclusions and future such that the Neumann series recasts the solution as Conclusions and future such that the Neumann series recasts the solution as

• Identified classes of matrices for which convergence is (1)

• Identified classes of matrices for which convergence is (1) guaranteed(1) guaranteedguaranteed

• Competitive performance compared to restarted GMRES for • Competitive performance compared to restarted GMRES for • One level fixed point schemes are renown for their deteriorated

• Competitive performance compared to restarted GMRES for different choices of preconditioner• One level fixed point schemes are renown for their deteriorated different choices of preconditioner• One level fixed point schemes are renown for their deteriorated

asymptotic convergence ratedifferent choices of preconditioner

asymptotic convergence rate Convergence analysis on specific classes of problemsasymptotic convergence rate• Multilevel schemes can accelerate convergence

Convergence analysis on specific classes of problems• Multilevel schemes can accelerate convergence Performance assessment at extreme• Multilevel schemes can accelerate convergence Performance assessment at extreme• Purely deterministic algorithms struggle to deal with inherently

Performance assessment at extreme Use of GPU accelerations• Purely deterministic algorithms struggle to deal with inherently Use of GPU accelerations

random faulty phenomena References Use of GPU accelerations

random faulty phenomena Referencesrandom faulty phenomena ReferencesReferences• “Anderson acceleration of the Jacobi iterative method: an efficientOne level relaxation scheme • “Anderson acceleration of the Jacobi iterative method: an efficientOne level relaxation scheme alternative to Krylov methods for large, sparse alternative to Krylov methods for large, sparse P. P. Pratapa, P. Suryanarayana, and J. E. Pask P. P. Pratapa, P. Suryanarayana, and J. E. Pask P. P. Pratapa, P. Suryanarayana, and J. E. Pask J. Comput. Phys., 306, pp. 4354, 2016.J. Comput. Phys., 306, pp. 4354, 2016.• “Converge analysis of Anderson-type accelerationDeterministic accelerations Stochastic accelerations • “Converge analysis of Anderson-type accelerationDeterministic accelerations

improve convergenceStochastic accelerations

enhance resilience• “Converge analysis of Anderson-type accelerationiteration”, M. Lupo Pasini – In preparationimprove convergence enhance resilience iteration”, M. Lupo Pasini – In preparationiteration”, M. Lupo Pasini – In preparation

Deterministic and stochastic acceleration techniques for Richardson-type iterationsDeterministic and stochastic acceleration techniques for Richardson-type iterationsDeterministic and stochastic acceleration techniques for Richardson-type iterationsDeterministic and stochastic acceleration techniques for Richardson-type iterationsDeterministic and stochastic acceleration techniques for Richardson-type iterations, Steven P. Hamilton2, Stuart R. Slattery2, Phanish Suryanarayana3, Steven P. Hamilton2, Stuart R. Slattery2, Phanish Suryanarayana3, Steven P. Hamilton , Stuart R. Slattery , Phanish Suryanarayana

Department of Mathematics and Computer Science, 400 Dowman Drive, Atlanta, GA 30322Department of Mathematics and Computer Science, 400 Dowman Drive, Atlanta, GA 30322Department of Mathematics and Computer Science, 400 Dowman Drive, Atlanta, GA 30322Department of Mathematics and Computer Science, 400 Dowman Drive, Atlanta, GA 30322Bethel Valley Rd, Oak Ridge, TN 37830Bethel Valley Rd, Oak Ridge, TN 37830Bethel Valley Rd, Oak Ridge, TN 37830

Georgia Institute of Technology, School of Civil and Environmental Engineering, 790 Atlantic Drive NW, Atlanta, GA 30332Georgia Institute of Technology, School of Civil and Environmental Engineering, 790 Atlantic Drive NW, Atlanta, GA 30332Georgia Institute of Technology, School of Civil and Environmental Engineering, 790 Atlantic Drive NW, Atlanta, GA 30332

Anderson-Richardson (AAR) Monte Carlo Linear Solvers (MCLS)Anderson-Richardson (AAR) Monte Carlo Linear Solvers (MCLS)Anderson-Richardson (AAR) Monte Carlo Linear Solvers (MCLS)MCLS use random walks defined on a transition matrix P and a computed by successive MCLS use random walks defined on a transition matrix P and a computed by successive MCLS use random walks defined on a transition matrix P and a sequence of weights w = H / P , the statistical estimator

computed by successive matrix sequence of weights wi,j = Hi,j / Pi,j , the statistical estimator matrix sequence of weights wi,j = Hi,j / Pi,j , the statistical estimator

redefines (1) asmatrix

redefines (1) asredefines (1) as

asasas

• Convergence guaranteed for strictly diagonally dominant, M • Convergence guaranteed for strictly diagonally dominant, M • Convergence guaranteed for strictly diagonally dominant, M matrices, generalized diagonally dominant matricesminimize the residual norm. matrices, generalized diagonally dominant matricesminimize the residual norm. matrices, generalized diagonally dominant matrices

• Choice of preconditioners affects convergenceMultiple Richardson’s steps without optimization benefit • Choice of preconditioners affects convergenceMultiple Richardson’s steps without optimization benefit • Choice of preconditioners affects convergence• We use adaptive methods to select the number of histories• We use adaptive methods to select the number of histories

positive definite matrices guaranteed Reference solution Tolerance = 0.5positive definite matrices guaranteed Reference solution Tolerance = 0.5positive definite matrices guaranteed

Tolerance = 0.1 Tolerance = 0.01Tolerance = 0.1 Tolerance = 0.01

Conclusions and future developmentsand future developments Conclusions and future developmentsand future developments Conclusions and future developmentsand future developments• Identified classes of matrices and preconditioners for which Identified classes of matrices for which convergence is • Identified classes of matrices and preconditioners for which Identified classes of matrices for which convergence is

MCLS are guaranteed to converge a prioriMCLS are guaranteed to converge a prioriMCLS are guaranteed to converge a priori• Difficulty in ensuring a priori convergence of MCLS for a general Competitive performance compared to restarted GMRES for • Difficulty in ensuring a priori convergence of MCLS for a general Competitive performance compared to restarted GMRES for • Difficulty in ensuring a priori convergence of MCLS for a general

problemCompetitive performance compared to restarted GMRES for

preconditioner problempreconditioner problempreconditioner Algorithm scalability has not yet been analyzed Convergence analysis on specific classes of problems Algorithm scalability has not yet been analyzed Convergence analysis on specific classes of problems Testing MCLS on large parallel architectures and evaluating extreme scale Testing MCLS on large parallel architectures and evaluating extreme scale Testing MCLS on large parallel architectures and evaluating

their resilience in presence of faults still to be addressedextreme scale

their resilience in presence of faults still to be addressedtheir resilience in presence of faults still to be addressed

ReferencesReferences ReferencesReferences ReferencesReferences ReferencesReferences• “A Monte Carlo synthetic-acceleration method for solving the thermalacceleration of the Jacobi iterative method: an efficient • “A Monte Carlo synthetic-acceleration method for solving the thermalacceleration of the Jacobi iterative method: an efficientradiation diffusion equation“ . T. M Evans, S. W. Mosher, S. R. Slattery, S. , sparse linear systems”, radiation diffusion equation“ . T. M Evans, S. W. Mosher, S. R. Slattery, S. P. Hamilton – J. Comp. Phys., Vol. 258, 2014, pp. 338-358.

, sparse linear systems”, P. P. Pratapa, P. Suryanarayana, and J. E. Pask – P. Hamilton – J. Comp. Phys., Vol. 258, 2014, pp. 338-358.P. P. Pratapa, P. Suryanarayana, and J. E. Pask – P. Hamilton – J. Comp. Phys., Vol. 258, 2014, pp. 338-358.

• “Analysis of Monte Carlo Accelerated Iterative Methods for Sparse Linear P. P. Pratapa, P. Suryanarayana, and J. E. Pask –

• “Analysis of Monte Carlo Accelerated Iterative Methods for Sparse Linear Systems”. M. Benzi, T. M. Evans, S. P. Hamilton, M. Lupo Pasini, S. R. acceleration for Richardson’s Systems”. M. Benzi, T. M. Evans, S. P. Hamilton, M. Lupo Pasini, S. R. acceleration for Richardson’s Systems”. M. Benzi, T. M. Evans, S. P. Hamilton, M. Lupo Pasini, S. R. Slattery – Numer. Linear Algebra Appl. Vol. 24, Issue 3, 2017.

acceleration for Richardson’s preparation. Slattery – Numer. Linear Algebra Appl. Vol. 24, Issue 3, 2017. preparation. Slattery – Numer. Linear Algebra Appl. Vol. 24, Issue 3, 2017. preparation.