In this paper, we explore polynomial accelerators that are well-suited for parallel computations, specifically as smoothers in Algebraic MultiGrid (AMG) preconditioners for symmetric positive definite matrices. These accelerators address a minimax problem, initially formulated in Lottes (Numer. Lin. Alg. with Appl. 30(6), e2518 2023), aiming to achieve an optimal (or near-optimal) bound for a polynomial-dependent constant involved in the AMG V-cycle error bound, without requiring information about the matrices’ spectra. Lottes focuses on Chebyshev polynomials of the 4th-kind and defines the relevant recurrence formulas applicable to a general convergent basic smoother. In this paper, we demonstrate the efficacy of these accelerations for large-scale applications on modern GPU-accelerated supercomputers. Furthermore, we formulate a variant of the aforementioned minimax problem, which naturally leads to solutions relying on Chebyshev polynomials of the 1st-kind as accelerators for a basic smoother. For all the polynomial accelerations, we describe efficient GPU kernels for their application and demonstrate their comparable effectiveness on standard benchmarks at very large scales.

Optimal polynomial smoothers for parallel AMG

Durastante Fabio
Membro del Collaboration Group
;
Filippone Salvatore
Membro del Collaboration Group
;
Massei Stefano
Membro del Collaboration Group
;
2025-01-01

Abstract

In this paper, we explore polynomial accelerators that are well-suited for parallel computations, specifically as smoothers in Algebraic MultiGrid (AMG) preconditioners for symmetric positive definite matrices. These accelerators address a minimax problem, initially formulated in Lottes (Numer. Lin. Alg. with Appl. 30(6), e2518 2023), aiming to achieve an optimal (or near-optimal) bound for a polynomial-dependent constant involved in the AMG V-cycle error bound, without requiring information about the matrices’ spectra. Lottes focuses on Chebyshev polynomials of the 4th-kind and defines the relevant recurrence formulas applicable to a general convergent basic smoother. In this paper, we demonstrate the efficacy of these accelerations for large-scale applications on modern GPU-accelerated supercomputers. Furthermore, we formulate a variant of the aforementioned minimax problem, which naturally leads to solutions relying on Chebyshev polynomials of the 1st-kind as accelerators for a basic smoother. For all the polynomial accelerations, we describe efficient GPU kernels for their application and demonstrate their comparable effectiveness on standard benchmarks at very large scales.
2025
D'Ambra, Pasqua; Durastante, Fabio; Filippone, Salvatore; Massei, Stefano; Thomas, Stephen
File in questo prodotto:
Non ci sono file associati a questo prodotto.

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11568/1315327
 Attenzione

Attenzione! I dati visualizzati non sono stati sottoposti a validazione da parte dell'ateneo

Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? 0
social impact