Numerical Reproducibility for the Parallel Reduction on Multi- and Many-Core Architectures - Archive ouverte HAL Access content directly
Journal Articles Parallel Computing Year : 2015

Numerical Reproducibility for the Parallel Reduction on Multi- and Many-Core Architectures

(1) , (2) , (3) , (3)
1
2
3

Abstract

On modern multi-core, many-core, and heterogeneous architectures, floating-point computations, especially reductions, may become non-deterministic and, therefore, non-reproducible mainly due to the non-associativity of floating-point operations. We introduce an approach to compute the correctly rounded sums of large floating-point vectors accurately and efficiently, achieving deterministic results by construction. Our multi-level algorithm consists of two main stages: first, a filtering stage that relies on fast vectorized floating-point expansion; second, an accumulation stage based on superaccumulators in a high-radix carry-save representation. We present implementations on recent Intel desktop and server processors, Intel Xeon Phi co-processors, and both AMD and NVIDIA GPUs. We show that numerical reproducibility and bit-perfect accuracy can be achieved at no additional cost for large sums that have dynamic ranges of up to 90 orders of magnitude by leveraging arithmetic units that are left underused by standard reduction algorithms.
Fichier principal
Vignette du fichier
superaccumulator.pdf (632.55 Ko) Télécharger le fichier
Origin : Files produced by the author(s)
Loading...

Dates and versions

lirmm-01206348 , version 1 (06-06-2019)

Identifiers

Cite

Caroline Collange, David Defour, Stef Graillat, Roman Iakymchuk. Numerical Reproducibility for the Parallel Reduction on Multi- and Many-Core Architectures. Parallel Computing, 2015, 49, pp.83-97. ⟨10.1016/j.parco.2015.09.001⟩. ⟨lirmm-01206348⟩
398 View
357 Download

Altmetric

Share

Gmail Facebook Twitter LinkedIn More