Skip to main content
U.S. Department of Energy
Office of Scientific and Technical Information

Mixed-Precision S/DGEMM Using the TF32 and TF64 Frameworks on Low-Precision AI Tensor Cores

Conference ·
Using NVIDIA graphics processing units (GPUs) equipped with Tensor Cores has enabled the significant acceleration of general matrix multiplication (GEMM) for applications in machine learning (ML) and artificial intelligence (AI) and in high-performance computing (HPC) generally. The use of such power-efficient, specialized accelerators can provide a performance increase between 8 × and 20 ×, albeit with a loss in precision. However, a high level of precision is required in many large scientific and HPC applications, and computing in single or double precision is still necessary for many of these applications to maintain accuracy. Fortunately, mixed-precision methods can be employed to maintain a higher level of numerical precision while also taking advantage of the performance increases from computing with lower-precision AI cores. With this in mind, we extend the state of the art by using NVIDIA’s new TF32 framework. This new framework not only burdens some constraints of the previous frameworks, such as costly 32 16-bit castings but also provides an equivalent precision and performance by using a much simpler approach. We also propose a new framework called TF64 that attempts double-precision arithmetic with low-precision Tensor Cores. Although this framework does not exist yet, we validated the correctness of this idea and achieved an equivalent of 64-bit precision on 32-bit hardware.
Research Organization:
Oak Ridge National Laboratory (ORNL), Oak Ridge, TN (United States)
Sponsoring Organization:
USDOE
DOE Contract Number:
AC05-00OR22725
OSTI ID:
2438716
Country of Publication:
United States
Language:
English

Similar Records

Towards Precision-Aware Fault Tolerance Approaches for Mixed-Precision Applications
Conference · Sat Nov 12 23:00:00 EST 2022 · OSTI ID:1963399

Climbing the Summit and Pushing the Frontier of Mixed Precision Benchmarks at Extreme Scale
Conference · Tue Nov 01 00:00:00 EDT 2022 · OSTI ID:1997799

Understanding Mixed Precision GEMM with MPGemmFI: Insights into Fault Resilience
Conference · Sat Sep 28 00:00:00 EDT 2024 · OSTI ID:2479157

Related Subjects