An improved MAGMA GEMM for Fermi graphics processing units

Rajib Nath, Stanimire Tomov, Jack Dongarra

Research output: Contribution to journalArticlepeer-review

141 Scopus citations

Abstract

We present an improved matrix-matrix multiplication routine (General Matrix Multiply [GEMM]) in the MAGMA BLAS library that targets the NVIDIA Fermi graphics processing units (GPUs) using Compute Unified Data Architecture (CUDA). We show how to modify the previous MAGMA GEMM kernels in order to make a more efficient use of the Fermi's new architectural features, most notably their extended memory hierarchy and memory sizes. The improved kernels run at up to 300 GFlop/s in double precision and up to 645 GFlop/s in single precision arithmetic (on a C2050), which is correspondingly 58% and 63% of the theoretical peak. We compare the improved kernels with the currently available version in CUBLAS 3.1. Further, we show the effect of the new kernels on higher-level dense linear algebra (DLA) routines such as the one-sided matrix factorizations, and compare their performances with corresponding, currently available routines running on homogeneous multicore systems.

Original languageEnglish
Pages (from-to)511-515
Number of pages5
JournalInternational Journal of High Performance Computing Applications
Volume24
Issue number4
DOIs
StatePublished - Nov 2010

Keywords

  • CUDA matrix mutiply
  • Fermi
  • GPU BLAS
  • dense linear algebra
  • hybrid computing

Fingerprint

Dive into the research topics of 'An improved MAGMA GEMM for Fermi graphics processing units'. Together they form a unique fingerprint.

Cite this