Skip to main content

Section 5.3 Friday Sept. 23

Subsection 5.3.1 Friday 8:30 - 9:00: Coffee and muffins

Subsection 5.3.2 Friday 9:00 - 10:30 Session 5

Subsubsection 5.3.2.1 GEMM SUP Optimizations

Mithun Mohan, co-author Bhaskar Nallani
AMD (India)

Abstract

Optimizations targeting multi-threaded GEMM SUP path, with emphasis on removing scaling bottlenecks.

Subsubsection 5.3.2.2 Efficient TRSM Implementation for Small Dimensions

Bhaskar Nallani, co-author Satish Nuggu
AMD (India)

Abstract

This approach gave the best TRSM performance for small sizes by reducing packing overhead, efficiently using prefetching and diagonal element packing.

Subsubsection 5.3.2.3 Automatic Generation of a Family of Matrix Multiplication Routines with Apache TVM

Adrián Castelló
Universitat Politecnica de Valencia

Abstract:

We explore the utilization of the Apache TVM open source framework to automatically generate a family of algorithms that mimic the approach taken by popular linear algebra libraries, such as GotoBLAS2, BLIS and OpenBLAS, in order to obtain a blocked formulation of the general matrix multiplication (gemm). In addition, we fully automatize the generation process, by also leveraging the Apache TVM framework to derive an ample variety of the internal micro-kernels for gemm. This is in contrast with the convention in high performance libraries, which hand-encode a single micro-kernel per architecture, in general using assembly. In global, the combination of our TVM-generated blocked algorithms and micro-kernels for gemm 1) im- proves portability, maintainability and, globally, the software life cycle; 2) provides high flexibility to be easily tailored and optimized to different data types, processor architectures, and gemm shapes; and 3) features a minimal memory footprint. The approach thus provides a solution that, for example, is especially appealing for deep learning tasks in edge devices.

Subsection 5.3.3 Friday 10:30-10:45 Break

Subsection 5.3.4 Friday 10:45 - 12:15 Session 6

Subsubsection 5.3.4.1 Programming Parallel Dense Matrix Factorizations (and Inversion) for New-Generation NUMA Architectures

Sandra Catalán
Universidad Complutense de Madrid

Abstract:

We address the programmability complexity derived from the emergence of new-generation shared-memory NUMA architectures employing dense matrix factorizations and matrix inversion (DMFI) as a use case.

Our methodology pursues performance portability across different NUMA configurations by proposing multi-domain implementations for DMFI plus a hybrid task- and loop-level parallelization that configures multi-threaded executions to fix core-to-data binding, exploiting locality while maintaining the ease of development usually associated with shared-memory programming. In addition, we introduce a generalization of the multi-domain implementations for DMFI that offers support for virtually any NUMA topology in present and future architectures.

Our experimentations of the DMFI routines on two modern architectures (AMD Rome and Huawei Kunpeng 920), with configurable NUMA topologies, reports performance across architectures and inter- and intra-socket NUMA configurations competitive with state-of-the-art message-passing implementations.

Subsubsection 5.3.4.2 Portable code generation and semi-automatic scheduling for BLIS microkernels with Exo

Julian Bellavita and Grace Dinh
UC Berkeley

Subsubsection 5.3.4.3 Optimizing BLIS for a NUMA architecture: A glimpse of the coming NUMApocalypse

Leick Robinson
Oracle

Abstract

As Moore’s Law approaches its useful limit, chip manufacturers are increasingly turning to higher core counts to achieve greater processing power. This, along with increasing demands for more memory, leads to a natural intermingling of cores and memory in ever more skewed NUMA architectures.

This presentation takes a rapid-fire, visually rich journey through our explorations at Oracle into bringing out the best performance from BLIS on one such architecture, the 160-core Ampere Altra, as well as how this can guide and inform the future of BLIS in an increasingly NUMA landscape.

Subsection 5.3.5 12:15 - ???? Wrap Up

slides/BLIS_Retreat_AMD_GEMM_SUP_Optimizations_MithunAndBhasker.pdf
slides/BLIS_Retreat_AMD_TRSM_Small_BhaskerAndSatish.pdf
slides/Adrian_slides.pdf
slides/Julian_slides.pdf
slides/BLISandtheNUMApocalypse_static.pdf