*********************************
There is now a CONTENT FREEZE for Mercury while we switch to a new platform. It began on Friday, March 10 at 6pm and will end on Wednesday, March 15 at noon. No new content can be created during this time, but all material in the system as of the beginning of the freeze will be migrated to the new platform, including users and groups. Functionally the new site is identical to the old one. webteam@gatech.edu
*********************************
Title: Building Efficient Tensor Accelerators for Sparse Andirregular Workloads
Committee:
Dr. Krishna, Advisor
Dr. Kim, Chair
Dr. Vuduc
Dr. Rajamanickam
Abstract: The objective of the proposed research is to create novel methodologies and architectures for building efficient sparse hardware accelerators that target machine learning and scientific workloads. Tensor kernels enable a vast amount of applications, including image classification and data analytic. One popular kernel that is often targeted for hardware acceleration is matrix multiplication. However, the dimensions, amount of sparsity and structure of each matrix operand varies across workloads. For example, in deep learning, tall-skinny matrices often occur because of the use of mini-batches and weight factorization. Additionally, the sparsity level of each matrix can span from 0% up to 99.99..% sparse across various workloads. Both the irregularity and sparsity have significant implications on the compute efficiency of current state-of-the-art accelerators. Many accelerators optimize for workloads of a certain size and sparsity level, but consequently, the accelerators will perform significantly worse for workloads of other sizes and sparsity levels. This is problematic for accelerators within HPC datacenters, as they are expected to run a wide range of workloads and adapt for future applications. The thesis proposes microarchitectures and techniques that maximize accelerator efficiency across workloads.