Ph.D. Dissertation Defense - Duckhwan Kim

*********************************
There is now a CONTENT FREEZE for Mercury while we switch to a new platform. It began on Friday, March 10 at 6pm and will end on Wednesday, March 15 at noon. No new content can be created during this time, but all material in the system as of the beginning of the freeze will be migrated to the new platform, including users and groups. Functionally the new site is identical to the old one. webteam@gatech.edu
*********************************

Event Details
  • Date/Time:
    • Thursday August 3, 2017 - Friday August 4, 2017
      12:00 pm - 1:59 pm
  • Location: Room 2100, Klaus
  • Phone:
  • URL:
  • Email:
  • Fee(s):
    N/A
  • Extras:
Contact
No contact information submitted.
Summaries

Summary Sentence: NeuroCube: Energy-efficient Programmable Digital Deep Learning Accelerator based on Processor in Memory Platform

Full Summary: No summary paragraph submitted.

TitleNeuroCube: Energy-efficient Programmable Digital Deep Learning Accelerator based on Processor in Memory Platform

Committee:

Dr. Saibal Mukhopadhyay, ECE, Chair , Advisor

Dr. Sudhakar Yalamanchili, ECE

Dr. Hyesoon Kim, CoC

Dr. Asif Khan, ECE

Dr. Sek Chai, SRI International

Abstract:

This thesis presents a programmable and scalable digital deep learning accelerator based on 3D high density memory integrated with logic tier for efficient deep learning computing. The proposed architecture consists of clusters of processing engines, connected by 2D mesh network as a processing tier, which is integrated in 3D with multiple tiers of DRAM. The PE clusters access multiple memory channels (vaults) in parallel. The operating principle, referred to as the memory centric computing, embeds specialized state-machines within the vault controllers of HMC to drive data into the PE clusters. Moreover, applying approximate computing during the inference to save more power and the relationship between on-chip training conditions and approximate inference allows energy optimization for both inference and training. The proposed architecture is synthesized in 15nm FinFet technology and its area and power analysis is provided. The performance of the Neurocube is evaluated and illustrated through the mapping of different state of art Deep Neural Network and estimating the subsequent power and performance for both training and inference.

Additional Information

In Campus Calendar
No
Groups

ECE Ph.D. Dissertation Defenses

Invited Audience
Public
Categories
Other/Miscellaneous
Keywords
Phd Defense, graduate students
Status
  • Created By: Daniela Staiculescu
  • Workflow Status: Published
  • Created On: Jul 25, 2017 - 4:29pm
  • Last Updated: Jul 25, 2017 - 4:29pm