Return to search

COMPILER FOR A TRACE-BASED DEEP NEURAL NETWORK ACCELERATOR

Deep Neural Networks (DNNs) are the algorithm of choice for various applications that require modeling large datasets, such as image classification, object detection and natural language processing. DNNs present highly parallel workloads<br>that lead to the need of custom hardware accelerators. Deep Learning (DL) models specialized on different tasks require a programmable custom hardware, and a compiler to efficiently translate various DNNs into an efficient dataflow to be executed on the accelerator. Given a DNN oriented custom instructions set, various compilation phases are needed to generate efficient code and maintain generality to support<br>many models. Different compilation phases need to have different levels of hardware awareness so that it exploits the hardware’s full potential, while abiding with the hardware constraints. The goal of this work is to present a compiler workflow and its hardware aware optimization passes for a custom DNN hardware accelerator. The compiler uses model definition files created from popular frameworks to generate custom instructions. Different levels of hardware aware code optimizations are applied to improve performance and data reuse. The software also exposes an interface to run the accelerator implemented on various FPGA platforms, proving an end-to-end solution.

  1. 10.25394/pgs.8220602.v1
Identiferoai:union.ndltd.org:purdue.edu/oai:figshare.com:article/8220602
Date12 October 2021
CreatorsAndre Xian Ming Chang (6789503)
Source SetsPurdue University
Detected LanguageEnglish
TypeText, Thesis
RightsCC BY 4.0
Relationhttps://figshare.com/articles/thesis/COMPILER_FOR_A_TRACE-BASED_DEEP_NEURAL_NETWORK_ACCELERATOR/8220602

Page generated in 0.0021 seconds