Skip navigation
Please use this identifier to cite or link to this item: http://arks.princeton.edu/ark:/88435/dsp01s4655k92r
Full metadata record
DC FieldValueLanguage
dc.contributor.advisorJha, Niraj K.
dc.contributor.authorTuli, Shikhar
dc.contributor.otherElectrical and Computer Engineering Department
dc.date.accessioned2024-02-21T17:21:08Z-
dc.date.available2024-02-21T17:21:08Z-
dc.date.created2023-01-01
dc.date.issued2024
dc.identifier.urihttp://arks.princeton.edu/ark:/88435/dsp01s4655k92r-
dc.description.abstractOver the past decade, artificial intelligence (AI) has gained significant interest in industry and academia. Deep neural network (DNN) models have exploded in size over the years. Wider access to graphical processing units (GPUs) and machine learning (ML) accelerators, along with increasing dataset sizes, have fueled this growth. However, efficient evaluation of such models on hardware in terms of accuracy, peak power draw, energy consumption, and chip area of the integrated circuit remains challenging, requiring long design cycles and domain expertise. Increasing model sizes exacerbate this problem. In this thesis, we propose a set of frameworks targeting this challenge from various perspectives. We propose FlexiBERT, the first wide-scale design space for heterogeneous and flexible transformer architectures. We then propose AccelTran, a state-of-the-art transformer accelerator. Taking motivation from AccelTran, we propose ELECTOR, a design space of transformer accelerators, and implement transformer-accelerator co-design leveraging our co-design technique, namely BOSHCODE. We also propose EdgeTran, a co-search technique to find out the best-performing pair, i.e., the transformer model and the edge-AI device. We apply this framework to convolutional neural networks (CNNs) as well (CODEBench). Finally, we discuss two extensions of BOSHCODE: DINI for data imputation and BREATHE for generic multi-objective optimization in vector and graphical search spaces. These works expand the scope of the proposed approach to a much more diverse set of applications.
dc.format.mimetypeapplication/pdf
dc.language.isoen
dc.publisherPrinceton, NJ : Princeton University
dc.subjectBayesian Optimization
dc.subjectElectronic Design Automation
dc.subjectMachine Learning
dc.subject.classificationComputer engineering
dc.subject.classificationElectrical engineering
dc.titleEnhancing Deep Neural Networks in Diverse Resource-Constrained Hardware Settings
dc.typeAcademic dissertations (Ph.D.)
pu.date.classyear2024
pu.departmentElectrical and Computer Engineering
Appears in Collections:Electrical Engineering

Files in This Item:
File Description SizeFormat 
Tuli_princeton_0181D_14849.pdf13.05 MBAdobe PDFView/Download


Items in Dataspace are protected by copyright, with all rights reserved, unless otherwise indicated.