Cited time in webofscience Cited time in scopus

Full metadata record

DC Field Value Language
dc.contributor.author Park, Sehun -
dc.contributor.author Kim, Jae-joon -
dc.contributor.author Kung, Jaeha -
dc.date.accessioned 2021-10-18T12:00:11Z -
dc.date.available 2021-10-18T12:00:11Z -
dc.date.created 2021-08-19 -
dc.date.issued 2022-07 -
dc.identifier.issn 2168-6750 -
dc.identifier.uri http://hdl.handle.net/20.500.11750/15594 -
dc.description.abstract We propose a HW-SW co-optimization technique to perform energy-efficient spGEMM operations for deep learning. First, we present an automated pruning algorithm, named AutoRelax, that allows some level of relaxation to achieve higher compression ratio. Since the benefit of the proposed pruning algorithm may be limited by the sparsity level of a given weight matrix, we present additional steps to further improve its efficiency. Along with the software approach, we also present a hardware architecture for processing sparse GEMM operations to maximize the benefit of the proposed pruning algorithm and sparse matrix format. To validate the efficiency of our co-optimization methodology, we evaluated the proposed method on three benchmarks, language modeling, speech recognition and image classification. As a result, our approach improved on-chip performance of spGEMM operations by 9.5027.57% and achieved energy reductions of 15.3533.28% considering DRAM accesses over other sparse accelerators. IEEE -
dc.language English -
dc.publisher IEEE Computer Society -
dc.title AutoRelax: HW-SW Co-Optimization for Efficient SpGEMM Operations with Automated Relaxation in Deep Learning -
dc.type Article -
dc.identifier.doi 10.1109/TETC.2021.3089848 -
dc.identifier.scopusid 2-s2.0-85112176332 -
dc.identifier.bibliographicCitation IEEE Transactions on Emerging Topics in Computing, v.10, no.3, pp.1428 - 1442 -
dc.description.isOpenAccess FALSE -
dc.subject.keywordAuthor Computational modeling -
dc.subject.keywordAuthor Data models -
dc.subject.keywordAuthor Deep learning -
dc.subject.keywordAuthor Hardware -
dc.subject.keywordAuthor Indexes -
dc.subject.keywordAuthor Sparse matrices -
dc.subject.keywordAuthor Training -
dc.subject.keywordPlus Energy efficiency -
dc.subject.keywordPlus Modeling languages -
dc.subject.keywordPlus Speech recognition -
dc.subject.keywordPlus Energy reduction -
dc.subject.keywordPlus Hardware architecture -
dc.subject.keywordPlus Higher compression ratios -
dc.subject.keywordPlus Its efficiencies -
dc.subject.keywordPlus Level of relaxation -
dc.subject.keywordPlus Pruning algorithms -
dc.subject.keywordPlus Software approach -
dc.subject.keywordPlus Sparse matrix formats -
dc.subject.keywordPlus Deep learning -
dc.citation.endPage 1442 -
dc.citation.number 3 -
dc.citation.startPage 1428 -
dc.citation.title IEEE Transactions on Emerging Topics in Computing -
dc.citation.volume 10 -
Files in This Item:

There are no files associated with this item.

Appears in Collections:
Department of Electrical Engineering and Computer Science Intelligent Digital Systems Lab 1. Journal Articles

qrcode

  • twitter
  • facebook
  • mendeley

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE