Full metadata record
DC Field | Value | Language |
---|---|---|
dc.contributor.author | Park, Sehun | - |
dc.contributor.author | Kim, Jae-joon | - |
dc.contributor.author | Kung, Jaeha | - |
dc.date.accessioned | 2021-10-18T12:00:11Z | - |
dc.date.available | 2021-10-18T12:00:11Z | - |
dc.date.created | 2021-08-19 | - |
dc.date.issued | 2022-07 | - |
dc.identifier.issn | 2168-6750 | - |
dc.identifier.uri | http://hdl.handle.net/20.500.11750/15594 | - |
dc.description.abstract | We propose a HW-SW co-optimization technique to perform energy-efficient spGEMM operations for deep learning. First, we present an automated pruning algorithm, named AutoRelax, that allows some level of relaxation to achieve higher compression ratio. Since the benefit of the proposed pruning algorithm may be limited by the sparsity level of a given weight matrix, we present additional steps to further improve its efficiency. Along with the software approach, we also present a hardware architecture for processing sparse GEMM operations to maximize the benefit of the proposed pruning algorithm and sparse matrix format. To validate the efficiency of our co-optimization methodology, we evaluated the proposed method on three benchmarks, language modeling, speech recognition and image classification. As a result, our approach improved on-chip performance of spGEMM operations by 9.5027.57% and achieved energy reductions of 15.3533.28% considering DRAM accesses over other sparse accelerators. IEEE | - |
dc.language | English | - |
dc.publisher | IEEE Computer Society | - |
dc.title | AutoRelax: HW-SW Co-Optimization for Efficient SpGEMM Operations with Automated Relaxation in Deep Learning | - |
dc.type | Article | - |
dc.identifier.doi | 10.1109/TETC.2021.3089848 | - |
dc.identifier.scopusid | 2-s2.0-85112176332 | - |
dc.identifier.bibliographicCitation | IEEE Transactions on Emerging Topics in Computing, v.10, no.3, pp.1428 - 1442 | - |
dc.description.isOpenAccess | FALSE | - |
dc.subject.keywordAuthor | Computational modeling | - |
dc.subject.keywordAuthor | Data models | - |
dc.subject.keywordAuthor | Deep learning | - |
dc.subject.keywordAuthor | Hardware | - |
dc.subject.keywordAuthor | Indexes | - |
dc.subject.keywordAuthor | Sparse matrices | - |
dc.subject.keywordAuthor | Training | - |
dc.subject.keywordPlus | Energy efficiency | - |
dc.subject.keywordPlus | Modeling languages | - |
dc.subject.keywordPlus | Speech recognition | - |
dc.subject.keywordPlus | Energy reduction | - |
dc.subject.keywordPlus | Hardware architecture | - |
dc.subject.keywordPlus | Higher compression ratios | - |
dc.subject.keywordPlus | Its efficiencies | - |
dc.subject.keywordPlus | Level of relaxation | - |
dc.subject.keywordPlus | Pruning algorithms | - |
dc.subject.keywordPlus | Software approach | - |
dc.subject.keywordPlus | Sparse matrix formats | - |
dc.subject.keywordPlus | Deep learning | - |
dc.citation.endPage | 1442 | - |
dc.citation.number | 3 | - |
dc.citation.startPage | 1428 | - |
dc.citation.title | IEEE Transactions on Emerging Topics in Computing | - |
dc.citation.volume | 10 | - |
There are no files associated with this item.