Detail View

Improving Hardware Efficiency of a Sparse Training Accelerator by Restructuring a Reduction Network
Citations

WEB OF SCIENCE

Citations

SCOPUS

Metadata Downloads

Title
Improving Hardware Efficiency of a Sparse Training Accelerator by Restructuring a Reduction Network
Issued Date
2023-06-27
Citation
Shin, Banseok. (2023-06-27). Improving Hardware Efficiency of a Sparse Training Accelerator by Restructuring a Reduction Network. IEEE Interregional NEWCAS Conference, NEWCAS 2023, 191480. doi: 10.1109/NEWCAS57931.2023.10198090
Type
Conference Paper
ISBN
9798350300246
ISSN
2474-9672
Abstract
Deep learning is used in various applications including recommendation system, natural language processing, and image processing. When training deep learning models, there is inherent sparsity in input or weight matrices. Therefore, efficiently processing sparse general matrix multiplication (spGEMM) is the key to improve training efficiency. In this paper, we present an improved spGEMM accelerator by restructuring a reduction network and using a proper matrix tiling strategy. The overall hardware area is reduced by 21.8% and the power consumption is reduced by 37.5% with the proposed reduction network. When the stationary matrix's sparsity is 80% and the streaming matrix's sparsity is 99%, the required clock cycles reduce by 80% using the proposed matrix tiling method. © 2023 IEEE.
URI
http://hdl.handle.net/20.500.11750/47937
DOI
10.1109/NEWCAS57931.2023.10198090
Publisher
IEEE CAS Society
Show Full Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Total Views & Downloads