Detail View

All-Rounder: A Flexible AI Accelerator With Diverse Data Format Support and Morphable Structure for Multi-DNN Processing
Citations

WEB OF SCIENCE

Citations

SCOPUS

Metadata Downloads

DC Field Value Language
dc.contributor.author Noh, Seock-Hwan -
dc.contributor.author Lee, Seungpyo -
dc.contributor.author Shin, Banseok -
dc.contributor.author Park, Sehun -
dc.contributor.author Jang, Yongjoo -
dc.contributor.author Kung, Jaeha -
dc.date.accessioned 2025-04-16T14:40:17Z -
dc.date.available 2025-04-16T14:40:17Z -
dc.date.created 2025-03-13 -
dc.date.issued 2025-05 -
dc.identifier.issn 1063-8210 -
dc.identifier.uri http://hdl.handle.net/20.500.11750/58297 -
dc.description.abstract Recognizing the explosive increase in the use of artificial intelligence (AI)-based applications, several industrial companies developed custom application-specific integrated circuits (ASICs) (e.g., Google TPU, IBM RaPiD, and Intel NNP-I/NNP-T) and constructed a hyperscale cloud infrastructure with them. These ASICs perform operations of the inference or training process of AI models which are requested by users. Since the AI models have different data formats and types of operations, the ASICs need to support diverse data formats and various operation shapes. However, the previous ASIC solutions do not or less fulfill these requirements. To overcome these limitations, we first present an area-efficient multiplier, named all-in-one multiplier, which supports multiple bit-widths for both integer (INT) and floating-point (FP) data types. Then, we build a multiply-and-accumulation (MAC) array equipped with these multipliers with multiformat support. In addition, the MAC array can be partitioned into multiple blocks that can be flexibly fused to support various deep neural network (DNN) operation types. We evaluate the practical effectiveness of the proposed MAC array by making an accelerator out of it, named All-rounder. According to our evaluation, the proposed all-in-one multiplier occupies 1.49× smaller area compared to the baselines with dedicated multipliers for each data format. Then, we compare the performance and energy efficiency of the proposed All-rounder with three different accelerators showing consistent speedup and higher efficiency across various AI benchmarks from vision to large language model (LLM)-based language tasks. © 2025 IEEE. All rights reserved. -
dc.language English -
dc.publisher Institute of Electrical and Electronics Engineers -
dc.title All-Rounder: A Flexible AI Accelerator With Diverse Data Format Support and Morphable Structure for Multi-DNN Processing -
dc.type Article -
dc.identifier.doi 10.1109/TVLSI.2025.3540346 -
dc.identifier.wosid 001477390300017 -
dc.identifier.scopusid 2-s2.0-85219334718 -
dc.identifier.bibliographicCitation Noh, Seock-Hwan. (2025-05). All-Rounder: A Flexible AI Accelerator With Diverse Data Format Support and Morphable Structure for Multi-DNN Processing. IEEE Transactions on Very Large Scale Integration (VLSI) Systems, 33(5), 1264–1277. doi: 10.1109/TVLSI.2025.3540346 -
dc.description.isOpenAccess FALSE -
dc.subject.keywordAuthor multitenant execution -
dc.subject.keywordAuthor Deep neural networks -
dc.subject.keywordAuthor hardware acceleration -
dc.subject.keywordAuthor multiple data format support -
dc.citation.endPage 1277 -
dc.citation.number 5 -
dc.citation.startPage 1264 -
dc.citation.title IEEE Transactions on Very Large Scale Integration (VLSI) Systems -
dc.citation.volume 33 -
dc.description.journalRegisteredClass scie -
dc.description.journalRegisteredClass scopus -
dc.relation.journalResearchArea Computer Science; Engineering -
dc.relation.journalWebOfScienceCategory Computer Science, Hardware & Architecture; Engineering, Electrical & Electronic -
dc.type.docType Article -
Show Simple Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Total Views & Downloads