Detail View

Title
Deep Partitioned Training from Near-Storage Computing to DNN Accelerators
DGIST Authors
Jang, YongjooKim, SejinKim, DaehoonLee, SungjinKung, Jaeha
Issued Date
2021-01
Citation
Jang, Yongjoo. (2021-01). Deep Partitioned Training from Near-Storage Computing to DNN Accelerators. doi: 10.1109/LCA.2021.3081752
Type
Article
Author Keywords
Computational modelingData modelsDNN acceleratorsIndexesKernelNear-storage computingParallel processingRandom access memoryTrainingTraining deep neural networksWorkload partitioning
Keywords
Virtual storageBatch sizesComputing devicesFpga prototypesTraining timeStorage as a service (STaaS)Deep neural networksRecommender systems
ISSN
1556-6056
Abstract
In this paper, we present deep partitioned training to accelerate computations involved in training DNN models. This is the first work that partitions a DNN model across storage devices, an NPU and a host CPU forming a unified compute node for training workloads. To validate the benefit of using the proposed system during DNN training, a trace-based simulator or an FPGA prototype is used to estimate the overall performance and obtain the layer index to be partitioned that provides the minimum latency. As a case study, we select two benchmarks, i.e., vision-related tasks and a recommendation system. As a result, the training time reduces by 12.2~31.0% with four near-storage computing devices in vision-related tasks with a mini-batch size of 512 and 40.6~44.7% with one near-storage computing device in the selected recommendation system with a mini-batch size of 64. CCBY
URI
http://hdl.handle.net/20.500.11750/15436
DOI
10.1109/LCA.2021.3081752
Publisher
Institute of Electrical and Electronics Engineers
Show Full Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Related Researcher

김대훈
Kim, Daehoon김대훈

Department of Electrical Engineering and Computer Science

read more

Total Views & Downloads