Detail View
Deep Partitioned Training from Near-Storage Computing to DNN Accelerators
- Department of Electrical Engineering and Computer Science
- Computer Architecture and Systems Lab
- 1. Journal Articles
- Department of Electrical Engineering and Computer Science
- Data-Intensive Computing Systems Laboratory
- 1. Journal Articles
- Department of Electrical Engineering and Computer Science
- Intelligent Digital Systems Lab
- 1. Journal Articles
WEB OF SCIENCE
SCOPUS
- Title
- Deep Partitioned Training from Near-Storage Computing to DNN Accelerators
- DGIST Authors
- Jang, Yongjoo ; Kim, Sejin ; Kim, Daehoon ; Lee, Sungjin ; Kung, Jaeha
- Issued Date
- 2021-01
- Citation
- Jang, Yongjoo. (2021-01). Deep Partitioned Training from Near-Storage Computing to DNN Accelerators. doi: 10.1109/LCA.2021.3081752
- Type
- Article
- Author Keywords
- Computational modeling ; Data models ; DNN accelerators ; Indexes ; Kernel ; Near-storage computing ; Parallel processing ; Random access memory ; Training ; Training deep neural networks ; Workload partitioning
- Keywords
- Virtual storage ; Batch sizes ; Computing devices ; Fpga prototypes ; Training time ; Storage as a service (STaaS) ; Deep neural networks ; Recommender systems
- ISSN
- 1556-6056
- Abstract
-
In this paper, we present deep partitioned training to accelerate computations involved in training DNN models. This is the first work that partitions a DNN model across storage devices, an NPU and a host CPU forming a unified compute node for training workloads. To validate the benefit of using the proposed system during DNN training, a trace-based simulator or an FPGA prototype is used to estimate the overall performance and obtain the layer index to be partitioned that provides the minimum latency. As a case study, we select two benchmarks, i.e., vision-related tasks and a recommendation system. As a result, the training time reduces by 12.2~31.0% with four near-storage computing devices in vision-related tasks with a mini-batch size of 512 and 40.6~44.7% with one near-storage computing device in the selected recommendation system with a mini-batch size of 64. CCBY
더보기
- Publisher
- Institute of Electrical and Electronics Engineers
File Downloads
- There are no files associated with this item.
공유
Related Researcher
- Kim, Daehoon김대훈
-
Department of Electrical Engineering and Computer Science
Total Views & Downloads
???jsp.display-item.statistics.view???: , ???jsp.display-item.statistics.download???:
