WEB OF SCIENCE
SCOPUS
Metadata Downloads
| DC Field | Value | Language |
|---|---|---|
| dc.contributor.author | Hwang, Sangwoo | - |
| dc.contributor.author | Lee, Seunghyun | - |
| dc.contributor.author | Park, Dahoon | - |
| dc.contributor.author | Lee, Donghun | - |
| dc.contributor.author | Kung, Jaeha | - |
| dc.date.accessioned | 2025-06-12T10:40:17Z | - |
| dc.date.available | 2025-06-12T10:40:17Z | - |
| dc.date.created | 2025-04-07 | - |
| dc.date.issued | 2024-12-13 | - |
| dc.identifier.issn | 1049-5258 | - |
| dc.identifier.uri | https://scholar.dgist.ac.kr/handle/20.500.11750/58413 | - |
| dc.description.abstract | Event-driven spiking neural networks (SNNs) are promising neural networks that reduce the energy consumption of continuously growing AI models. Recently, keeping pace with the development of transformers, transformer-based SNNs were presented. Due to the incompatibility of self-attention with spikes, however, existing transformer-based SNNs limit themselves by either restructuring self-attention architecture or conforming to non-spike computations. In this work, we propose a novel transformer-to-SNN conversion method that outputs an end-to-end spike-based transformer, named SpikedAttention. Our method directly converts the well-trained transformer without modifying its attention architecture. For the vision task, the proposed method converts Swin Transformer into an SNN without post-training or conversion-aware training, achieving state-of-the-art SNN accuracy on ImageNet dataset, i.e., 80.0% with 28.7M parameters. Considering weight accumulation, neuron potential update, and on-chip data movement, SpikedAttention reduces energy consumption by 42% compared to the baseline ANN, i.e., Swin-T. Furthermore, for the first time, we demonstrate that SpikedAttention successfully converts a BERT model to an SNN with only 0.3% accuracy loss on average consuming 58% less energy on GLUE benchmark. Our code is available at Github (https://github.com/sangwoohwang/SpikedAttention). © 2024 Neural information processing systems foundation. All rights reserved. | - |
| dc.language | English | - |
| dc.publisher | Neural Information Processing Systems Foundation (NeurIPS Foundation) | - |
| dc.relation.ispartof | Advances in Neural Information Processing Systems 37 (NeurIPS 2024) | - |
| dc.title | SpikedAttention: Training-Free and Fully Spike-Driven Transformer-to-SNN Conversion with Winner-Oriented Spike Shift for Softmax Operation | - |
| dc.type | Conference Paper | - |
| dc.identifier.scopusid | 2-s2.0-105000520731 | - |
| dc.identifier.bibliographicCitation | Hwang, Sangwoo. (2024-12-13). SpikedAttention: Training-Free and Fully Spike-Driven Transformer-to-SNN Conversion with Winner-Oriented Spike Shift for Softmax Operation. Conference on Neural Information Processing Systems (poster), 1–24. | - |
| dc.identifier.url | https://neurips.cc/virtual/2024/poster/94181 | - |
| dc.citation.conferenceDate | 2024-12-09 | - |
| dc.citation.conferencePlace | CN | - |
| dc.citation.conferencePlace | Vancouver | - |
| dc.citation.endPage | 24 | - |
| dc.citation.startPage | 1 | - |
| dc.citation.title | Conference on Neural Information Processing Systems (poster) | - |