Detail View

SpikedAttention: Training-Free and Fully Spike-Driven Transformer-to-SNN Conversion with Winner-Oriented Spike Shift for Softmax Operation
Citations

WEB OF SCIENCE

Citations

SCOPUS

Metadata Downloads

DC Field Value Language
dc.contributor.author Hwang, Sangwoo -
dc.contributor.author Lee, Seunghyun -
dc.contributor.author Park, Dahoon -
dc.contributor.author Lee, Donghun -
dc.contributor.author Kung, Jaeha -
dc.date.accessioned 2025-06-12T10:40:17Z -
dc.date.available 2025-06-12T10:40:17Z -
dc.date.created 2025-04-07 -
dc.date.issued 2024-12-13 -
dc.identifier.issn 1049-5258 -
dc.identifier.uri https://scholar.dgist.ac.kr/handle/20.500.11750/58413 -
dc.description.abstract Event-driven spiking neural networks (SNNs) are promising neural networks that reduce the energy consumption of continuously growing AI models. Recently, keeping pace with the development of transformers, transformer-based SNNs were presented. Due to the incompatibility of self-attention with spikes, however, existing transformer-based SNNs limit themselves by either restructuring self-attention architecture or conforming to non-spike computations. In this work, we propose a novel transformer-to-SNN conversion method that outputs an end-to-end spike-based transformer, named SpikedAttention. Our method directly converts the well-trained transformer without modifying its attention architecture. For the vision task, the proposed method converts Swin Transformer into an SNN without post-training or conversion-aware training, achieving state-of-the-art SNN accuracy on ImageNet dataset, i.e., 80.0% with 28.7M parameters. Considering weight accumulation, neuron potential update, and on-chip data movement, SpikedAttention reduces energy consumption by 42% compared to the baseline ANN, i.e., Swin-T. Furthermore, for the first time, we demonstrate that SpikedAttention successfully converts a BERT model to an SNN with only 0.3% accuracy loss on average consuming 58% less energy on GLUE benchmark. Our code is available at Github (https://github.com/sangwoohwang/SpikedAttention). © 2024 Neural information processing systems foundation. All rights reserved. -
dc.language English -
dc.publisher Neural Information Processing Systems Foundation (NeurIPS Foundation) -
dc.relation.ispartof Advances in Neural Information Processing Systems 37 (NeurIPS 2024) -
dc.title SpikedAttention: Training-Free and Fully Spike-Driven Transformer-to-SNN Conversion with Winner-Oriented Spike Shift for Softmax Operation -
dc.type Conference Paper -
dc.identifier.scopusid 2-s2.0-105000520731 -
dc.identifier.bibliographicCitation Hwang, Sangwoo. (2024-12-13). SpikedAttention: Training-Free and Fully Spike-Driven Transformer-to-SNN Conversion with Winner-Oriented Spike Shift for Softmax Operation. Conference on Neural Information Processing Systems (poster), 1–24. -
dc.identifier.url https://neurips.cc/virtual/2024/poster/94181 -
dc.citation.conferenceDate 2024-12-09 -
dc.citation.conferencePlace CN -
dc.citation.conferencePlace Vancouver -
dc.citation.endPage 24 -
dc.citation.startPage 1 -
dc.citation.title Conference on Neural Information Processing Systems (poster) -
Show Simple Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Total Views & Downloads