Cited time in webofscience Cited time in scopus

Skipformer: Evolving Beyond Blocks for Extensively Searching On-Device Language Models With Learnable Attention Window

Title
Skipformer: Evolving Beyond Blocks for Extensively Searching On-Device Language Models With Learnable Attention Window
Author(s)
Bodenham, MatthewKung, Jaeha
Issued Date
2024-09
Citation
IEEE Access, v.12, pp.124428 - 124439
Type
Article
Author Keywords
Computational modelingTransformersNatural language processingComputer architectureContext modelingTrainingLanguage modelsneural architecture searchon-device inferencetransformers
Abstract
Deployment of language models to resource-constrained edge devices is an uphill battle against their ever-increasing size. The task transferability of language models makes deployment to the edge an attractive application. Prior neural architecture search (NAS) works have produced hardware-efficient transformers, but often overlook some architectural features in favor of efficient NAS. We propose a novel evolutionary NAS with large and flexible search space to encourage the exploration of previously unexplored transformer architectures. Our search space allows architectures to vary through their depth and skip connections to transfer information anywhere inside the architecture; Skipformer, the top searched model, displays these novel architectural features. To further increase Skipformer efficiency, we learn a CUDA-accelerated attention window size at each self-attention layer during training. Skipformer achieves 23.3% speed up and requires 19.2% less memory on NVIDIA Jetson Nano with negligible accuracy loss on GLEU benchmark compared to GPT-2 Small.
URI
http://hdl.handle.net/20.500.11750/57487
DOI
10.1109/ACCESS.2024.3420232
Publisher
Institute of Electrical and Electronics Engineers Inc.
Files in This Item:
001311199700001.pdf

001311199700001.pdf

기타 데이터 / 5.23 MB / Adobe PDF download
Appears in Collections:
ETC 1. Journal Articles

qrcode

  • twitter
  • facebook
  • mendeley

Items in Repository are protected by copyright, with all rights reserved, unless otherwise indicated.

BROWSE