Detail View

A Transformer Model with Spatiotemporal Input Embedding for fNIRS data-Driven Neural Decoding
Citations

WEB OF SCIENCE

Citations

SCOPUS

Metadata Downloads

DC Field Value Language
dc.contributor.author Lee, Hyunmin -
dc.contributor.author Kim, Taehun -
dc.contributor.author An, Jinung -
dc.date.accessioned 2025-01-07T12:10:14Z -
dc.date.available 2025-01-07T12:10:14Z -
dc.date.created 2024-05-02 -
dc.date.issued 2024-02-26 -
dc.identifier.isbn 9798350309430 -
dc.identifier.issn 2572-7672 -
dc.identifier.uri http://hdl.handle.net/20.500.11750/57516 -
dc.description.abstract Model structures proposed as spatial embedding of fNIRS data have a problem of embedding channels in a row into one channel without considering the brain area. This paper proposes a neural decoding model, ORC-T, using input embedding that reflects fNIRS data's spatiotemporal characteristics. ORC-T embeds the fNIRS channels into new channels with a depth-wise convolution layer based on the fNIRS optod layout. The embedded data is transferred to the classifier via a 2D convolution layer and a transformer encoder. An open dataset (mental arithmetic, BNCI Horizon 2020) and a consciousness dataset were trained to validate model performance. The ORC-T achieved a k-fold CV accuracy of 81.78% and LOSO CV accuracy of 90.19% in open data, showing lower or equivalent performance than the existing model, fNIRS-T. In the consciousness state dataset, the ORC-T achieved a k-fold CV accuracy of 97.09% and LOSO CV accuracy of 59.43%, showing the equivalent performance as the fNIRS-T. The performance of ORC-T and fNIRS-T hardly decreased even though the number of channels input to the model was reduced. Considering the possibility that spatial embedding introduced in ORC-T did not effectively reflect the spatial information of the fNIRS channels, a model using a convolution layer with a kernel in the channel array's horizontal, vertical, and diagonal directions is in the planning process. © 2024 IEEE. -
dc.language English -
dc.publisher Institute of Electrical and Electronics Engineers Inc. -
dc.relation.ispartof International Winter Conference on Brain-Computer Interface, BCI -
dc.title A Transformer Model with Spatiotemporal Input Embedding for fNIRS data-Driven Neural Decoding -
dc.type Conference Paper -
dc.identifier.doi 10.1109/BCI60775.2024.10480484 -
dc.identifier.wosid 001206159700018 -
dc.identifier.scopusid 2-s2.0-85190276396 -
dc.identifier.bibliographicCitation Lee, Hyunmin. (2024-02-26). A Transformer Model with Spatiotemporal Input Embedding for fNIRS data-Driven Neural Decoding. 12th International Winter Conference on Brain-Computer Interface, BCI 2024, 1–4. doi: 10.1109/BCI60775.2024.10480484 -
dc.identifier.url https://brain.korea.ac.kr/bci2024/technicalprogram.php -
dc.citation.conferenceDate 2024-02-26 -
dc.citation.conferencePlace KO -
dc.citation.conferencePlace 강원 -
dc.citation.endPage 4 -
dc.citation.startPage 1 -
dc.citation.title 12th International Winter Conference on Brain-Computer Interface, BCI 2024 -
Show Simple Item Record

File Downloads

  • There are no files associated with this item.

공유

qrcode
공유하기

Related Researcher

안진웅
An, Jinung안진웅

Division of Intelligent Robotics

read more

Total Views & Downloads