https://www.mdu.se/

mdu.sePublications
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
FaCT-LSTM: Fast and Compact Ternary Architecture for LSTM Recurrent Neural Networks
University of Tehran, Tehran, Iran.
University of Tehran, Tehran, Iran.
Mälardalen University, School of Innovation, Design and Engineering, Embedded Systems.
University of Tehran, Tehran, Iran.
Show others and affiliations
2022 (English)In: IEEE design & test, ISSN 2168-2356, E-ISSN 2168-2364, Vol. 39, no 3, p. 45-53Article in journal (Refereed) Published
Abstract [en]

Long Short-Term Memory (LSTM) achieved great success in healthcare applications. However, its extensive computation cost and massive model size have become the major obstacles for the deployment of such a powerful algorithm in resource-limited embedded systems such as wearable devices. Quantization is a promising way to reduce the memory footprint and computational cost of neural networks. Although quantization achieved remarkable success in convolutional neural networks (CNNs), it still suffers from large accuracy loss in LSTM networks, especially in extremely low bitwidths. In this paper, we propose Fast and Compact Ternary LSTM (FaCT-LSTM), which bridges the accuracy gap between the full precision and quantized neural networks. We propose a hardware-friendly circuit to implement ternarized LSTM and eliminate computation-intensive floating-point operations. With the proposed ternarized LSTM architectures, our experiments on the ECG and EMG signals show ~0.88 to 2.04% accuracy loss in comparison to the full-precision counterparts while reducing latency and area for ~111× to 116× and ~29× to 33×, respectively. The proposed architectures also improves the memory footprint and bandwidth of the full precision signal classification, by 17×, and 31×, respectively.

Place, publisher, year, edition, pages
2022. Vol. 39, no 3, p. 45-53
Keywords [en]
Encoding;Computer architecture;Quantization (signal);Logic gates;Computational modeling;Multiplexing;Long short term memory;Long Short -Term Memory (LSTM);ECG;EMG;Wearable Devices;Quantization
National Category
Computer Sciences
Identifiers
URN: urn:nbn:se:mdh:diva-56397DOI: 10.1109/MDAT.2021.3070245ISI: 000803107400010Scopus ID: 2-s2.0-85103775592OAI: oai:DiVA.org:mdh-56397DiVA, id: diva2:1610019
Available from: 2021-11-09 Created: 2021-11-09 Last updated: 2022-11-25Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Sinaei, SimaDaneshtalab, Masoud

Search in DiVA

By author/editor
Sinaei, SimaDaneshtalab, Masoud
By organisation
Embedded Systems
In the same journal
IEEE design & test
Computer Sciences

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 291 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf