Low-Area and Low-Power VLSI Architectures for Long Short-Term Memory Networks

Mohammed A. Alhartomi*, Mohd. Tasleem Khan, Saeed Alzahrani, Ahmed Alzahmi, Rafi Ahamed Shaik, Jinti Hazarika, Ruwaybih Alsulami, Abdulaziz Alotaibi, Meshal Al-Harthi

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

1 Citation (Scopus)

Abstract

Long short-term memory (LSTM) networks are extensively used in various sequential learning tasks, including speech recognition. Their significance in real-world applications has prompted the demand for cost-effective and power-efficient designs. This paper introduces LSTM architectures based on distributed arithmetic (DA), utilizing circulant and block-circulant matrix-vector multiplications (MVMs) for network compression. The quantized weights-oriented approach for training circulant and block-circulant matrices is considered. By formulating fixed-point circulant/block-circulant MVMs, we explore the impact of kernel size on accuracy. Our DA-based approach employs shared full and partial methods of add-store/store-add followed by a select unit to realize an MVM. It is then coupled with a multi-partial strategy to reduce complexity for larger kernel sizes. Further complexity reduction is achieved by optimizing decoders of multiple select units. Pipelining in add-store enhances speed at the expense of a few pipelined registers. The results of the field-programmable gate array showcase the superiority of our proposed architectures based on the partial store-add method, delivering reductions of 98.71% in DSP slices, 33.59% in slice look-up tables, 13.43% in flip-flops, and 29.76% in power compared to the state-of-the-art.

Original languageEnglish
Pages (from-to)1000-1014
Number of pages15
JournalIEEE Journal on Emerging and Selected Topics in Circuits and Systems
Volume13
Issue number4
Early online date6 Nov 2023
DOIs
Publication statusPublished - Dec 2023

Keywords

  • Inner products
  • long-short term memory
  • matrix-vector multiplication
  • recurrent neural networks
  • VLSI

ASJC Scopus subject areas

  • Electrical and Electronic Engineering

Fingerprint

Dive into the research topics of 'Low-Area and Low-Power VLSI Architectures for Long Short-Term Memory Networks'. Together they form a unique fingerprint.

Cite this