A Sliding‐Kernel Computation‐In‐Memory Architecture for Convolutional Neural Network

Abstract: Presently described is a sliding‐kernel computation‐in‐memory (SKCIM) architecture conceptually involving two overlapping layers of functional arrays, one containing memory elements and artificial synapses for neuromorphic computation, the other is used for storing and sliding convolutional kernel matrices. A low‐temperature metal‐oxide thin‐film transistor (TFT) technology capable of monolithically integrating single‐gate TFTs, dual‐gate TFTs, and memory capacitors is deployed for the construction of a physical SKCIM system. Exhibiting an 88% reduction in memory access operations compared to state‐of‐the‐art systems, a 32 × 32 SKCIM system is applied to execute common convolution tasks. A more involved demonstration is the application of a 5‐layer, SKCIM‐based convolutional neural network to the classification of the modified national institute of standards and technology (MNIST) dataset of handwritten numerals, achieving an accuracy rate of over 95%.

Location
Deutsche Nationalbibliothek Frankfurt am Main
Extent
Online-Ressource
Language
Englisch

Bibliographic citation
A Sliding‐Kernel Computation‐In‐Memory Architecture for Convolutional Neural Network ; day:22 ; month:10 ; year:2024 ; extent:12
Advanced science ; (22.10.2024) (gesamt 12)

Creator
Hu, Yushen
Xie, Xinying
Lei, Tengteng
Shi, Runxiao
Wong, Man

DOI
10.1002/advs.202407440
URN
urn:nbn:de:101:1-2410221445527.569872128530
Rights
Open Access; Der Zugriff auf das Objekt ist unbeschränkt möglich.
Last update
15.08.2025, 7:26 AM CEST

Data provider

This object is provided by:
Deutsche Nationalbibliothek. If you have any questions about the object, please contact the data provider.

Associated

  • Hu, Yushen
  • Xie, Xinying
  • Lei, Tengteng
  • Shi, Runxiao
  • Wong, Man

Other Objects (12)