FeFET-based Process-in-Memory Architecture for Low-Power DNN Training

Loading...
Thumbnail Image

External File or Record

Can’t use the file because of accessibility barriers? Contact us

Date

Journal Title

Journal ISSN

Volume Title

Publisher

Abstract

Abstract—Although deep neural networks (DNNs) have become the cornerstone of Artificial Intelligence, the current training of DNNs still requires dozens of CPU hours. Prior works created various customized hardware accelerators for DNNs, however, most of these accelerators are designed to accelerate DNN inference and lack basic support for complex compute phases and sophisticated data dependency involved in DNN training. The major challenges for supporting DNN training come from various layers of the system stack: (1) the current de-facto training methods, error backpropagation (BP), requires all the weights and intermediate data to be stored in memory, and then sequentially consumed in backward paths. Therefore, weight updates are non-local and rely on upstream layers, which makes training parallelization extremely challenging and also incurs significant memory and computing overheads; (2) the power consumption of such CMOS accelerators can reach 200∼250 Watt. Though emerging memory technology based designs demonstrated a great potential in low-power DNN acceleration, their power efficiency is bottlenecked by CMOS analog-to-digital converters (ADCs).

Series and Number:

EducationalLevel:

Is Based On:

Target Name:

Teaches:

Table of Contents

Description

This record is for a postprint of an article published by IEEE in 2021 IEEE/ACM International Symposium on Nanoscale Architectures (NANOARCH) in 2021; the version of record is available at https://doi.org/10.1109/nanoarch53687.2021.9642234.

Keywords

Citation

Journal

2021 IEEE/ACM International Symposium on Nanoscale Architectures (NANOARCH)

Rights

This work may be protected by copyright unless otherwise stated.