The search functionality is under construction.

IEICE TRANSACTIONS on Electronics

Write Variation & Reliability Error Compensation by Layer-Wise Tunable Retraining of Edge FeFET LM-GA CiM

Shinsei YOSHIKIYO, Naoko MISAWA, Kasidit TOPRASERTPONG, Shinichi TAKAGI, Chihiro MATSUI, Ken TAKEUCHI

  • Full Text Views

    37

  • Cite this

Summary :

This paper proposes a layer-wise tunable retraining method for edge FeFET Computation-in-Memory (CiM) to compensate the accuracy degradation of neural network (NN) by FeFET device errors. The proposed retraining can tune the number of layers to be retrained to reduce inference accuracy degradation by errors that occur after retraining. Weights of the original NN model, accurately trained in cloud data center, are written into edge FeFET CiM. The written weights are changed by FeFET device errors in the field. By partially retraining the written NN model, the proposed method combines the error-affected layers of NN model with the retrained layers. The inference accuracy is thus recovered. After retraining, the retrained layers are re-written to CiM and affected by device errors again. In the evaluation, at first, the recovery capability of NN model by partial retraining is analyzed. Then the inference accuracy after re-writing is evaluated. Recovery capability is evaluated with non-volatile memory (NVM) typical errors: normal distribution, uniform shift, and bit-inversion. For all types of errors, more than 50% of the degraded percentage of inference accuracy is recovered by retraining only the final fully-connected (FC) layer of Resnet-32. To simulate FeFET Local-Multiply and Global-accumulate (LM-GA) CiM, recovery capability is also evaluated with FeFET errors modeled based on FeFET measurements. Retraining only FC layer achieves recovery rate of up to 53%, 66%, and 72% for FeFET write variation, read-disturb, and data-retention, respectively. In addition, just adding two more retraining layers improves recovery rate by 20-30%. In order to tune the number of retraining layers, inference accuracy after re-writing is evaluated by simulating the errors that occur after retraining. When NVM typical errors are injected, it is optimal to retrain FC layer and 3-6 convolution layers of Resnet-32. The optimal number of layers can be increased or decreased depending on the balance between the size of errors before retraining and errors after retraining.

Publication
IEICE TRANSACTIONS on Electronics Vol.E106-C No.7 pp.352-364
Publication Date
2023/07/01
Publicized
2022/12/19
Online ISSN
1745-1353
DOI
10.1587/transele.2022CDP0004
Type of Manuscript
Special Section PAPER (Special Section on Solid-State Circuit Design — Architecture, Circuit, Device and Design Methodology)
Category

Authors

Shinsei YOSHIKIYO
  The University of Tokyo
Naoko MISAWA
  The University of Tokyo
Kasidit TOPRASERTPONG
  The University of Tokyo
Shinichi TAKAGI
  The University of Tokyo
Chihiro MATSUI
  The University of Tokyo
Ken TAKEUCHI
  The University of Tokyo

Keyword