TY - JOUR
T1 - Mapping Transformation Enabled High-Performance and Low-Energy Memristor-Based DNNs
AU - Oli-Uz-zaman, Md
AU - Khan, Saleh Ahmad
AU - Yuan, Geng
AU - Liao, Zhiheng
AU - Fu, Jingyan
AU - Ding, Caiwen
AU - Wang, Yanzhi
AU - Wang, Jinhui
N1 - Publisher Copyright:
© 2022 by the authors. Licensee MDPI, Basel, Switzerland.
PY - 2022/3
Y1 - 2022/3
N2 - When deep neural network (DNN) is extensively utilized for edge AI (Artificial Intelli-gence), for example, the Internet of things (IoT) and autonomous vehicles, it makes CMOS (Com-plementary Metal Oxide Semiconductor)-based conventional computers suffer from overly large computing loads. Memristor-based devices are emerging as an option to conduct computing in memory for DNNs to make them faster, much more energy efficient, and accurate. Despite having excellent properties, the memristor-based DNNs are yet to be commercially available because of Stuck-At-Fault (SAF) defects. A Mapping Transformation (MT) method is proposed in this paper to mitigate Stuck-at-Fault (SAF) defects from memristor-based DNNs. First, the weight distribu-tion for the VGG8 model with the CIFAR10 dataset is presented and analyzed. Then, the MT method is used for recovering inference accuracies at 0.1% to 50% SAFs with two typical cases, SA1 (Stuck-At-One): SA0 (Stuck-At-Zero) = 5:1 and 1:5, respectively. The experiment results show that the MT method can recover DNNs to their original inference accuracies (90%) when the ratio of SAFs is smaller than 2.5%. Moreover, even when the SAF is in the extreme condition of 50%, it is still highly efficient to recover the inference accuracy to 80% and 21%. What is more, the MT method acts as a regulator to avoid energy and latency overhead generated by SAFs. Finally, the immunity of the MT Method against non-linearity is investigated, and we conclude that the MT method can benefit accuracy, energy, and latency even with high non-linearity LTP = 4 and LTD = −4.
AB - When deep neural network (DNN) is extensively utilized for edge AI (Artificial Intelli-gence), for example, the Internet of things (IoT) and autonomous vehicles, it makes CMOS (Com-plementary Metal Oxide Semiconductor)-based conventional computers suffer from overly large computing loads. Memristor-based devices are emerging as an option to conduct computing in memory for DNNs to make them faster, much more energy efficient, and accurate. Despite having excellent properties, the memristor-based DNNs are yet to be commercially available because of Stuck-At-Fault (SAF) defects. A Mapping Transformation (MT) method is proposed in this paper to mitigate Stuck-at-Fault (SAF) defects from memristor-based DNNs. First, the weight distribu-tion for the VGG8 model with the CIFAR10 dataset is presented and analyzed. Then, the MT method is used for recovering inference accuracies at 0.1% to 50% SAFs with two typical cases, SA1 (Stuck-At-One): SA0 (Stuck-At-Zero) = 5:1 and 1:5, respectively. The experiment results show that the MT method can recover DNNs to their original inference accuracies (90%) when the ratio of SAFs is smaller than 2.5%. Moreover, even when the SAF is in the extreme condition of 50%, it is still highly efficient to recover the inference accuracy to 80% and 21%. What is more, the MT method acts as a regulator to avoid energy and latency overhead generated by SAFs. Finally, the immunity of the MT Method against non-linearity is investigated, and we conclude that the MT method can benefit accuracy, energy, and latency even with high non-linearity LTP = 4 and LTD = −4.
KW - Accuracy
KW - Deep neural network
KW - Edge artificial intelligence
KW - Energy
KW - Latency
KW - Memristor
KW - Stuck-at-fault
UR - http://www.scopus.com/inward/record.url?scp=85124558619&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85124558619&partnerID=8YFLogxK
U2 - 10.3390/jlpea12010010
DO - 10.3390/jlpea12010010
M3 - Article
AN - SCOPUS:85124558619
SN - 2079-9268
VL - 12
JO - Journal of Low Power Electronics and Applications
JF - Journal of Low Power Electronics and Applications
IS - 1
M1 - 10
ER -