TY - GEN
T1 - Neural network classifiers using stochastic computing with a hardware-oriented approximate activation function
AU - Li, Bingzhe
AU - Qin, Yaobin
AU - Yuan, Bo
AU - Lilja, David J.
N1 - Publisher Copyright:
© 2017 IEEE.
Copyright:
Copyright 2018 Elsevier B.V., All rights reserved.
PY - 2017/11/22
Y1 - 2017/11/22
N2 - Neural networks are becoming prevalent in many areas, such as pattern recognition and medical diagnosis. Stochastic computing is one potential solution for neural networks implemented in low-power back-end devices such as solar-powered devices and Internet-of-things (IoT) devices. In this paper, we investigate a new architecture of stochastic neural networks with a hardware-oriented approximate activation function. The new proposed approximate activation function can be omitted while keeping the functionality well. Thus, it reduces the stochastic implementation complexity and hardware costs. Moreover, the new architecture significantly improves recognition error rates compared to previous stochastic neural networks with sigmoid function. Three classical types of neural networks are explored, multiple layer perceptron (MLP), restricted Boltzmann machine (RBM) and convolutional neural networks (CNN). The experimental results indicate the new proposed architecture achieves more than 25%, 60% and 3x reduction than previous stochastic neural networks, and more than 30x, 30x and 52% reduction than conventional binary neural networks, in terms of area, power and energy, respectively, while maintaining the similar error rates compared to the conventional neural networks.
AB - Neural networks are becoming prevalent in many areas, such as pattern recognition and medical diagnosis. Stochastic computing is one potential solution for neural networks implemented in low-power back-end devices such as solar-powered devices and Internet-of-things (IoT) devices. In this paper, we investigate a new architecture of stochastic neural networks with a hardware-oriented approximate activation function. The new proposed approximate activation function can be omitted while keeping the functionality well. Thus, it reduces the stochastic implementation complexity and hardware costs. Moreover, the new architecture significantly improves recognition error rates compared to previous stochastic neural networks with sigmoid function. Three classical types of neural networks are explored, multiple layer perceptron (MLP), restricted Boltzmann machine (RBM) and convolutional neural networks (CNN). The experimental results indicate the new proposed architecture achieves more than 25%, 60% and 3x reduction than previous stochastic neural networks, and more than 30x, 30x and 52% reduction than conventional binary neural networks, in terms of area, power and energy, respectively, while maintaining the similar error rates compared to the conventional neural networks.
KW - Approximate activation function
KW - Hardware implementation
KW - Neural network
KW - Stochastic computing
UR - http://www.scopus.com/inward/record.url?scp=85041678757&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85041678757&partnerID=8YFLogxK
U2 - 10.1109/ICCD.2017.23
DO - 10.1109/ICCD.2017.23
M3 - Conference contribution
AN - SCOPUS:85041678757
T3 - Proceedings - 35th IEEE International Conference on Computer Design, ICCD 2017
SP - 97
EP - 104
BT - Proceedings - 35th IEEE International Conference on Computer Design, ICCD 2017
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 35th IEEE International Conference on Computer Design, ICCD 2017
Y2 - 5 November 2017 through 8 November 2017
ER -