A Table Look-Up Based Ternary Neural Network Processor

2020 
Matrix calculations dominate deep neural network (DNN) operations. Edge devices with applications in computer vision and embedded processors tend to be slower, and therefore, they are not suitable for real-time performance. Recently, dedicated processors have been developed, most of which have product-sum operation circuits. However, the area of the arithmetic circuit is large. In this paper, we propose an inference processor with a table look-up method. By storing the output of the DNN in a table in advance, an arithmetic circuit can be eliminated. We introduce ternary weights that can be realized with a small amount of memory and aid in reduction of zero weights. We show the training algorithm for the ternary weight DNN and the memory storage method by functional decomposition. The overall architecture for a sparse weight is described. To elucidate the superiority of ternary neural networks, we have compared binarization and ternarization. Experimental results show that ternarization reduced memory usage by 92% compared to binarization, while the recognition accuracy slightly improved. Thorough comparison also revealed that our proposed look-up based processor is 41 times faster than an Arduino embedded processor.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    19
    References
    2
    Citations
    NaN
    KQI
    []