XNORAM: An Efficient Computing-in-Memory Architecture for Binary Convolutional Neural Networks with Flexible Dataflow Mapping

2020 
In this paper, an energy-efficient computing-inmemory architecture for binary convolutional neural networks, called XNORAM, is proposed. The XNORAM employs 6T feature cells and 10T weight cells to form one XNORAM column. Multiplexed XNOR operations are embedded in each column. To address the data reuse in convolutional neural networks, flexible dataflow mapping is supported on XNORAM to minimize the external data access. To verify the architecture, we design a 4-KB XNORAM prototype in 65nm CMOS technology. It achieves a throughput of 18. 5GOPs at 100-MHz clock rate and 1.0-V power supply. XNOR-AlexNet is performed on the design achieving 39.86 TOPS/W and 4.63 GOPS/KB utilization with only 1.3% accuracy loss comparing to the original XNOR-Net result on GPUs.
    • Correction
    • Source
    • Cite
    • Save
    • Machine Reading By IdeaReader
    10
    References
    0
    Citations
    NaN
    KQI
    []