Old Web
English
Sign In
Acemap
>
Paper
>
A 17-95.6 TOPS/W Deep Learning Inference Accelerator with Per-Vector Scaled 4-bit Quantization for Transformers in 5nm.
A 17-95.6 TOPS/W Deep Learning Inference Accelerator with Per-Vector Scaled 4-bit Quantization for Transformers in 5nm.
2022
Ben Keller
Rangharajan Venkatesan
Steve Dai
Stephen G. Tell
Brian Zimmer
William J. Dally
C. Thomas Gray
Brucek Khailany
Correction
Cite
Save
Machine Reading By IdeaReader
0
References
0
Citations
NaN
KQI
[]