Structured Quantization for Neural Network Language Model Compression Post date February 4, 2021 ← How TinyML Could Help Developing Countries → When will enter the era of In-memory Computing with thousandfold energy-efficiency? — an algorithm-architecture co-design approach