How to accelerate ML inference in silicon, while maintaining low-power consumption