Optimizing Inference Efficiency for Tiny DNNs