, ,
Optimizing resource utilization in target platforms is key to achieving high performance during DNN inference. While optimizations have been proposed for inference latency, memory footprint, and energy consumption, prior hardware-aware neural architecture ...
SPRINGER INTERNATIONAL PUBLISHING AG2022