Many Models at the Edge: Scaling Deep Inference via Model-Level Caching | IEEE Conference Publication | IEEE Xplore