Adaptive Neural Networks for Efficient Inference
This topic contains 0 replies, has 1 voice, and was last updated by arXiv 1 year, 7 months ago.

Adaptive Neural Networks for Efficient Inference
We present an approach to adaptively utilize deep neural networks in order to reduce the evaluation time on new examples without loss of accuracy. Rather than attempting to redesign or approximate existing networks, we propose two schemes that adaptively utilize networks. We first pose an adaptive network evaluation scheme, where we learn a system to adaptively choose the components of a deep network to be evaluated for each example. By allowing examples correctly classified using early layers of the system to exit, we avoid the computational time associated with full evaluation of the network. We extend this to learn a network selection system that adaptively selects the network to be evaluated for each example. We show that computational time can be dramatically reduced by exploiting the fact that many examples can be correctly classified using relatively efficient networks and that complex, computationally costly networks are only necessary for a small fraction of examples. We pose a global objective for learning an adaptive early exit or network selection policy and solve it by reducing the policy learning problem to a layerbylayer weighted binary classification problem. Empirically, these approaches yield dramatic reductions in computational cost, with up to a 2.8x speedup on stateoftheart networks from the ImageNet image recognition challenge with minimal (<1%) loss of top5 accuracy.
Adaptive Neural Networks for Efficient Inference
by Tolga Bolukbasi, Joseph Wang, Ofer Dekel, Venkatesh Saligrama
https://arxiv.org/pdf/1702.07811v2.pdf
You must be logged in to reply to this topic.