Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Adaptive Neural Networks for Efficient Inference

Authors: Tolga Bolukbasi, Joseph Wang, Ofer Dekel, Venkatesh Saligrama

ICML 2017 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Empirically, these approaches yield dramatic reductions in computational cost, with up to a 2.8x speedup on state-of-the-art networks from the Image Net image recognition challenge with minimal (< 1%) loss of top5 accuracy.
Researcher Affiliation Collaboration 1Boston University, Boston, MA, USA 2Amazon, Cambridge, MA, USA 3Microsoft Research, Redmond, WA, USA.
Pseudocode Yes Algorithm 1 Adaptive Network Learning Pseudocode
Open Source Code No The paper does not provide any statement or link regarding the availability of its source code.
Open Datasets Yes We evaluate our method on the Imagenet 2012 classification dataset (Russakovsky et al., 2015) which has 1000 object classes. We train using the 1.28 million training images and evaluate the system using 50k validation images.
Dataset Splits Yes We train using the 1.28 million training images and evaluate the system using 50k validation images.
Hardware Specification Yes We measure network times using the built-in tool in the Caffe library on a server that utilizes a Nvidia Titan X Pascal with Cu DNN 5.
Software Dependencies Yes We measure network times using the built-in tool in the Caffe library on a server that utilizes a Nvidia Titan X Pascal with Cu DNN 5.
Experiment Setup Yes To output a prediction following each convolutional layer, we train a single layer linear classifier after a global average pooling for each layer... We sweep the cost trade-off parameter in the range 0.0 to 0.1 to achieve different budget points.