Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Modular Networks: Learning to Decompose Neural Computation

Authors: Louis Kirsch, Julius Kunze, David Barber

NeurIPS 2018 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We apply modular networks both to image recognition and language modeling tasks, where we achieve superior performance compared to several baselines. Introspection reveals that modules specialize in interpretable contexts.
Researcher Affiliation Academia Louis Kirsch Department of Computer Science University College London EMAIL Julius Kunze Department of Computer Science University College London EMAIL David Barber Department of Computer Science University College London EMAIL now affiliated with IDSIA, The Swiss AI Lab (USI & SUPSI)
Pseudocode Yes Algorithm 1 Training modular networks with generalized EM
Open Source Code Yes A library to use modular layers in Tensor Flow can be found at http://louiskirsch.com/libmodular.
Open Datasets Yes We use the Penn Treebank2 dataset, consisting of 0.9 million words with a vocabulary size of 10,000. (Footnote 2: http://www.fit.vutbr.cz/~imikolov/rnnlm/simple-examples.tgz) and We applied our method to image classification on CIFAR10 [13]
Dataset Splits No No specific train/validation/test dataset splits (e.g., percentages, sample counts, or explicit mention of validation sets) are provided in the main text.
Hardware Specification No No specific hardware details (e.g., GPU/CPU models, memory) used for running the experiments are mentioned in the paper.
Software Dependencies No The paper mentions "Tensor Flow" in the context of their library, but does not provide specific version numbers for TensorFlow or any other software dependencies.
Experiment Setup Yes Except if noted otherwise, we use a controller consisting of a linear transformation followed by a softmax function for each of the K modules to select. Our modules are either linear transformations or convolutions, followed by a Re LU activation. Additional experimental details are given in the supplementary material.