On Exact Computation with an Infinitely Wide Neural Net
Authors: Sanjeev Arora, Simon S. Du, Wei Hu, Zhiyuan Li, Russ R. Salakhutdinov, Ruosong Wang
NeurIPS 2019 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We evaluate the performances of CNNs and their corresponding CNTKs on the CIFAR-10 dataset. The implementation details are in Section A. We also compare the performances between CNTKs and their corresponding random features. Due to space limit, we defer these results on random features to Section B. Results. We test two types of architectures, vanilla CNN and CNN with global average pooling (GAP), as described in Sections 4 and H. We also test CNTKs with only 2,000 training data to see whether their performances are consistent with CNTKs and CNNs using the full training set. The results are summarized in Table 1. |
| Researcher Affiliation | Academia | Princeton University and Institute for Advanced Study. Email: arora@cs.princeton.edu Institute for Advanced Study. Email: ssdu@ias.edu Princeton University. Email: huwei@cs.princeton.edu Princeton University. Email: zhiyuanli@cs.princeton.edu k Carnegie Mellon University. Email:rsalakhu@cs.cmu.edu Carnegie Mellon University. Email: ruosongw@andrew.cmu.edu |
| Pseudocode | No | The paper describes algorithmic steps but does not present them in a structured pseudocode or algorithm block. |
| Open Source Code | No | The paper does not provide an explicit statement about open-sourcing the code or a direct link to a code repository for the described methodology. |
| Open Datasets | Yes | We evaluate the performances of CNNs and their corresponding CNTKs on the CIFAR-10 dataset. |
| Dataset Splits | No | The paper does not provide specific dataset split information (exact percentages, sample counts, or detailed splitting methodology) for training, validation, and test sets. It mentions using 'the full CIFAR-10 dataset' and '2,000 training data' for some experiments, but does not detail how these sets are divided for training, validation, and testing. |
| Hardware Specification | No | The paper states 'We thank Amazon Web Services for providing compute time for the experiments in this paper.' but does not specify exact GPU/CPU models, processor types, or memory amounts used for running the experiments. |
| Software Dependencies | No | The paper does not provide specific ancillary software details, such as library names with version numbers, needed to replicate the experiment. |
| Experiment Setup | No | The paper describes the architectures (vanilla CNN, CNN with global average pooling) and depth (e.g., 11-layer, 21-layer) but does not provide specific hyperparameter values like learning rate, batch size, number of epochs, or optimizer settings for the experimental setup. |