Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
On the Expressive Power of Deep Polynomial Neural Networks
Authors: Joe Kileel, Matthew Trager, Joan Bruna
NeurIPS 2019 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We also present computational evidence that it is profitable in terms of expressiveness for layer widths to increase monotonically and then decrease monotonically. ... We have written code2 in the mathematical software Sage Math [12] that computes the dimension of Vd,r for a general architecture d and activation degree r. Our approach is based on randomly selecting parameters = (Wh, . . . , W1) and computing the rank of the Jacobian of Φd,r( ) in (1). ... We present examples of some of our computations in Tables 1 and 2. |
| Researcher Affiliation | Academia | Joe Kileel Princeton University Matthew Trager New York University Joan Bruna New York University |
| Pseudocode | No | The paper describes the steps taken for computational investigation of dimensions but does not present them in pseudocode or an algorithm block format. |
| Open Source Code | Yes | We have written code2 in the mathematical software Sage Math [12]... Available at https://github.com/mtrager/polynomial_networks. |
| Open Datasets | No | The paper conducts computational investigations rather than empirical studies on publicly available datasets for training models. The data used for the computations are generated based on the theoretical framework, not external datasets. |
| Dataset Splits | No | The paper conducts computational investigations of theoretical properties, not empirical studies involving training, validation, and test splits on datasets. |
| Hardware Specification | No | The paper does not provide any specific details about the hardware used for their computational investigations. |
| Software Dependencies | No | The paper mentions 'Sage Math [12]' but does not provide a specific version number or list other software dependencies with versions. |
| Experiment Setup | No | The paper describes the general approach for its computational investigations (e.g., 'randomly selecting parameters', 'leveraging backpropagation', 'finite field F = Z/p Z'), but it does not specify concrete hyperparameters, training configurations, or system-level settings typically found in experimental setups for machine learning models. |