Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Benign overfitting in leaky ReLU networks with moderate input dimension

Authors: Kedar Karhadkar, Erin George, Michael Murray, Guido F. Montufar, Deanna Needell

NeurIPS 2024 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental To further support our theory, we train shallow neural networks on the data model described in Definition 2.1 and record the numerical results. Scripts to reproduce these experiments can be found at https://github.com/kedar2/benign_overfitting. These experiments were run on the CPU of a Mac Book Pro M2 with 8GB of RAM.
Researcher Affiliation Academia Kedar Karhadkar1 Erin George1 Michael Murray1 Guido Montรบfar12 Deanna Needell1 EMAIL 1UCLA 2Max Planck Institute for Mathematics in the Sciences Equal contribution
Pseudocode No The paper contains mathematical equations and descriptions of algorithms, but no structured pseudocode blocks or algorithm boxes are presented.
Open Source Code Yes Scripts to reproduce these experiments can be found at https://github.com/kedar2/benign_overfitting.
Open Datasets No 2.1 Data model We study data generated as per the following data model. Definition 2.1. Suppose d, n, k N, (0, 1) and v Sd 1. If (X, y, y, x, y) D(d, n, k, , v) then 1. X Rn d is a random matrix whose rows, which we denote xi, satisfy xi = yiv + 1 ni, where ni N(0d, 1 d(Id vv T )) are mutually i.i.d..
Dataset Splits Yes Parameter settings: = 0.1, = 5/n, m = 64, k = 0.1n, number of trials = 5, size of validation sample = 1000.
Hardware Specification Yes These experiments were run on the CPU of a Mac Book Pro M2 with 8GB of RAM.
Software Dependencies No The paper describes the learning algorithm (gradient descent with hinge loss) but does not specify any particular software libraries or frameworks with version numbers (e.g., PyTorch, TensorFlow).
Experiment Setup Yes Parameter settings: = 0.1, = 5/n, m = 64, k = 0.1n, number of trials = 5, size of validation sample = 1000.