Compositional Curvature Bounds for Deep Neural Networks

Authors: Taha Entesari, Sina Sharifi, Mahyar Fazlyab

ICML 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Finally, we demonstrate the efficacy of our method on classification tasks using the MNIST and CIFAR-10 datasets.
Researcher Affiliation Academia Taha Entesari * 1 Sina Sharifi * 1 Mahyar Fazlyab 1 1Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, United States of America. Correspondence to: Mahyar Fazlyab <mahyarfazlyab@jhu.edu>.
Pseudocode Yes Algorithm 1 Compositional Curvature Estimation of Neural Networks
Open Source Code Yes Our code is available at https://github.com/o4lc/Compositional Curvature-Bounds-for-DNNs.
Open Datasets Yes Finally, we demonstrate the efficacy of our method on classification tasks using the MNIST and CIFAR-10 datasets.
Dataset Splits No The paper mentions training models on MNIST and CIFAR-10 and evaluating on test data, but it does not explicitly specify the proportions for training, validation, and test splits, nor does it explicitly state the use of standard splits for these datasets.
Hardware Specification No The paper does not provide specific hardware details such as GPU models, CPU types, or memory specifications used for running the experiments.
Software Dependencies No The paper describes implementation details and training strategies, but it does not specify software dependencies (e.g., libraries, frameworks) along with their version numbers.
Experiment Setup Yes For training on CIFAR-10, we choose (η, ϵ, λmin) = (0.05, 0.6, 0.01). The rest of the training details are as follows. We use the modified cross-entropy loss function from (Prach & Lampert, 2022), ... We use τ = 0.25. Furthermore, we train our models for 1000 epochs with a batch size of 256 with a cosine annealing strategy with an initial learning rate of 10 4 and a final learning rate 10 5, and report the average results on two seed in Table 1.