Alternating minimization for dictionary learning with random initialization
Authors: Niladri Chatterji, Peter L. Bartlett
NeurIPS 2017 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Theoretical | We present theoretical guarantees for an alternating minimization algorithm for the dictionary learning/sparse coding problem. Our guarantees are under a reasonable generative model that allows for dictionaries with growing operator norms, and can handle an arbitrary level of overcompleteness, while having sparsity that is information theoretically optimal. Our main contribution is to present new conditions under which alternating minimization for dictionary learning converges at a linear rate to the optimum. |
| Researcher Affiliation | Academia | Niladri S. Chatterji University of California, Berkeley chatterji@berkeley.edu; Peter L. Bartlett University of California, Berkeley peter@berkeley.edu |
| Pseudocode | Yes | Algorithm 1: Alternating Minimization for Dictionary Learning |
| Open Source Code | No | The paper does not provide any links to open-source code or explicitly state that code for the described methodology is available. |
| Open Datasets | No | The paper relies on a generative model for samples ( |
| Dataset Splits | No | The paper is theoretical and does not conduct empirical experiments with datasets. Therefore, it does not provide information about dataset splits for training, validation, or testing. |
| Hardware Specification | No | The paper is theoretical and does not conduct empirical experiments. Therefore, it does not specify any hardware used for running experiments. |
| Software Dependencies | No | The paper is theoretical and does not conduct empirical experiments. Therefore, it does not list specific software dependencies with version numbers. |
| Experiment Setup | No | The paper specifies parameters and their valid ranges for the theoretical algorithm (e.g., step size η, thresholds, tuning parameters λ, ν, γ), which are essential for its theoretical analysis and convergence guarantees. However, these are conditions for the theoretical model and are not described as concrete 'hyperparameter values' or 'training configurations' for an empirical experimental setup. The paper focuses on theoretical proofs rather than empirical implementation details. |