Provable Efficient Online Matrix Completion via Non-convex Stochastic Gradient Descent

Authors: Chi Jin, Sham M. Kakade, Praneeth Netrapalli

NeurIPS 2016 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Theoretical In this paper, we propose the first provable, efficient online algorithm for matrix completion. Our proofs introduce a general framework to show that SGD updates tend to stay away from saddle surfaces and could be of broader interests to other non-convex problems.
Researcher Affiliation Collaboration Chi Jin UC Berkeley chijin@cs.berkeley.edu Sham M. Kakade University of Washington sham@cs.washington.edu Praneeth Netrapalli Microsoft Research India praneeth@microsoft.com
Pseudocode Yes Algorithm 1 Online Algorithm for PSD Matrix Completion. Algorithm 2 Online Algorithm for Matrix Completion (Theoretical) Algorithm 3 Online Algorithm for Matrix Completion (Practical)
Open Source Code No The paper does not contain any explicit statements or links indicating that the source code for the described methodology is publicly available.
Open Datasets No The paper does not specify a publicly available or open dataset used for training. It refers to 'Initial set of uniformly random samples Ωinit' without providing access information.
Dataset Splits No The paper is theoretical and does not describe specific training, validation, or test dataset splits.
Hardware Specification No The paper is theoretical and does not describe the specific hardware used for running experiments.
Software Dependencies No The paper is theoretical and does not list any specific software dependencies with version numbers.
Experiment Setup No The paper discusses theoretical parameters like learning rate (η) and number of observations (T) but does not provide concrete hyperparameter values or system-level training settings for an empirical experimental setup.