Predtron: A Family of Online Algorithms for General Prediction Problems
Authors: Prateek Jain, Nagarajan Natarajan, Ambuj Tewari
NeurIPS 2015 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | A simulation study confirms the behavior predicted by our bounds and demonstrates the flexibility of the design choices in our framework. We now present simulation results to demonstrate the application of our proposed Predtron framework to subset ranking. We also demonstrate that empirical results match the trend predicted by our error bounds, hence hinting at tightness of our (upper) bounds. Figure 1 (a) shows LNDCG (on a test set) for our Predtron algorithm... |
| Researcher Affiliation | Collaboration | Prateek Jain Microsoft Research, INDIA prajain@microsoft.com Nagarajan Natarajan University of Texas at Austin, USA naga86@cs.utexas.edu Ambuj Tewari University of Michigan, Ann Arbor, USA tewaria@umich.edu |
| Pseudocode | Yes | Algorithm 1 Predtron: Extension of the Perceptron Algorithm to General Prediction Problems |
| Open Source Code | No | The paper does not contain any explicit statements about releasing open-source code for the described methodology, nor does it provide links to a code repository. |
| Open Datasets | No | We generated n data points X 2 Rm p0 using a Gaussian distribution with independent rows. The ith row of X represents a document and is sampled from a spherical Gaussian centered at µi. We selected a w 2 Rp0 and also a set of thresholds [ 1, . . . , m+1] to generate relevance scores... |
| Dataset Splits | No | The paper discusses generating synthetic data and evaluating on a 'test set', but it does not provide specific details about training, validation, and test splits (e.g., percentages, sample counts, or references to standard predefined splits). |
| Hardware Specification | No | The paper does not provide any specific details about the hardware (e.g., GPU/CPU models, memory, or cloud instances) used for running the experiments. |
| Software Dependencies | No | The paper does not provide any specific software dependencies with version numbers (e.g., library names, frameworks, or solvers with their corresponding versions). |
| Experiment Setup | No | The paper describes how the synthetic data was generated ('We generated n data points X 2 Rm p0 using a Gaussian distribution...'), and explores different choices for `pred_1` functions, varying 'n' (number of training points), 'm' (number of documents), and 'p0' (data dimensionality). However, it does not specify concrete training hyperparameters for the Predtron algorithm itself, such as learning rate, batch size, or number of epochs. |