Calibrated Data-Dependent Constraints with Exact Satisfaction Guarantees

Authors: Songkai Xue, Yuekai Sun, Mikhail Yurochkin

NeurIPS 2022 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental In Section 4, we validate our theory by simulating a resource-constrained newsvendor problem. In Section 5, we demonstrate the efficacy of our method by using it to train an algorithmically fair income classifier. In Figure 1, we plot frequencies of constraint satisfaction for each constraint and both constraints, all of which are averaged over 1000 replicates. In Figure 2, we have line plots for frequency of constraint satisfaction and box plots for classification error rate, all of which are summarized over 100 replicates.
Researcher Affiliation Collaboration Songkai Xue Department of Statistics University of Michigan sxue@umich.edu Yuekai Sun Department of Statistics University of Michigan yuekai@umich.edu Mikhail Yurochkin IBM Research MIT-IBM Watson AI Lab mikhail.yurochkin@ibm.com
Pseudocode Yes Algorithm 1 Dual ascent algorithm for (1.3)
Open Source Code Yes Did you include the code, data, and instructions needed to reproduce the main experimental results (either in the supplemental material or as a URL)? [Yes]
Open Datasets Yes using the Adult dataset from UCI [13]. [13] Dheeru Dua and Casey Graff. UCI machine learning repository, 2017. URL http://archive.ics.uci.edu/ml.
Dataset Splits No The paper mentions 'training data' in theoretical discussions and states using the Adult dataset for experiments, but it does not specify explicit training/validation/test dataset splits (e.g., percentages, sample counts, or citations to predefined splits) for its own experiments in the main text.
Hardware Specification No Did you include the total amount of compute and the type of resources used (e.g., type of GPUs, internal cluster, or cloud provider)? [No]
Software Dependencies No The paper mentions using 'standard stochastic optimization algorithms' and 'logistic regression model for classification', but it does not specify any software names with version numbers (e.g., Python, PyTorch, TensorFlow versions, or specific solver versions) that would be needed to reproduce the experiment.
Experiment Setup No The paper states using a 'logistic regression model for classification' and notes the 'nominal probability' values tested. However, it does not explicitly provide concrete hyperparameter values (e.g., learning rate, batch size, number of epochs) or detailed training configurations for the experiments in the main text.