Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
Counterfactual Prediction for Outcome-Oriented Treatments
Authors: Hao Zou, Bo Li, Jiangang Han, Shuiping Chen, Xuetao Ding, Peng Cui
ICML 2022 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | Extensive experimental results on both synthetic datasets and semi-synthetic datasets demonstrate the effectiveness of our method. |
| Researcher Affiliation | Collaboration | 1Department of Computer Science and Technology, Tsinghua University, Beijing, China; email:EMAIL, EMAIL 2School of Economics and Management, Tsinghua University, Beijing, China; email: EMAIL 3Meituan, Beijing, China; EMAIL. |
| Pseudocode | Yes | Algorithm 1 Outcome-Oriented Sample Re-weighting(OOSR) |
| Open Source Code | No | The paper does not provide any statement or link indicating the availability of open-source code for the described methodology. |
| Open Datasets | Yes | The confounder feature is obtained from a real-world dataset TCGA (Weinstein et al., 2013). |
| Dataset Splits | No | A sample set of size 10000 is randomly generated as held-out test-set to compute the out-of-sample metric. |
| Hardware Specification | No | The paper does not provide specific hardware details (e.g., GPU/CPU models, memory) used for running its experiments. |
| Software Dependencies | No | The paper mentions software components like 'neural networks', 'ELU activation function', and 'SGD optimizer', but does not provide specific software dependencies with version numbers (e.g., PyTorch 1.9, TensorFlow 2.x). |
| Experiment Setup | Yes | The predictive model is a neural networks with two hidden layers of size 20. The policy networks in IPS-Bandit Net is of the same architecture. We use the ELU activation function. The predictive models are trained by SGD optimizer for 60000 iterations in synthetic experiments, 100000 iterations in setting 1 of semi-synthetic experiments and 300000 iterations in setting 2 of semi-synthetic experiments. The policy networks are trained for 4000 epochs. For our algorithm, in each experiment, the length of the first stage is 40% of the training process, and the length of each round in the second stage is 5% of the training process. We choose "lambda" = 10.0 and "tau" = 0.2. |