Learning to search efficiently for causally near-optimal treatments
Authors: Samuel Håkansson, Viktor Lindblom, Omer Gottesman, Fredrik D. Johansson
NeurIPS 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | The methods are evaluated on synthetic and real-world healthcare data and compared to model-free reinforcement learning. We find that our methods compare favorably to the model-free baseline while offering a more transparent trade-off between search time and treatment efficacy. |
| Researcher Affiliation | Academia | Samuel H akansson University of Gothenburg samuel.hakansson@gu.se Viktor Lindblom Chalmers University of Technology viklindb@student.chalmers.se Omer Gottesman Brown University omer_gottesman@brown.edu Fredrik D. Johansson Chalmers University of Technology fredrik.johansson@chalmers.se |
| Pseudocode | No | The paper describes dynamic programming and greedy approximation algorithms but does not include a clearly labeled pseudocode or algorithm block. |
| Open Source Code | Yes | Implementations can be found at: https://github.com/Healthy-AI/Treatment Exploration |
| Open Datasets | Yes | We evaluate our proposed methods using synthetic and real-world healthcare data... MIMIC-III database (Johnson et al., 2016). We consider training sets in a low-data regime with 50 samples and a high-data regime of 75000, with fixed test set size of 3000 samples. The cohort was split randomly into a training and test set with a 70/30 ratio and experiments were repeated over five such splits. |
| Dataset Splits | Yes | We consider training sets in a low-data regime with 50 samples and a high-data regime of 75000, with fixed test set size of 3000 samples. The cohort was split randomly into a training and test set with a 70/30 ratio and experiments were repeated over five such splits. |
| Hardware Specification | No | The paper does not provide specific hardware details (e.g., CPU/GPU models, memory) used for running the experiments. |
| Software Dependencies | No | The paper mentions methods like "function approximation using random forests" and "historical kernel-smoothing" but does not specify software dependencies with version numbers (e.g., Python version, library names and their versions). |
| Experiment Setup | Yes | Here, CDP and CG use δ = 0.4, = 0 and the upper bound of (10) and NDP λ = 0.35. We sweep all hyperparameters uniformly over 10 values; for CDP, CG, δ 2 [0, 1], for NDP H, λ 2 [0, 0.5] and for NDP F, λ 2 [0, 1]. |