Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Composite Feature Selection Using Deep Ensembles

Authors: Fergus Imrie, Alexander Norcliffe, Pietro Liรณ, Mihaela van der Schaar

NeurIPS 2022 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We evaluate Comp FS using several synthetic and semi-synthetic datasets where ground truth feature importances and group structure are known. In addition, we illustrate our method on an image dataset (MNIST) and a real-world cancer dataset (METABRIC).
Researcher Affiliation Academia Fergus Imrie University of California, Los Angeles EMAIL Alexander Norcliffe University of Cambridge EMAIL Pietro Liรฒ University of Cambridge EMAIL Mihaela van der Schaar University of Cambridge The Alan Turing Institute University of California, Los Angeles EMAIL
Pseudocode No The paper describes its methodology in prose and figures, but it does not include a dedicated section or block explicitly labeled as 'Pseudocode' or 'Algorithm'.
Open Source Code Yes The code for our method and experiments is available on Github. 2 3
Open Datasets Yes We use several of the datasets constructed by [62], some of which were also used by [75]. (Section 5.2 Semi-Synthetic Experiments) We investigate Comp FS on the MNIST dataset [48]. (Section 5.3 Image Dataset: MNIST) Finally, we assess Comp FS on a real-world dataset, METABRIC [21, 68] (Section 5.4 Real-World Data: METABRIC)
Dataset Splits Yes For all datasets, we hold out 10% of the data for validation and use 10% of the data for testing, using the remaining 80% for training.
Hardware Specification Yes All experiments were run on a single NVIDIA GeForce RTX 3090 GPU or on a commercially available laptop with an Intel i7-10750H CPU and 16GB of RAM.
Software Dependencies No The paper mentions the use of an Adam optimizer but does not specify software dependencies with version numbers (e.g., specific Python versions, machine learning frameworks like PyTorch or TensorFlow with their versions, or library versions).
Experiment Setup Yes For all our experiments, we use a two-layer feedforward neural network with 128 hidden units and ReLU non-linearities for the group encoder and predictor network. We use a batch size of 128 and train with the Adam optimizer [44] with learning rate 0.001. (Appendix C) and Hyperparameters for each experiment are provided in Table 4. (Question 3b in checklist, Table 4 provides specific values for learning rate, batch size, epochs, etc.)