Regional Tree Regularization for Interpretability in Deep Neural Networks

Authors: Mike Wu, Sonali Parbhoo, Michael Hughes, Ryan Kindle, Leo Celi, Maurizio Zazzi, Volker Roth, Finale Doshi-Velez6413-6421

AAAI 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Across many datasets, including two healthcare applications, we show our approach delivers simpler explanations than other regularization schemes without compromising accuracy. Specifically, our regional regularizer finds many more desirable optima compared to global analogues.
Researcher Affiliation Collaboration Mike Wu,1 Sonali Parbhoo,2,3 Michael C. Hughes,4 Ryan Kindle,5 Leo Celi,6 Maurizio Zazzi,7 Volker Roth,2 Finale Doshi-Velez3 1Stanford University, wumike@stanford.edu 2University of Basel, volker.roth@unibas.ch 3Harvard University SEAS, {sparbhoo, finale}@seas.harvard.edu 4Tufts University, michael.hughes@tufts.edu 5Massachusetts General Hospital, rkindle@mgh.harvard.edu 6Massachusetts Institute of Technology, lceli@mit.edu 7University of Siena, maurizio.zazzi@unisi.it
Pseudocode Yes Algorithm 1 APL (Wu et al., 2018)
Open Source Code Yes Py Torch implementation is available at https://github.com/mhw32/regional-tree-regularizer-public.
Open Datasets Yes We now apply regional tree regularization to four datasets from the UC Irvine repository (Dheeru and Karra Taniskidou, 2017). [...] The critical care task, performed with the MIMIC dataset (Johnson et al., 2016), [...] The HIV task, performed with the EUResist dataset (Zazzi et al., 2011)
Dataset Splits Yes convergence is measured by APL and accuracy on a validation set that does not change for at least 10 epochs
Hardware Specification No Computations were supported by the FAS Research Computing Group at Harvard and sci CORE (http://scicore.unibas.ch/) scientific computing core facility at University of Basel. This describes the facility, not specific hardware components like CPU/GPU models or memory.
Software Dependencies No The paper mentions software like 'Scikit-Learn' and 'Py Torch' but does not specify version numbers for these or any other software dependencies.
Experiment Setup Yes We train each regularizer with an exhaustive set of strengths: λ = 0.0001, 0.0005, 0.001, 0.005, 0.01, 0.02, 0.05, 0.1, 0.2, 0.5, 1.0, 2.0, 5.0, 10.0. Three runs with different random seeds were used to avoid local optima.