Kernel-Based Tests for Likelihood-Free Hypothesis Testing
Authors: Patrik Robert Gerber, Tianze Jiang, Yury Polyanskiy, Rui Sun
NeurIPS 2023 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | In this work we (a) introduce a generalization where unlabeled samples come from a mixture of the two classes a case often encountered in practice; (b) study the minimax sample complexity for non-parametric classes of densities under maximum mean discrepancy (MMD) separation; and (c) investigate the empirical performance of kernels parameterized by neural networks on two tasks: detection of the Higgs boson and detection of planted DDPM generated images amidst CIFAR-10 images. For both problems we confirm the existence of the theoretically predicted asymmetric m vs n trade-off. |
| Researcher Affiliation | Academia | Patrik RĂ³bert Gerber Department of Mathematics, MIT Cambridge, MA 02139 prgerber@mit.edu Tianze Jiang Department of Mathematics, MIT Cambridge, MA 02139 tjiang@mit.edu Yury Polyanskiy Department of EECS, MIT Cambridge, MA 02139 yp@mit.edu Rui Sun Department of Mathematics, MIT Cambridge, MA 02139 eruisun@mit.edu |
| Pseudocode | Yes | Algorithm 1 m LFHT with a learned deep kernel |
| Open Source Code | Yes | Our code can be found at https://github.com/Sr-11/LFI. |
| Open Datasets | Yes | detection of the Higgs boson and detection of planted DDPM generated images amidst CIFAR-10 images. |
| Dataset Splits | Yes | Consider splitting the data into three parts: (Xtr, Y tr) is used for training (optimizing) the kernel; (Xev, Y ev) is used to evaluate our test statistic at test time; and (Xcal, Y cal) is used for calibrating the distribution of the test statistic under the null hypothesis. |
| Hardware Specification | Yes | Our code is implemented in Python 3.7 (Py Torch 1.1) and was ran on an NVIDIA RTX 3080 GPU equipped with a standard torch library and dataset extensions. |
| Software Dependencies | Yes | Our code is implemented in Python 3.7 (Py Torch 1.1) and was ran on an NVIDIA RTX 3080 GPU equipped with a standard torch library and dataset extensions. |
| Experiment Setup | Yes | We use 80 training epochs for most of our code from the CNN architecture (for classifiers, this is well after interpolating the training data and roughly when validation loss stops decreasing), and a batch size of 32 which has a slight empirical benefit compared to larger batch sizes. The learning rates are tuned separately in MMD methods for optimality, whereas for classifiers they follow the discriminator s original setting from [40]. In Phase 2 of Algorithm 1, we choose k = 1000 for the desired precision while not compromising runtime. |