Meta Two-Sample Testing: Learning Kernels for Testing with Limited Data
Authors: Feng Liu, Wenkai Xu, Jie Lu, Danica J. Sutherland
NeurIPS 2021 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We provide both theoretical justification and empirical evidence that our proposed meta-testing schemes outperform learning kernel-based tests directly from scarce observations, and identify when such schemes will be successful. |
| Researcher Affiliation | Academia | Feng Liu Australian AI Institute, UTS feng.liu@uts.edu.au Gatsby Unit, UCL xwk4813@gmail.com Jie Lu Australian AI Institute, UTS jie.lu@uts.edu.au Danica J. Sutherland UBC and Amii dsuth@cs.ubc.ca |
| Pseudocode | Yes | Algorithm 1 Meta Kernel Learning (Meta-KL), Algorithm 2 Testing with a Kernel Learner, Algorithm 3 Meta Multi-Kernel Learning (Meta-MKL) |
| Open Source Code | Yes | Implementation details are in Appendix B.1; the code is available at github.com/fengliu90/Meta Testing. |
| Open Datasets | Yes | We distinguish the standard datasets of CIFAR-10 and CIFAR-100 [52] from the attempted replication CIFAR-10.1 [53], similar to Liu et al. [16]. |
| Dataset Splits | Yes | Here, one divides the observed data into training and testing splits, identifies a kernel on the training data by maximizing a power criterion ˆJ, then runs an MMD test on the testing data (as illustrated in Figure 1a). [...] we set the number of training samples (Str P , Str Q ) to 50, 100, 150 per mode, and the number of testing samples (Ste P , Ste Q ) from 50 to 250. |
| Hardware Specification | No | No specific hardware details (GPU/CPU models, memory, etc.) are mentioned in the paper. |
| Software Dependencies | No | We use PyTorch [55] and Adam [56] for our implementations. |
| Experiment Setup | Yes | We use a learning rate of 1e-4 and 5 inner gradient steps for Meta-KL. We run 1000 outer iterations for the meta-training. [...] The deep neural network φ in Eq. (8) is composed of 3 fully connected layers with 1024, 512, 256 units respectively, and ReLU activation functions. |