Uniform Convergence Rates for Kernel Density Estimation
Authors: Heinrich Jiang
ICML 2017 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Theoretical | We (1) derive finite-sample high-probability density estimation bounds for multivariate KDE under mild density assumptions which hold uniformly in x Rd and bandwidth matrices. We apply these results to (2) mode, (3) density level set, and (4) class probability estimation and attain optimal rates up to logarithmic factors. We then (5) provide an extension of our results under the manifold hypothesis. Finally, we (6) give uniform convergence results for local intrinsic dimension estimation. |
| Researcher Affiliation | Industry | 1Google. Correspondence to: Heinrich Jiang <heinrich.jiang@gmail.com>. |
| Pseudocode | No | The paper does not contain pseudocode or clearly labeled algorithm blocks. |
| Open Source Code | No | The paper does not provide any statement or link indicating that open-source code for the methodology is available. |
| Open Datasets | No | The paper is theoretical and does not use specific publicly available datasets for training or empirical evaluation. It refers to 'samples drawn from f' but no actual dataset names or access information. |
| Dataset Splits | No | The paper is theoretical and does not mention training/validation/test dataset splits, as it does not conduct empirical experiments. |
| Hardware Specification | No | The paper is theoretical and does not describe the hardware used, as it does not conduct empirical experiments. |
| Software Dependencies | No | The paper is theoretical and does not describe specific software dependencies with version numbers. |
| Experiment Setup | No | The paper is theoretical and does not provide details about an experimental setup, hyperparameters, or system-level training settings. |