Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Fast k-Nearest Neighbour Search via Dynamic Continuous Indexing

Authors: Ke Li, Jitendra Malik

ICML 2016 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We show appealing theoretical properties and demonstrate empirically that the proposed algorithm outperforms locality-sensitivity hashing (LSH) in terms of approximation quality, speed and space ef๏ฌciency. ... We compare the performance of the proposed algorithm to that of LSH... We compare the performance of the proposed algorithm and LSH on the CIFAR-100 and MNIST datasets...
Researcher Affiliation Academia Ke Li EMAIL Jitendra Malik EMAIL University of California, Berkeley, CA 94720, United States
Pseudocode Yes Algorithm 1 Algorithm for data structure construction; Algorithm 2 Algorithm for k-nearest neighbour retrieval
Open Source Code No The paper does not provide any explicit statement or link indicating that the source code for the described methodology is publicly available.
Open Datasets Yes We compare the performance of the proposed algorithm and LSH on the CIFAR-100 and MNIST datasets, which consist of 32 32 colour images of various real-world objects and 28 28 grayscale images of handwritten digits respectively.
Dataset Splits No The paper describes combining the training and test sets and then selecting query points using cross-validation. It does not mention a separate validation split for model tuning.
Hardware Specification No The paper does not provide any specific hardware details (e.g., GPU models, CPU types, memory) used for running the experiments.
Software Dependencies No The paper mentions using 'self-balancing binary search trees or skip lists' and comparing against 'Exact Euclidean LSH (E2LSH)', but it does not specify any software names with version numbers for reproducibility (e.g., Python, PyTorch, or specific library versions).
Experiment Setup Yes We adopt the recommended guidelines for choosing parameters of LSH and used 24 hashes per table and 100 tables. For the proposed method, we used m = 25 and L = 2 on CIFAR-100 and m = 15 and L = 3 on MNIST, which we found to work well in practice.