Towards Establishing Guaranteed Error for Learned Database Operations

Authors: Sepanta Zeighami, Cyrus Shahabi

ICLR 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental In this paper, we embark on the first theoretical study of such guarantees for learned methods, presenting the necessary conditions for such guarantees to hold when using machine learning to perform indexing, cardinality estimation and range-sum estimation. and 4 EMPIRICAL RESULTS We present experiments comparing our bounds with the error obtained by training different models on datasets sampled from different distributions.
Researcher Affiliation Academia Sepanta Zeighami UC Berkeley zeighami@berkeley.edu Cyrus Shahabi University of Southern California shahabi@usc.edu
Pseudocode No The paper describes methods and proofs using mathematical notation and prose but does not include any clearly labeled pseudocode or algorithm blocks.
Open Source Code No The paper does not provide any statements about open-sourcing its code or links to a code repository for the methodology described.
Open Datasets No The paper states: 'We consider 1-dimensional datasets sampled from uniform and 2-component Gaussian mixture model distributions.' It describes how data was generated but does not provide concrete access information (link, DOI, or specific citation with authors/year) to a pre-existing publicly available dataset.
Dataset Splits No The paper discusses training and testing concepts in the context of learned models but does not provide specific train/validation/test dataset splits (e.g., percentages, counts, or references to standard splits) for its own empirical experiments.
Hardware Specification No The paper does not provide any specific details about the hardware (e.g., GPU models, CPU types) used for running its experiments.
Software Dependencies No The paper mentions types of models and general techniques but does not list specific software dependencies (libraries, frameworks) with their version numbers.
Experiment Setup No The paper mentions 'empirical hyperparameter tuning' generally but does not provide specific experimental setup details such as concrete hyperparameter values (e.g., learning rate, batch size, epochs) or training configurations for the models evaluated in its empirical section.