Anomaly Detection for Tabular Data with Internal Contrastive Learning

Authors: Tom Shenkar, Lior Wolf

ICLR 2022 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Our experiments show that our method leads by a sizable accuracy gap in comparison to the literature and that the same default rule of hyperparameters selection provides state-of-the-art results across benchmarks.
Researcher Affiliation Academia Tom Shenkar & Lior Wolf Blavatnik School of Computer Science, Tel Aviv University tomshenkar@gmail.com,wolf@cs.tau.ac.il
Pseudocode No No pseudocode or algorithm blocks were found.
Open Source Code Yes The full implementation of our method and scripts for reproducing the experiments are attached as a supplementary zip file. This archive includes a README file and a list of requirements that support seamless reproducibility.
Open Datasets Yes The second set employs the Multi-dimensional point datasets from the Outlier Detection Data Sets (ODDS)1. It contains 31 datasets, including two of the four datasets above. 1http://odds.cs.stonybrook.edu/, accessed January 2021
Dataset Splits No Following Zong et al. (2018); Bergman & Hoshen (2020), the training set contains a random subset of 50% of the normal data. The test set contains the rest of the normal data, as well as all the anomalies. No explicit mention of a validation set split was found.
Hardware Specification Yes The google colab infrastructure was used to run the experiments. GPU: Tesla K80, 12GB GDDR5 VRAM; CPU: Single core Xeon Processors @2.3Ghz; RAM: 24 GB. For the GOAD baseline that required larger memory, we used a 32GB GPU and 512GB RAM.
Software Dependencies No The paper mentions a "list of requirements" in a supplementary zip file but does not explicitly list specific software dependencies with version numbers in the main text.
Experiment Setup Yes We fix τ = 0.01 and u = 200. The value of u, which is often much larger than k, provides enough capacity throughout all experiments, without the need to tune it for each problem. We set k proportionally to the input dimension d. For d smaller than 40, we set k = 2, for d in the range [40, 160] we employ k = 10, and for d > 160, k takes the value d 150. Training employs the Adam optimizer (Kingma & Ba, 2014) with a learning rate of 10 3 .