RenderGAN: Generating Realistic Labeled Data

Authors: Leon Sixt, Benjamin Wild, Tim Landgraf

ICLR 2017 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Training a DCNN on data generated by the Render GAN yields considerably better performance than training it on various baselines. We furthermore include a previously used computer vision pipeline in the evaluation. The networks detections are used as feature to track the honeybees over time. When we use detections from the DCNN instead of the computer vision pipeline, the accuracy of assigning the true id increases from 55% to 96%.
Researcher Affiliation Academia Leon Sixt, Benjamin Wild, & Tim Landgraf Fachbereich Mathematik und Informatik Freie Universit at Berlin Berlin, Germany {leon.sixt, benjamin.wild, tim.landgraf}@fu-berlin.de
Pseudocode No The paper does not contain structured pseudocode or algorithm blocks.
Open Source Code Yes We will outline only the most important parts. See our code available online for all the details of the networks1. 1https://github.com/berleon/deepdecoder
Open Datasets No The paper mentions using
Dataset Splits Yes As the data is correlated in time (same ids, similar lighting conditions), we assign the data from one time span completely to either the train or test set. The data from three time spans forms the train set (40K). The test set (26K) contains data from the remaining two time spans. ... For the training with generated data, we use the real training set as the validation set. When training on real data, the test set is also used for validation. We could alternatively reduce the real training set further and form an extra validation set, but this would harm the performance of the DCNN trained on the real data.
Hardware Specification No The paper does not provide specific hardware details (e.g., exact GPU/CPU models, memory amounts) used for running its experiments.
Software Dependencies No The paper mentions using Adam as an optimizer but does not provide specific version numbers for any software dependencies.
Experiment Setup Yes Training: We train generator and discriminator as in the normal GAN setting. We use 2.4M unlabeled images of tags to train the Render GAN. We use Adam (Kingma & Ba, 2014) as an optimizer with a starting learning rate of 0.0002, which we reduce in epoch 200, 250, and 300 by a factor of 0.25. ... An epoch consists of 1000 batches a 128 samples. We use early stopping to select the best parameters of the networks. ... We use the 34-layer Res Net architecture (He et al., 2015) but start with 16 feature maps instead of 64. The DCNNs are evaluated on the mean Hamming distance (MHD) i.e. the expected value of bits decoded wrong. ... For us γ = 15 worked well.