Toward a neuro-inspired creative decoder

Authors: Payel Das, Brian Quanz, Pin-Yu Chen, Jae-wook Ahn, Dhruv Shah

IJCAI 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Our experiments on different image datasets (MNIST, FMNIST, MNIST+FMNIST, Wiki Art and Celeb A) reveal that atypical co-activation of highly activated and weakly activated neurons in a deep decoder promotes generation of novel and meaningful artifacts.
Researcher Affiliation Industry Payel Das , Brian Quanz , Pin-Yu Chen , Jae-wook Ahn and Dhruv Shah IBM Research, Yorktown Heights, NY, USA {daspa,blquanz}@us.ibm.com, pin-yu.chen@ibm.com, jaewook.ahn@us.ibm.com, dhruv.shah@ibm.com
Pseudocode Yes Algorithm 1 Low-active method
Open Source Code No The paper does not provide an explicit statement or link indicating that the source code for the described methodology is publicly available.
Open Datasets Yes We show the performance of the proposed method against MNIST digits , FMINST fashion objects , and on a combined MNIST plus FMINST dataset. We also present results on the Wiki Art art images and Celeb A faces.
Dataset Splits No The paper uses standard datasets like MNIST and FMNIST and mentions training and testing but does not explicitly state the specific dataset splits (e.g., percentages or counts for train/validation/test sets) used for reproducibility.
Hardware Specification No The paper does not provide specific details about the hardware used to run the experiments, such as GPU/CPU models or memory specifications.
Software Dependencies No The paper mentions deep learning frameworks like VAE and GAN but does not specify software dependencies with version numbers (e.g., Python, PyTorch, or TensorFlow versions).
Experiment Setup Yes For F/MNIST the encoder network consisted of 3 fully-connected layers (1000, 500, 250) before the z output (50 for F/MNIST and 100 for the combination), with the decoder architecture the reverse of the encoder. RELU activations were used; dropout equal to 0.10 for fully-connected layers was used during training only. ... Unless otherwise stated, results in the main paper were obtained by perturbing five neurons during decoding. For the low-active method, we used neurons whose activations (see Method Section) were within the 1st and 15th percentiles of the neuron percent activations (ak j ) for the layer.