Unsupervised Disentangled Representation Learning with Analogical Relations
Authors: Zejian Li, Yongchuan Tang, Yongxing He
IJCAI 2018 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | In Section 5 we demonstrate the experiment results and compare our method with other methods along the subspace score. In this section, we present the experiment results on five image datasets, including MNIST [Le Cun et al., 1998], Celeb A [Liu et al., 2015], Flower [Nilsback and Zisserman, 2008], CUB [Wah et al., 2011] and Chairs [Aubry et al., 2014]. Specifically, we compare our methods with other state-of-the-art methods along the subspace score. Table 1 reports the subspace score of models. |
| Researcher Affiliation | Academia | Zejian Li, Yongchuan Tang , Yongxing He College of Computer Science, Zhejiang University, Hangzhou 310027, China {zejianlee, yctang, heyongxing}@zju.edu.cn |
| Pseudocode | No | The paper does not contain any clearly labeled pseudocode or algorithm blocks. |
| Open Source Code | Yes | Our source code will be available on https://github.com/Zejian Li/analogical-training. |
| Open Datasets | Yes | In this section, we present the experiment results on five image datasets, including MNIST [Le Cun et al., 1998], Celeb A [Liu et al., 2015], Flower [Nilsback and Zisserman, 2008], CUB [Wah et al., 2011] and Chairs [Aubry et al., 2014]. |
| Dataset Splits | No | The paper mentions usage of 'train' and 'test' in the context of models and results, but does not provide specific train/validation/test dataset splits with percentages, counts, or a clear splitting methodology. |
| Hardware Specification | No | The paper does not provide specific hardware details such as GPU/CPU models, processor types, or memory amounts used for running experiments. |
| Software Dependencies | No | The paper mentions 'Py Torch' and 'scikit-learn' but does not provide specific version numbers for these software dependencies. |
| Experiment Setup | Yes | For Ana GAN, we use the network architecture of DCGAN [Radford et al., 2015]. We use the WGAN-GP loss [Gulrajani et al., 2017] instead of the original GAN loss and the parameter num critic is set as 3. At the first 100 epoch, we only optimize V (D, G)... Both the noise and the code are sampled from the standard normal distribution. We use Adam optimizer [Kingma and Ba, 2014] with a learning rate of 0.00002 and a momentum of 0.5. The batch size is 32. Ana VAE shares most of the configuration in Ana GAN. The encoder network Q borrows the major structure of D in Ana GAN. The learning rate for the Adam optimizer is 0.0001... We add a dropout layer [Srivastava et al., 2014] after each nonlinear activation layer in R to avoid overfitting. To compute the subspace score, a cluster of ten sample sequences is generated for each factor and each sequence has five samples. The sequence is generated by varying the corresponding component of the code from -2 to 2 with the interval 1 but keeping other components fixed. We compute the subspace score over five different sets of generated samples to get the average. |