Equivariant Neural Rendering
Authors: Emilien Dupont, Miguel Bautista Martin, Alex Colburn, Aditya Sankar, Josh Susskind, Qi Shan
ICML 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We perform experiments on Shape Net benchmarks (Chang et al., 2015) as well as on two new datasets designed to challenge the model on more complex scenes. |
| Researcher Affiliation | Collaboration | 1University of Oxford, UK 2Apple Inc, USA. Correspondence to: Emilien Dupont <dupont@stats.ox.ac.uk>, Qi Shan <qshan@apple.com>. |
| Pseudocode | No | The paper describes the model architecture and training process in text and diagrams (e.g., Figure 4, Figure 5) but does not include structured pseudocode or algorithm blocks. |
| Open Source Code | Yes | The code and datasets are available at https://github.com/ apple/ml-equivariant-neural-rendering. |
| Open Datasets | Yes | We perform experiments on Shape Net benchmarks (Chang et al., 2015) as well as on two new datasets designed to challenge the model on more complex scenes... The code and datasets are available at https://github.com/ apple/ml-equivariant-neural-rendering. |
| Dataset Splits | Yes | We evaluate our model on the Shape Net chairs class by following the experimental setup given in Sitzmann et al., using the same train/validation/test splits. |
| Hardware Specification | Yes | At inference time, SRNs require solving an optimization problem in order to fit a scene to the model. As such, inferring a scene representation from a single input image (on a Tesla V100 GPU) takes 2 minutes with SRNs but only 22ms for our model (three orders of magnitude faster). |
| Software Dependencies | No | The paper mentions using "Pytorch and Tensorflow" frameworks and the "Mitsuba renderer (Jakob, 2010)" but does not specify version numbers for these software dependencies. |
| Experiment Setup | Yes | For all experiments, the images are of size 128 128 and the scene representations are of size 64 32 32 32. For both the 2D and 3D parts of the network we use residual layers for convolutions that preserve the dimension of the input and strided convolutions for downsampling layers. We use the Leaky Re LU nonlinearity (Maas et al.) and Group Norm (Wu & He, 2018) for normalization. |