Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
NCDL: A Framework for Deep Learning on non-Cartesian Lattices
Authors: Joshua Horacsek, Usman Alim
NeurIPS 2023 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | In this section, we delineate the experiments we conducted to evaluate NCDL. First, we compare NCDL with the closest existing hexagonal convolution library [34]. Subsequently, we explore the use of non-dyadic down/up sampling within bottlenecked architectures. All experiments are conducted on an AMD Ryzen 9 3900X (3.8GHz) with 128GB of DDR4 RAM operating at 3200 MHz, accompanied by an NVIDIA RTX 3090 with 24GB of RAM. |
| Researcher Affiliation | Academia | Joshua J. Horacsek Department of Computer Science University of Calgary Calgary, Alberta EMAIL Usman R. Alim Department of Computer Science University of Calgary Calgary, Alberta EMAIL |
| Pseudocode | No | The paper contains mathematical definitions and propositions, but no structured pseudocode or algorithm blocks are present. |
| Open Source Code | Yes | Ultimately, we present a software library called Non-Cartesian Deep Learning (NCDL) which is an open source, concrete implementation of the lattice tensor container and the associated spatio-temporal operations defined over lattice tensors. ... Our implementation is available at https://www.ncdl.ai. |
| Open Datasets | Yes | We train our models using the Celeb A dataset [30]. ... We train our models on the DUTS salient object detection dataset [40] |
| Dataset Splits | No | Employing a straightforward L1 loss, we measure validation L1, L2, PSNR, SSIM [41] and LPIPs metrics [46]. ... we measure validation BCE, L1, L2 and SSIM. The paper mentions โvalidationโ but does not provide specific details on the dataset splits (e.g., percentages or sample counts) used for training, validation, or testing. |
| Hardware Specification | Yes | All experiments are conducted on an AMD Ryzen 9 3900X (3.8GHz) with 128GB of DDR4 RAM operating at 3200 MHz, accompanied by an NVIDIA RTX 3090 with 24GB of RAM. |
| Software Dependencies | No | NCDL library is implemented on top of Py Torch [31]". The paper mentions PyTorch but does not provide specific version numbers for software dependencies needed for replication. |
| Experiment Setup | Yes | The network is trained with the Adam optimizer, default parameters, and a batch size of 8. We train for 300,000 iterations, as convergence was observed at this point, and take an average of 5 runs. ... The network is trained with the Adam optimizer, with default parameters, and a batch size of 8. We train for 120,000 iterations, as convergence was observed in the validation data at this point. |