Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Sparse Bayesian Generative Modeling for Compressive Sensing

Authors: Benedikt Bรถck, Sadaf Syed, Wolfgang Utschick

NeurIPS 2024 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We support our approach theoretically through the concept of variational inference and validate it empirically using different types of compressible signals. and 4 Experiments
Researcher Affiliation Academia Benedikt Bรถck, Sadaf Syed, Wolfgang Utschick TUM School of Computation, Information and Technology Technical University of Munich EMAIL
Pseudocode Yes N Pseudo-Code for the Training and Inference of the CSVAE and CSGMM
Open Source Code Yes 1Source code is available at https://github.com/beneboeck/sparse-bayesian-gen-mod.
Open Datasets Yes We use the MNIST dataset (N = 784) for evaluation [42, (CC-BY-SA 3.0 license)]. and We also use a dataset of 64 ร— 64 cropped celeb A images (N = 3 ร— 642 = 12288) [44] and evaluate on the Fashion MNIST dataset (N = 784) in Appendix L [45, (MIT license)].
Dataset Splits Yes We once reduce the learning rate by a factor of 2 during training and stop the training, when the modified ELBO in (15) for a validation set of 5000 samples does not increase.
Hardware Specification Yes All models have been simulated on an NVIDIA A40 GPU except for the proposed CSGMM, whose experiments have been conducted on an Intel(R) Xeon(R) Gold 6134 CPU @ 3.20GHz.
Software Dependencies No The paper mentions software like 'Adam' for optimization and 'Pywavelets' for wavelet analysis, but does not provide specific version numbers for any software dependencies.
Experiment Setup Yes For CSGMM, we set the number K of components to 32... The CSVAE encoders and decoders contain two fully-connected layers with Re LU activation... The latent dimension is set to 16, the learning rate is set to 2 ร— 10โˆ’5, and the batch size is set to 64. We use the Adam optimizer for optimization [47].