Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Low-Rank Similarity Mining for Multimodal Dataset Distillation

Authors: Yue Xu, Zhilin Lin, Yusong Qiu, Cewu Lu, Yong-Lu Li

ICML 2024 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental 4. Experiments, The results on Flickr30k and COCO are shown in Tab. 2 and 3.
Researcher Affiliation Academia 1Shanghai Jiao Tong University. Correspondence to: Yong-Lu Li <yonglu EMAIL>.
Pseudocode Yes Algorithm 1 Low-Rank Similarity Mining (Lo RS)
Open Source Code Yes Our code is available at https: //github.com/silicx/Lo RS_Distill.
Open Datasets Yes We evaluate our method on Flickr30k (Plummer et al., 2015) and COCO (Lin et al., 2014)
Dataset Splits Yes We evaluate our method on Flickr30k (Plummer et al., 2015) and COCO (Lin et al., 2014) following the strong baseline (Wu et al., 2023) which exploits the MTT (Cazenavette et al., 2022) algorithm. The model performance is commonly measured by the recall of top K retrieval (R@K).
Hardware Specification Yes The experiments are conducted on one RTX 4090 GPU, revealing the efficiency of the method.
Software Dependencies No The paper mentions using NFNet and BERT-base models, and common deep learning frameworks are implied, but no specific software dependencies with version numbers are provided.
Experiment Setup Yes In the distillation stage, the images are resized to 224 224 and the text embeddings are 768-d. the synthetic data is learned with SGD and momentum 0.5. The rest hyperparameters including learning rate and Lo RS parameters vary among different datasets and synthetic data sizes, which are listed in Appendix Sec. F due to page limit.