Efficient Learning of Mahalanobis Metrics for Ranking
Authors: Daryl Lim, Gert Lanckriet
ICML 2014 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | To evaluate our proposed method, we conducted two sets of experiments. |
| Researcher Affiliation | Academia | Daryl K. H. Lim DKLIM@UCSD.EDU Department of Electrical and Computer Engineering, University of California, San Diego, CA 92093 USA |
| Pseudocode | Yes | Algorithm 1 Symmetric gradient update |
| Open Source Code | No | The paper does not provide a link to its own source code or explicitly state that it is open-source. It references a third-party code for LMNN. |
| Open Datasets | Yes | Image Net (Deng et al., 2009), CAL10K (Tingle et al., 2010) and Magna Tagatune (Law et al., 2009)... covertype dataset from the UCI repository |
| Dataset Splits | Yes | For each method, the hyperparameters with the best MAP performance on the validation set were selected. |
| Hardware Specification | No | The paper does not provide specific details regarding the hardware (e.g., CPU, GPU models, memory, or cloud resources) used for running the experiments. |
| Software Dependencies | No | The paper mentions various algorithms and implementations used (e.g., MLR-ADMM, LMNN 2.4 code) but does not provide specific version numbers for its own software dependencies or general software environment required for replication. |
| Experiment Setup | Yes | For both variants of FRML, the trade-off parameter λ was fixed at 0.1 and we used a minibatch of size 5. |