Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].
Subspace Clustering via New Low-Rank Model with Discrete Group Structure Constraint
Authors: Feiping Nie, Heng Huang
IJCAI 2016 | Venue PDF | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | Both synthetic and real world datasets demonstrate our proposed model s effectiveness. 5 Experiments 5.1 Experimental Results on Synthetic Data 5.2 Experiments on Real World Data |
| Researcher Affiliation | Academia | Department of Computer Science and Engineering University of Texas at Arlington EMAIL, EMAIL |
| Pseudocode | Yes | Algorithm 1 Algorithm to solve problem (5). Algorithm 2 Algorithm to solve problem (19). |
| Open Source Code | No | No explicit statement or link to open-source code for the methodology is provided. |
| Open Datasets | Yes | We ο¬rst test our model on the Hopkins 155 motion dataset. This dataset consists of 155 sequences... (cites [Costeira et al., 1997]), face datasets including JAFFE (cites [Lyons et al., 1998]), MSRA (cites [Liu et al., 2007]), XM2VTS [XM2, ] and another human palm image dataset called PALM [Yan et al., 2007] are used. [XM2, ] links to http://www.ee.surrey.ac.uk/cvssp/xm2vtsdb/. |
| Dataset Splits | No | The paper discusses data preprocessing and initialization methods but does not provide specific training/validation/test dataset splits (e.g., percentages or sample counts) or mention cross-validation. |
| Hardware Specification | No | No specific hardware details (e.g., CPU/GPU models, memory specifications, or cloud instance types) used for running experiments are mentioned. |
| Software Dependencies | No | No specific software dependencies with version numbers (e.g., Python 3.x, PyTorch 1.x) are mentioned for replication. |
| Experiment Setup | Yes | We run our algorithm with ten different initializations and select the results with the best objective values. We use the PCA to project the coordinates in each sequence into the dimensions ranging from 5 to 20. Then we use K-means method to get our initialized Gi(1 i k). In the data, we also add 5% level noises to deviate from the subspaces. |