Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Towards Intelligent Visual Understanding under Minimal Supervision

Authors: Dingwen Zhang

IJCAI 2016 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental The experimental results have demonstrated the effectiveness of the proposed algorithms. Comprehensive evaluations of three benchmark datasets and comparisons with nine state-of-the-art algorithms demonstrate the superiority of this work.
Researcher Affiliation Academia Dingwen Zhang School of Automation, Northwestern Polytechnical University EMAIL
Pseudocode No The paper does not contain any pseudocode or clearly labeled algorithm blocks.
Open Source Code No The paper does not provide any links to open-source code or explicitly state that the code is publicly available.
Open Datasets Yes Comprehensive evaluations of three benchmark datasets and comparisons with nine state-of-the-art algorithms demonstrate the superiority of this work. [Han et al., 2015 a] J. Han, D. Zhang, X. Hu, L. Guo, F. Wu. Background Prior-Based Salient Object Detection via Deep Reconstruction Residual. TCSVT, 25(8): 1309-1321, 2015. [Zhang et al., 2015 a] D. Zhang, J. Han, C. Li, J. Wang. Co-saliency detection via looking deep and wide. In CVPR, pages 2994-3002, 2015.
Dataset Splits No The paper mentions evaluating on benchmark datasets but does not specify any training, validation, or test splits, nor does it refer to standard splits with citations.
Hardware Specification No The paper does not provide any specific details about the hardware (e.g., GPU, CPU models, memory) used for running the experiments.
Software Dependencies No The paper mentions the use of "deep learning architectures" but does not specify any software names with version numbers (e.g., libraries, frameworks, or specific solvers).
Experiment Setup No The paper describes the general approaches (e.g., "stacked denoising autoencoders with deep learning architectures"), but it does not provide specific experimental setup details such as hyperparameters (learning rate, batch size, number of epochs), optimizer settings, or model initialization details.