Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Collective Biobjective Optimization Algorithm for Parallel Test Paper Generation

Authors: Minh Luan Nguyen, Siu Cheung Hui, Alvis C. M. Fong

IJCAI 2015 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experiment results have shown that CBO has drastically outperformed the current techniques in terms of paper quality and runtime efficiency.
Researcher Affiliation Collaboration Minh Luan Nguyen Data Analytics Department Institute for Infocomm Research Singapore; Siu Cheung Hui School of Computer Engineering Nanyang Technological University Singapore; Alvis C. M. Fong School of Computing Science University of Glasgow Scotland, UK
Pseudocode Yes Algorithm 1: Infeasible Allocation Detection; Algorithm 2: Total Quality Maximization
Open Source Code No The paper does not provide any statement or link indicating that the source code for the described methodology is publicly available.
Open Datasets No As there is no benchmark data available, we generated 4 large-sized synthetic datasets, namely D1, D2, D3 and D4, for performance evaluation.
Dataset Splits No The paper describes generating synthetic datasets and using them to evaluate the algorithm's performance in generating test papers, but it does not specify traditional training/validation/test dataset splits as would be relevant for training a machine learning model.
Hardware Specification No The paper does not provide any specific details regarding the hardware specifications (e.g., CPU, GPU models, memory) used to run the experiments.
Software Dependencies No The paper does not provide specific software dependencies with version numbers needed to replicate the experiment.
Experiment Setup Yes We vary the parameters in order to have different test criteria in the test specifications. The number of topics is specified between 2 and 40. The total time is set between 20 and 240 minutes, and it is also set proportional to the number of selected topics for each specification. The average difficulty degree is specified randomly between 3 and 9. We have conducted the experiments according to 5 different values of k, i.e., k = 1, 5, 10, 15, 20.