Efficient Pruning of Large Knowledge Graphs

Authors: Stefano Faralli, Irene Finocchi, Simone Paolo Ponzetto, Paola Velardi

IJCAI 2018 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We apply our algorithm to the tasks of pruning automatically acquired taxonomies using benchmarking data from a Sem Eval evaluation exercise, as well as the extraction of a domain-adapted taxonomy from the Wikipedia category hierarchy. The results show the superiority of our approach over state-of-art algorithms in terms of both output quality and computational efficiency.
Researcher Affiliation Academia 1 University of Rome Unitelma Sapienza 2 University of Rome Sapienza 3 University of Mannheim
Pseudocode Yes Algorithm 1: CRUMBTRAIL
Open Source Code No The paper does not provide concrete access to source code for the methodology described in this paper.
Open Datasets Yes We use the gold-standard taxonomies and the taxonomy learning systems from Task 17 of the Sem Eval 2015 challenge2. 2http://alt.qcri.org/semeval2015/task17/ ... We apply the four pruning algorithms to the entire Wikipedia category graph...
Dataset Splits No The paper evaluates performance against gold-standard and reference graphs but does not specify explicit training/validation/test dataset splits for reproducibility in the context of model training or hyperparameter tuning.
Hardware Specification No The paper mentions issues with CLE running out of memory 'on a multi-core machine' but does not provide specific hardware details like CPU/GPU models, processor types, or memory amounts used for the experiments.
Software Dependencies No The paper does not provide specific ancillary software details with version numbers, such as library or solver names and their versions.
Experiment Setup No The paper describes the algorithms and evaluation methods but does not provide specific experimental setup details such as hyperparameter values or training configurations for the algorithms themselves.