Efficient Pruning of Large Knowledge Graphs
Authors: Stefano Faralli, Irene Finocchi, Simone Paolo Ponzetto, Paola Velardi
IJCAI 2018 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We apply our algorithm to the tasks of pruning automatically acquired taxonomies using benchmarking data from a Sem Eval evaluation exercise, as well as the extraction of a domain-adapted taxonomy from the Wikipedia category hierarchy. The results show the superiority of our approach over state-of-art algorithms in terms of both output quality and computational efficiency. |
| Researcher Affiliation | Academia | 1 University of Rome Unitelma Sapienza 2 University of Rome Sapienza 3 University of Mannheim |
| Pseudocode | Yes | Algorithm 1: CRUMBTRAIL |
| Open Source Code | No | The paper does not provide concrete access to source code for the methodology described in this paper. |
| Open Datasets | Yes | We use the gold-standard taxonomies and the taxonomy learning systems from Task 17 of the Sem Eval 2015 challenge2. 2http://alt.qcri.org/semeval2015/task17/ ... We apply the four pruning algorithms to the entire Wikipedia category graph... |
| Dataset Splits | No | The paper evaluates performance against gold-standard and reference graphs but does not specify explicit training/validation/test dataset splits for reproducibility in the context of model training or hyperparameter tuning. |
| Hardware Specification | No | The paper mentions issues with CLE running out of memory 'on a multi-core machine' but does not provide specific hardware details like CPU/GPU models, processor types, or memory amounts used for the experiments. |
| Software Dependencies | No | The paper does not provide specific ancillary software details with version numbers, such as library or solver names and their versions. |
| Experiment Setup | No | The paper describes the algorithms and evaluation methods but does not provide specific experimental setup details such as hyperparameter values or training configurations for the algorithms themselves. |