Cross-Lingual Knowledge Validation Based Taxonomy Derivation from Heterogeneous Online Wikis

Authors: Zhigang Wang, Juanzi Li, Shuangjie Li, Mingyang Li, Jie Tang, Kuo Zhang, Kun Zhang

AAAI 2014 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental The proposed approach successfully overcome the above issues, and experiments show that our approach significantly outperforms the designed state-of-the-art comparison methods.
Researcher Affiliation Collaboration Zhigang Wang Juanzi Li Shuangjie Li Mingyang Li Jie Tang Kuo Zhang Kun Zhang Department of Computer Science and Technology, Tsinghua University, Beijing, China {wzhigang, ljz, lsj, lmy, tangjie}@keg.cs.tsinghua.edu.cn Sogou Incorporation, Beijing, China {zhangkuo, zhangkun}@sogou-inc.com
Pseudocode No The paper describes the learning process with mathematical equations and steps, but it does not include a formally labeled 'Pseudocode' or 'Algorithm' block.
Open Source Code No The paper states: 'The data sets are available at http://xlore.org/publications.action.' This refers to data, not source code for the methodology.
Open Datasets Yes The data sets are available at http://xlore.org/publications.action.
Dataset Splits Yes To demonstrate the better generalization ability of DAB model, we conduct 2-fold crossvalidation on the labeled dataset. Besides, in each iteration, we separate the cross-lingual validated results Vt into 2 fold and add one of them into the testing dataset, and use the other part to expand the pool Pt+1.
Hardware Specification No The paper does not provide specific hardware details (e.g., CPU, GPU models, memory) used for running the experiments.
Software Dependencies No The paper mentions 'Weka (Hall et al. 2009)' and 'Stanford Parser (Green et al. 2011)' but does not provide specific version numbers for these or other software dependencies.
Experiment Setup Yes Both the comparison methods and DAB model use the default settings of Decision Tree in Weka (Hall et al. 2009). We use the Stanford Parser (Green et al. 2011) for head word extraction. The Ada Boost and DAB methods run 20 iterations. ... the threshold θ is experimentally set as 0.93. ... The parameter δ is used to limit the update speed, where in each iteration no more than δ m examples are replaced from At. δ is experimentally set as 0.2.