Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Graph Transformer for Graph-to-Sequence Learning

Authors: Deng Cai, Wai Lam7464-7471

AAAI 2020 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experiments on the applications of text generation from Abstract Meaning Representation (AMR) and syntax-based neural machine translation show the superiority of our proposed model.
Researcher Affiliation Academia Deng Cai, Wai Lam The Chinese University of Hong Kong EMAIL, EMAIL
Pseudocode No The paper does not contain structured pseudocode or algorithm blocks.
Open Source Code Yes Code available at https://github.com/jcyk/gtos.
Open Datasets Yes For this AMR-to-text generation task, we use two benchmarks, namely the LDC2015E86 dataset and the LDC2017T10 dataset. Both the English-German and the English-Czech datasets from the WMT16 translation task. http://www.statmt.org/wmt16/translation-task.html
Dataset Splits Yes Table 1: Data statistics of all four datasets. #train/dev/test indicates the number of instances in each set...
Hardware Specification No The paper does not provide specific hardware details such as GPU/CPU models, processor types, or memory amounts used for running its experiments.
Software Dependencies No The paper mentions software components like 'Adam optimizer' and implies 'Python' for the code, but does not provide specific version numbers for any libraries, frameworks, or programming languages used (e.g., PyTorch 1.x, Python 3.x).
Experiment Setup Yes Table 2: Hyper-parameters settings. (Table includes number of filters, width of filters, embedding sizes, number of heads, hidden state sizes, feed-forward hidden size). Also mentions beam size of 8, dropout rate of 0.2, UNK token rate of 0.33, Adam optimizer with β1 = 0.9 and beta2 = 0.999, and the same learning rate schedule of Vaswani et al.(2017).