Variational Template Machine for Data-to-Text Generation
Authors: Rong Ye, Wenxian Shi, Hao Zhou, Zhongyu Wei, Lei Li
ICLR 2020 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | Experiments on datasets from a variety of different domains show that VTM is able to generate more diversely while keeping a good fluency and quality. |
| Researcher Affiliation | Collaboration | Fudan University {rye18,zywei}@fudan.edu.cn Byte Dance AI Lab {shiwenxian,zhouhao.nlp,lileilab}@.bytedance.com |
| Pseudocode | Yes | Algorithm 1 Training procedure |
| Open Source Code | No | The paper does not provide an explicit statement about releasing source code for the VTM model, nor does it include a link to a code repository. |
| Open Datasets | Yes | We perform the experiment on SPNLG (Reed et al., 2018)3 and WIKI (Lebret et al., 2016; Wang et al., 2018b). ... https://nlds.soe.ucsc.edu/sentence-planning-NLG ... https://eaglew.github.io/patents/ |
| Dataset Splits | Yes | The statistics for the number of table-text pairs and raw texts in the training, validation and test sets are shown in Table 2. ... Table 2: Dataset statistics in our experiments. ... Valid ... SPNLG 20, 495 ... WIKI 72, 831 |
| Hardware Specification | Yes | We train and test the models on a single Tesla V100 GPU. |
| Software Dependencies | No | The paper mentions using 'Adam optimizer' but does not specify versions for programming languages or other software libraries (e.g., Python, PyTorch/TensorFlow) used for implementation. |
| Experiment Setup | Yes | Word embeddings are randomly initialized with 300-dimension. During training, we use Adam optimizer (Kingma & Ba, 2015) with the initial learning rate as 0.001. Details on hyperparameters are listed in Appendix D. ... For the model trained on WIKI dataset, the the dimension of latent template variable is set as 100, and the dimension of latent content variable is set as 200. ... For the hyperparameters of total loss Ltot, we set λMI = 0.5, λpt = 1.0 and λpc = 0.5. |