Show Me How To Revise: Improving Lexically Constrained Sentence Generation with XLNet
Authors: Xingwei He, Victor O.K. Li12989-12997
AAAI 2021 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | Experimental results have demonstrated that our proposed model performs much better than the previous work in terms of sentence fluency and diversity. |
| Researcher Affiliation | Academia | Xingwei He, Victor O.K. Li Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, China hexingwei15@gmail.com, vli@eee.hku.hk |
| Pseudocode | Yes | Algorithm 1 Constrained Sentence Generation with XLNet |
| Open Source Code | Yes | Our code, pre-trained models and Appendix are available at https://github.com/NLPCode/MCMCXLNet. |
| Open Datasets | Yes | We used One-Billion-Word corpus1 to construct the synthetic dataset. 1http://www.statmt.org/lm-benchmark/ |
| Dataset Splits | Yes | We selected 6M, 0.3M and 1K sentences from One-Billion-Word corpus as the training, validation, and test sets, respectively. |
| Hardware Specification | No | The paper does not provide specific hardware details (e.g., GPU/CPU models, memory) used for running experiments. |
| Software Dependencies | No | The paper mentions 'Hugging Face' and 'GPT-2 small (117M)' but does not provide specific version numbers for these or other software dependencies. |
| Experiment Setup | Yes | The experiment setups for language models and the classifier are shown in the Appendix. We set N to 20. We set K to 50. All MCMC-based models were run for 200 steps. We ran 20 iterations for Bayesian MCMC, TIGS, L-MCMC, and L-MCMC-C. |