Notice: The reproducibility variables underlying each score are classified using an automated LLM-based pipeline, validated against a manually labeled dataset. LLM-based classification introduces uncertainty and potential bias; scores should be interpreted as estimates. Full accuracy metrics and methodology are described in [1].

Sentiment-Controllable Chinese Poetry Generation

Authors: Huimin Chen, Xiaoyuan Yi, Maosong Sun, Wenhao Li, Cheng Yang, Zhipeng Guo

IJCAI 2019 | Venue PDF | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Experimental results show our model can control the sentiment of not only a whole poem but also each line, and improve the poetry diversity against the state-of-the-art models without losing quality.
Researcher Affiliation Academia 1Department of Computer Science and Technology, Tsinghua University 2Institute for Artificial Intelligence, Tsinghua University 3Beijing National Research Center for Information Science and Technology, Tsinghua University 4State Key Lab on Intelligent Technology and Systems, Tsinghua University
Pseudocode No The paper describes its methods using mathematical formulations and textual explanations but does not include explicit pseudocode or algorithm blocks.
Open Source Code Yes Our source code and the Fine-grained Sentiment Poetry Corpus will be available at https://github.com/THUNLP-AIPoet.
Open Datasets Yes Due to the lack of off-the-shelf sentimental poetry corpus, we first build a fine-grained manually-labelled sentimental Chinese corpus. ... Our source code and the Fine-grained Sentiment Poetry Corpus will be available at https://github.com/THUNLP-AIPoet.
Dataset Splits Yes For unlabelled data, we randomly select 4,500 poems for validation and testing respectively and the rest for training. ... For labelled data, we use 500 poems for validation and testing respectively.
Hardware Specification No The paper does not provide specific hardware details such as GPU/CPU models, memory, or other computer specifications used for running the experiments.
Software Dependencies No The paper mentions 'Adam [Kingma and Ba, 2015] with mini-batches (batch size 64) is used for optimization.' but does not specify versions for any programming languages, libraries, or frameworks (e.g., Python, PyTorch, TensorFlow).
Experiment Setup Yes The dimensions of word embedding, sentiment embedding and latent variable are 256, 32, 128 respectively. The hidden state size is 512 for the encoder, decoder and content sequence; 64 for the sentiment sequence. Adam [Kingma and Ba, 2015] with mini-batches (batch size 64) is used for optimization. We also use dropout (keep ratio=0.75) to avoid overfitting. For testing, all models generate poems with beam search (beam size = 20).