Data Augmented Graph Neural Networks for Personality Detection

Authors: Yangfu Zhu, Yue Xia, Meiling Li, Tingting Zhang, Bin Wu

AAAI 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Extensive experiments on three real-world datasets, Youtube, PAN2015, and My Personality demonstrate the effectiveness of our Semi-Per GCN in personality detection, especially in scenarios with limited labeled users.
Researcher Affiliation Academia Beijing University of Posts and Telecommunications, Beijing, China zhuyangfu,meilinglee,zhangtingting,wubin@bupt.edu.cn, 1216918224@qq.com
Pseudocode No The paper describes the model architecture and equations (e.g., Xk+1 = σ(AXkWk), L = Ld + λLc) but does not include any explicit pseudocode or algorithm blocks.
Open Source Code No The paper does not contain any statement about releasing source code or providing a link to a code repository for the described methodology.
Open Datasets Yes We conduct experiments on the Youtube Personality (Biel et al. 2013), PAN2015 (Rangel Pardo et al. 2015), and Mypersonality datasets (Celli et al. 2013; Xue et al. 2018) with Big Five taxonomy.
Dataset Splits Yes All the hyperparameters are tuned over the validation set to obtain the optimized results.
Hardware Specification Yes We use Pytorch to implement all the deep learning models on our three 2080Ti GPU cards.
Software Dependencies No The paper mentions using 'Pytorch' and 'bert-base-cased' but does not specify version numbers for these software dependencies or any other libraries.
Experiment Setup Yes Empirically, we use a batch size of 16,16, and 64 for the labeled data and a batch size of 32, 32, and 112 for the unlabeled data in Youtube, PAN2015, and My Personality datasets respectively. Adam is utilized as the optimizer and the learning rate of our model is set to 0.0001, 0.0003, and 0.0003 in PAN2015, Youtube, and My Personality datasets respectively. The pre-trained language models BERT are employed to initialize the word node embeddings by the bert-base-cased (Devlin et al. 2018), and the dimensions of word nodes, LIWC nodes, and user nodes are set to 200.