More the Merrier: Towards Multi-Emotion and Intensity Controllable Response Generation
Authors: Mauajama Firdaus, Hardik Chauhan, Asif Ekbal, Pushpak Bhattacharyya12821-12829
AAAI 2021 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | The detailed evaluation shows that our proposed approach attains superior performance compared to the baseline models. Experiments Implementation details For all the models including baselines, the batch size is set to 32. The utterance encoder is a bidirectional GRU with 600 hidden units in each direction. |
| Researcher Affiliation | Academia | Mauajama Firdaus, Hardik Chauhan, Asif Ekbal and Pushpak Bhattacharyya Department of Computer Science and Engineering Indian Institute of Technology Patna, India (mauajama.pcs16,hardik,asif,pb)@iitp.ac.in |
| Pseudocode | No | No structured pseudocode or algorithm blocks were found. |
| Open Source Code | No | No explicit statement or link providing access to the source code for the methodology was found. |
| Open Datasets | No | Due to the unavailability of multi-emotion intensity labeled data for our proposed task, we create a large-scale dialogue dataset, MEIMD from 8 English TV series having 34k conversations that have been labeled with multiple emotions and their intensities. We create a large-scale Multiple Emotion and Intensity aware Multi-party Dialogue (MEIMD) dataset. |
| Dataset Splits | No | The paper does not explicitly provide training/test/validation dataset splits needed to reproduce the experiment. |
| Hardware Specification | No | No specific hardware details (GPU/CPU models, processor types, memory amounts, or detailed computer specifications) used for running experiments were found. |
| Software Dependencies | No | The paper mentions optimizers and embeddings (AMS-Grad, Glove) but does not provide specific version numbers for software dependencies or libraries. |
| Experiment Setup | Yes | For all the models including baselines, the batch size is set to 32. The utterance encoder is a bidirectional GRU with 600 hidden units in each direction. The context encoder and decoder are both GRUs with 600 hidden units. All the model parameters are randomly initialised using a Gaussian distribution with Xavier scheme (Glorot and Bengio 2010). We use 0.45 as dropout rate and perform gradient clipping when gradient norm is over 3. |