Ease-of-Teaching and Language Structure from Emergent Communication
Authors: Fushan Li, Michael Bowling
NeurIPS 2019 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | We explore the connection between ease-of-teaching and the structure of the language through empirical experiments. |
| Researcher Affiliation | Academia | Fushan Li Department of Computing Science University of Alberta Edmonton, Canada fushan@ualberta.ca Michael Bowling Department of Computing Science University of Alberta Edmonton, Canada mbowling@ualberta.ca |
| Pseudocode | No | No structured pseudocode or algorithm blocks were found in the paper. |
| Open Source Code | No | The paper does not provide any explicit statement or link indicating that the source code for their methodology is publicly available. |
| Open Datasets | No | The paper describes the attributes used to create objects for the game (colors and shapes), which are generated within their experimental setup, but it does not refer to a pre-existing, publicly available dataset with concrete access information (link, DOI, or specific citation to the dataset itself). |
| Dataset Splits | No | The paper describes a game setup where objects are generated based on attributes, and agents are trained over a certain number of iterations. It does not define explicit train/validation/test dataset splits from a fixed dataset, but rather describes a continuous training and evaluation process involving dynamically generated game instances. |
| Hardware Specification | No | The paper does not specify the hardware used for running experiments, such as GPU or CPU models. |
| Software Dependencies | No | The paper mentions methods like Adam and REINFORCE but does not provide specific version numbers for any software dependencies, libraries, or frameworks used in the implementation. |
| Experiment Setup | Yes | For training, we use the Adam (Kingma and Ba, 2014) optimizer with learning rate 0.001 for both S and L. We use a batch size of 100 to compute policy gradients. We use λS = 0.1 and λL = 0.05 in all our experiments. The dimensionalities of the hidden states in both gS and gL are 100. In the reset regime, the listener is trained with the speaker for 6k (i.e., 6000) iterations and then we reinitialize the listener. |