Ease-of-Teaching and Language Structure from Emergent Communication

Authors: Fushan Li, Michael Bowling

NeurIPS 2019 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental We explore the connection between ease-of-teaching and the structure of the language through empirical experiments.
Researcher Affiliation Academia Fushan Li Department of Computing Science University of Alberta Edmonton, Canada fushan@ualberta.ca Michael Bowling Department of Computing Science University of Alberta Edmonton, Canada mbowling@ualberta.ca
Pseudocode No No structured pseudocode or algorithm blocks were found in the paper.
Open Source Code No The paper does not provide any explicit statement or link indicating that the source code for their methodology is publicly available.
Open Datasets No The paper describes the attributes used to create objects for the game (colors and shapes), which are generated within their experimental setup, but it does not refer to a pre-existing, publicly available dataset with concrete access information (link, DOI, or specific citation to the dataset itself).
Dataset Splits No The paper describes a game setup where objects are generated based on attributes, and agents are trained over a certain number of iterations. It does not define explicit train/validation/test dataset splits from a fixed dataset, but rather describes a continuous training and evaluation process involving dynamically generated game instances.
Hardware Specification No The paper does not specify the hardware used for running experiments, such as GPU or CPU models.
Software Dependencies No The paper mentions methods like Adam and REINFORCE but does not provide specific version numbers for any software dependencies, libraries, or frameworks used in the implementation.
Experiment Setup Yes For training, we use the Adam (Kingma and Ba, 2014) optimizer with learning rate 0.001 for both S and L. We use a batch size of 100 to compute policy gradients. We use λS = 0.1 and λL = 0.05 in all our experiments. The dimensionalities of the hidden states in both gS and gL are 100. In the reset regime, the listener is trained with the speaker for 6k (i.e., 6000) iterations and then we reinitialize the listener.