Training Agent for First-Person Shooter Game with Actor-Critic Curriculum Learning
Authors: Yuxin Wu, Yuandong Tian
ICLR 2017 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | In this section, we show the training procedure (Sec. 5.1), evaluate our AIs with ablation analysis (Sec. 5.2) and Vi ZDoom AI Competition (Sec. 5.3). |
| Researcher Affiliation | Collaboration | Yuxin Wu Carnegie Mellon University ppwwyyxx@gmail.com Yuandong Tian Facebook AI Research yuandong@fb.com |
| Pseudocode | No | The paper describes the model and training procedure in detail, but does not provide structured pseudocode or an algorithm block. |
| Open Source Code | No | The paper mentions "tensorpack" which has a GitHub link (5https://github.com/ppwwyyxx/tensorpack), but this is a framework used by the authors, not the specific code implementation for *this* paper's methodology. There is no explicit statement or link for the paper's own source code. |
| Open Datasets | No | The paper uses the Vi ZDoom platform and describes custom scenarios like "Flat Map" and "CIGTrack1", but does not provide any public access information (link, citation, repository) for the specific datasets or scenarios used in their experiments. |
| Dataset Splits | No | The paper mentions using "100 episodes" or "300 episodes" for evaluation but does not specify a formal train/validation/test split for a dataset. |
| Hardware Specification | Yes | The training procedure runs on Intel Xeon CPU E5-2680v2 at 2. 80GHz, and 2 Titan X GPUs. |
| Software Dependencies | No | Our training procedure is implemented with Tensor Flow [Abadi et al. (2016)] and tensorpack5. (No version numbers provided for TensorFlow or tensorpack.) |
| Experiment Setup | Yes | We use Adam [Kingma & Ba (2014)] with ǫ = 10 3 for training. Batch size is 128, discount factor γ = 0.99, learning rate α = 10 4 and the policy learning rate β = 0.08α. |