SpikeZIP-TF: Conversion is All You Need for Transformer-based SNN
Authors: Kang You, Zekai Xu, Chen Nie, Zhijie Deng, Qinghai Guo, Xiang Wang, Zhezhi He
ICML 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details
| Reproducibility Variable | Result | LLM Response |
|---|---|---|
| Research Type | Experimental | 4. Experiments, 4.2. Results Comparison, Table 5. Comparison on Image Net. |
| Researcher Affiliation | Collaboration | 1School of Electronic Information and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China 2Huawei Technologies, Shenzhen, China. |
| Pseudocode | No | The paper does not contain any clearly labeled pseudocode or algorithm blocks. |
| Open Source Code | Yes | The code is publicly available at: https://github.com/IntelligentComputing-Research-Group/SpikeZIP-TF |
| Open Datasets | Yes | Vision Benchmarks. Various vision datasets are adopted for evaluation. 1) static vision datasets, including CIFAR10/100 (Krizhevsky et al., 2009) and Image Net (Deng et al., 2009). 2) neuromorphic vision dataset: We evaluate Spike ZIP-TF on CIFAR10-DVS (Hongmin et al., 2017). |
| Dataset Splits | No | The paper lists the datasets used (e.g., Image Net, CIFAR10/100) but does not explicitly provide specific train/validation/test dataset splits (e.g., percentages or sample counts) within the main text or appendices for reproducibility. |
| Hardware Specification | No | The paper mentions 'Nvidia' GPUs and 'GPU Time(h)' in Table 7 regarding training cost, but does not specify exact GPU models (e.g., A100, V100), CPU models, or other specific hardware configurations used for running experiments. |
| Software Dependencies | No | The paper does not explicitly list specific software dependencies with version numbers (e.g., 'Python 3.8', 'PyTorch 1.9') used for the experiments. |
| Experiment Setup | Yes | The paper includes detailed hyperparameter configurations in Appendix A4, such as in Table A3 'The hyperparameter of end-to-end finetuning with Vi T-S Re LU on Image Net', listing 'optimizer Adam W', 'base learning rate 1e-4', 'batch size 64/256', 'training epochs 100', etc. |