Convolutional Channel-Wise Competitive Learning for the Forward-Forward Algorithm

Authors: Andreas Papachristodoulou, Christos Kyrkou, Stelios Timotheou, Theocharis Theocharides

AAAI 2024 | Conference PDF | Archive PDF | Plain Text | LLM Run Details

Reproducibility Variable Result LLM Response
Research Type Experimental Our method outperforms recent FF-based models on image classification tasks, achieving testing errors of 0.58%, 7.69%, 21.89%, and 48.77% on MNIST, Fashion MNIST, CIFAR-10 and CIFAR-100 respectively. Our approach is benchmarked against FF-based and non Backpropagation methods reported in the literature on image classification tasks.
Researcher Affiliation Academia KIOS Research and Innovation Center of Excellence, Department of Electrical and Computer Engineering, University of Cyprus {papachristodoulou.andreas, kyrkou.christos, stimo, ttheocharides}@ucy.ac.cy
Pseudocode Yes Algorithm 1: Interleaved Layer Training (ILT) Strategy
Open Source Code Yes Our source code and supplementary material are available at https://github.com/andreaspapac/CwComp.
Open Datasets Yes The performance of the various models was evaluated on three benchmark datasets commonly used in literature, namely MNIST (Le Cun et al. 1998), Fashion MNIST (Xiao, Rasul, and Vollgraf 2017), and CIFAR-10 (Krizhevsky and Hinton 2009). Our method outperforms all FF-based models on basic image classification tasks, demonstrating quicker convergence rates and significantly better performance, achieving testing errors of 0.58%, 7.69%, 21.89%, and 48.77% on MNIST (Le Cun et al. 1998), Fashion-MNIST (Xiao, Rasul, and Vollgraf 2017), CIFAR10, and CIFAR-100 (Krizhevsky and Hinton 2009) respectively.
Dataset Splits No The paper does not explicitly provide details about training/validation/test dataset splits (e.g., percentages, sample counts for each split) nor does it specify how a validation set was used for the standard benchmark datasets.
Hardware Specification Yes Our models were trained on an NVIDIA-RTX 4080 GPU and the i9-13900K Intel CPU for 20 epochs on MNIST, 50 epochs on Fashion-MNIST, and CIFAR-10.
Software Dependencies No The paper mentions 'torchinfo (Harris 2016)' as an open-source toolbox for calculating parameters and operations, but it does not specify a version number for torchinfo or any other software dependencies crucial for reproducibility.
Experiment Setup Yes In the experiments, the learning rate used for both the convolutional and dense layers was set to 0.01, and the batch size to 128. For comprehensive details on the model configurations, refer to the supplementary material.