Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Training Convolutional Neural Networks with the Forward-Forward algorithm (2312.14924v3)

Published 22 Dec 2023 in cs.CV and cs.AI

Abstract: The recent successes in analyzing images with deep neural networks are almost exclusively achieved with Convolutional Neural Networks (CNNs). The training of these CNNs, and in fact of all deep neural network architectures, uses the backpropagation algorithm where the output of the network is compared with the desired result and the difference is then used to tune the weights of the network towards the desired outcome. In a 2022 preprint, Geoffrey Hinton suggested an alternative way of training which passes the desired results together with the images at the input of the network. This so called Forward Forward (FF) algorithm has up to now only been used in fully connected networks. In this paper, we show how the FF paradigm can be extended to CNNs. Our FF-trained CNN, featuring a novel spatially-extended labeling technique, achieves a classification accuracy of 99.16% on the MNIST hand-written digits dataset. We show how different hyperparameters affect the performance of the proposed algorithm and compare the results with CNN trained with the standard backpropagation approach. Furthermore, we use Class Activation Maps to investigate which type of features are learnt by the FF algorithm.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (37)
  1. Layer normalization. arXiv.1607.06450, 2016.
  2. Gradients without backpropagation. arXiv:2202.08587, 2022.
  3. A study of forward-forward algorithm for self-supervised learning. arXiv:2309.11955, 2023.
  4. 2022 roadmap on neuromorphic computing and engineering. Neuromorphic Computing and Engineering, 2:022501, 2022. doi: 10.1088/2634-4386/ac4a83.
  5. µ-FF: On-Device Forward-Forward Training Algorithm for Microcontrollers. In 2023 IEEE International Conference on Smart Computing (SMARTCOMP), pages 49–56, 2023. doi: 10.1109/SMARTCOMP58114.2023.00024.
  6. Error-driven input modulation: Solving the credit assignment problem without a backward pass. In Proceedings of the 39th International Conference on Machine Learning, Proceedings of Machine Learning Research, pages 4937–4955, 2022.
  7. Efficient biologically plausible adversarial training. arXiv:2309.17348, 2023.
  8. Feed-forward optimization with delayed feedback for neural networks. arXiv:2304.13372, 2023.
  9. Extending the forward forward algorithm. arXiv:2307.04205, 2023.
  10. Histopathological image analysis: A review. IEEE Reviews in Biomedical Engineering, 2:147–171, 2009. doi: 10.1109/RBME.2009.2034865.
  11. Geoffrey Hinton. The forward-forward algorithm: Some preliminary investigations. arXiv:2212.13345, 2022.
  12. The Hugging Face Community. Anatomy of model’s memory, 2023. URL https://huggingface.co/docs/transformers/v4.36.0/model_memory_anatomy#anatomy-of-models-memory.
  13. Hebbian deep learning without feedback. arXiv:2209.11883, 2023.
  14. MNIST handwritten digit database, 2010. URL http://yann.lecun.com/exdb/mnist/.
  15. Symba: Symmetric backpropagation-free contrastive learning with forward-forward algorithm for optimizing convergence. arXiv:2303.08418, 2023.
  16. Backpropagation and the brain. Nature Reviews Neuroscience, 21(6):335–346, 2020. doi: 10.1038/s41583-020-0277-3.
  17. Sindy Loewe. The Forward-Forward Fully-Connected Network Implementation, 2023. URL https://github.com/loeweX/Forward-Forward.
  18. Layer collaboration in the forward-forward algorithm. arXiv:2305.12393, 2023.
  19. Fine-tuning language models with just forward passes. arXiv:2305.17333, 2023.
  20. Multiplexed gradient descent: Fast online training of modern datasets on hardware neural networks without backpropagation. APL Machine Learning, 1, 2023. doi: 10.1063/5.0157645.
  21. Arild Nøkland. Direct Feedback Alignment Provides Learning in Deep Neural Networks. In D. Lee, M. Sugiyama, U. Luxburg, I. Guyon, and R. Garnett, editors, Advances in Neural Information Processing Systems, volume 29. Curran Associates, Inc., 2016.
  22. Forward–forward training of an optical neural network. Opt. Lett., 48(20):5249–5252, 2023. doi: 10.1364/OL.496884.
  23. Alexander Ororbia. Contrastive-signal-dependent plasticity: Forward-forward learning of spiking neural systems. arXiv:2303.18187, 2023.
  24. The predictive forward-forward algorithm. arXiv:2301.01452, 2023.
  25. Forward-forward algorithm for hyperspectral image classification: A preliminary study. arXiv:2307.00231, 2023.
  26. Graph neural networks go forward-forward. arXiv:2302.05282, 2023.
  27. Deep Convolutional Neural Networks for Image Classification: A Comprehensive Review. Neural Computation, 29(9):2352–2449, 2017. doi: 10.1162/neco˙a˙00990.
  28. The forward-forward algorithm as a feature extractor for skin lesion classification: A preliminary study. arXiv:2307.00617, 2023.
  29. Learning representations by back-propagating errors. Nature, 323(6088):533–536, 1986.
  30. Equilibrium Propagation: Bridging the Gap between Energy-Based Models and Backpropagation. Frontiers in Computational Neuroscience, 11, 2017. doi: https://doi.org/10.3389/fncom.2017.00024.
  31. Artificial intelligence in histopathology: enhancing cancer research and clinical oncology. Nature Cancer, 3(9):1026–1038, 2022. doi: 10.1038/s43018-022-00436-4.
  32. Emergent representations in networks trained with the forward-forward algorithm. arXiv:2305.18353, 2023.
  33. Layer-wise feedback propagation. arXiv:2308.12053, 2023.
  34. Yukun Yang. A theory for the sparsity emerged in the forward forward algorithm. arXiv:2311.05667, 2023.
  35. The cascaded forward algorithm for neural network training. arXiv:2303.09728, 2023a.
  36. Tensor-compressed back-propagation-free training for (physics-informed) neural networks. arXiv:2308.09858, 2023b.
  37. Hongchao Zhou. Activation learning by local competitions. arXiv:2209.13400, 2022.
Citations (5)

Summary

We haven't generated a summary for this paper yet.

X Twitter Logo Streamline Icon: https://streamlinehq.com