Submitted:
05 March 2025
Posted:
20 March 2025
You are already at the latest version
Abstract
Keywords:
1. Introduction
2. Fundamentals of Large-scale Spiking Neural Networks
2.1. Mathematical Model of Spiking Neurons
- is the membrane potential at time t,
- is the resting membrane potential,
- R is the membrane resistance,
- is the input current,
- is the membrane time constant, with C being the membrane capacitance [19].
2.2. Spike Timing and Synaptic Plasticity
- is the change in synaptic weight,
- is the spike timing difference,
- and are learning rate parameters,
- and are time constants governing weight updates.
2.3. Comparison of Spiking Neuron Models
2.4. Challenges in Scaling SNNs
- Training Stability: The non-differentiability of spikes requires surrogate gradient techniques for efficient training.
- Hardware Constraints: Neuromorphic processors need optimized spike communication mechanisms.
- Energy Efficiency: Reducing redundant spike activity is crucial for deploying SNNs at scale.
3. Training Large-scale Spiking Neural Networks
3.1. Backpropagation in SNNs: Surrogate Gradient Methods
3.2. Spike-Timing-Dependent Plasticity (STDP)
- is the weight update,
- is the time difference between post- and pre-synaptic spikes,
- and are learning rates for potentiation and depression,
- and are decay time constants.
3.3. Hybrid Learning Approaches
3.4. Comparison of Training Methods
3.5. Challenges in Training Large-Scale SNNs
- Gradient Stability: Surrogate gradients can lead to vanishing or exploding gradients, requiring careful tuning.
- Hardware Constraints: Implementing STDP or hybrid approaches efficiently on neuromorphic hardware remains difficult [30].
4. Neuromorphic Hardware for Large-scale SNNs
4.1. Neuromorphic Computing Paradigm
- Event-driven computation: Instead of continuous activations, computations occur only when spikes are generated.
- In-memory processing: Memory and computation are tightly coupled, minimizing data movement [37].
- Parallelism: Massive parallelism enables real-time processing with low power consumption [38].
4.2. Comparison of Neuromorphic Architectures
4.3. Mathematical Model for Neuromorphic Efficiency
- is the number of processed spikes per second,
- is the computation time per spike [42].
4.4. Challenges in Large-scale Neuromorphic Computing
- Memory Bottlenecks: Storing and accessing large synaptic weight matrices is costly in terms of power and latency.
- Communication Overhead: Routing spike events across large-scale networks requires optimized event-driven architectures [44].
- Hardware-software Co-design: Developing algorithms tailored for neuromorphic platforms is necessary for achieving optimal performance [45].
4.5. Future Directions in Neuromorphic Hardware
- 3D neuromorphic chips: Stacking memory and processing units vertically can reduce data movement and improve efficiency [47].
- Hybrid analog-digital designs: Combining analog computation with digital precision can enhance scalability [48].
- Brain-inspired learning mechanisms: Implementing efficient on-chip learning algorithms can enable real-time adaptive systems.
5. Applications of Large-Scale Spiking Neural Networks
5.1. Neuroscientific Modeling and Brain Simulation
- is the total number of synapses,
- is the number of neurons.
5.2. Event-Driven Perception in Robotics
- L is the latency per event,
- E is the energy consumption per event,
- is the spiking throughput [54].
5.3. Neuromorphic Computer Vision
5.4. Edge AI and Low-Power IoT Devices
- Always-on wake-up detectors, where SNNs process low-power auditory or visual signals [61].
- Energy-efficient speech recognition, leveraging spike-based audio processing.
- is the power consumption per inference,
- is the total energy per inference,
- is the number of spikes generated per inference [62].
5.5. Future Prospects and Open Challenges
- Scalability in real-world applications: Deploying SNNs beyond simulations requires further hardware and algorithmic optimizations [64].
- Algorithm-hardware co-design: Efficient neuromorphic processing demands synergy between SNN models and specialized neuromorphic hardware [65].
- Data-driven learning in SNNs: While SNNs excel in unsupervised learning (e.g., STDP), supervised and reinforcement learning methods remain underdeveloped [66].
6. Future Research Directions and Open Challenges
6.1. Scalable Training Algorithms for Large-Scale SNNs
- Surrogate Gradient Methods: Approximate the derivative of the spiking function to enable backpropagation [71].
- Spike-Timing-Dependent Plasticity (STDP): A biologically inspired unsupervised learning rule [72].
- Hybrid ANN-to-SNN Conversion: Train deep artificial neural networks (ANNs) and convert them into SNNs [73].
- is the true loss function [32].
- is the surrogate loss function.
- is the membrane potential of the neuron.
6.2. Optimizing SNN Energy Efficiency
- is the number of active neurons per inference.
- is the total number of neurons.
6.3. Scalability of Neuromorphic Hardware
6.4. Bridging the Gap Between Theory and Real-world Applications
6.5. Conclusion
7. Conclusion
7.1. Key Takeaways
- Hybrid learning approaches show promise. ANN-to-SNN conversion and surrogate gradient methods have improved training performance, but fully unlocking the potential of SNNs requires novel biologically plausible learning rules [96].
7.2. Future Outlook
- Enhancing training algorithms to improve convergence and accuracy in deep SNN architectures [99].
- Developing energy-efficient neuromorphic hardware capable of scaling up SNN computations [100].
- Bridging the gap between SNN theory and practice by integrating them into mainstream AI applications [101].
References
- Maass, W. Lower bounds for the computational power of networks of spiking neurons. Neural Computation 1996, 8, 1–40. [Google Scholar] [CrossRef]
- Brette, R.; Gerstner, W. Adaptive exponential integrate-and-fire model as an effective description of neuronal activity. Journal of Neurophysiology 2005, 94, 3637–3642. [Google Scholar] [CrossRef] [PubMed]
- Van Rullen, R.; Thorpe, S.J. Rate coding versus temporal order coding: what the retinal ganglion cells tell the visual cortex. Neural Computation 2001, 13, 1255–1283. [Google Scholar] [CrossRef]
- Liu, C.; Chen, P.; Zhuang, f.B.; Shen, C.; Zhang, B.; Ding, W. SA-BNN: State-aware binary neural network. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2021, Vol. 35, pp. 2091–2099.
- Wang, P.; He, X.; Li, G.; Zhao, T.; Cheng, J. Sparsity-inducing binarized neural networks. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2020, Vol. 34, pp. 12192–12199.
- Ng, K.T.; Boussaid, F.; Bermak, A. A CMOS single-chip gas recognition circuit for metal oxide gas sensor arrays. IEEE Transactions on Circuits and Systems 2011, 58, 1569–1580. [Google Scholar] [CrossRef]
- Google. Cloud TPU.
- Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. International Journal of Computer Vision 2015, 115, 211–252. [Google Scholar] [CrossRef]
- Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Proceedings of the International Conference on Artificial Intelligence and Statistics. JMLR Workshop and Conference Proceedings, 2010, pp. 249–256.
- iniLabs. Dynamic Audio Sensor.
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2016, pp. 770–778.
- Zhu, R.J.; Zhao, Q.; Zhang, T.; Deng, H.; Duan, Y.; Zhang, M.; Deng, L.J. TCJA-SNN: Temporal-channel joint attention for spiking neural networks. arXiv 2022, arXiv:2206.10177. [Google Scholar] [CrossRef]
- Yan, Z.; Zhou, J.; Wong, W.F. Near Lossless Transfer Learning for Spiking Neural Networks. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2021, Vol. 35, pp. 10577–10584.
- Guo, L.; Gao, Z.; Qu, J.; Zheng, S.; Jiang, R.; Lu, Y.; Qiao, H. Transformer-based Spiking Neural Networks for Multimodal Audio-Visual Classification. IEEE Transactions on Cognitive and Developmental Systems 2023, 16, 1077–1086. [Google Scholar] [CrossRef]
- Li, Y.; Deng, S.; Dong, X.; Gong, R.; Gu, S. A free lunch from ANN: Towards efficient, accurate spiking neural networks calibration. In Proceedings of the Proceedings of the International Conference on Machine Learning, 2021, pp. 6316–6325.
- Bal, M.; Sengupta, A. SpikingBERT: Distilling bert to train spiking language models using implicit differentiation. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2024, Vol. 38, pp. 10998–11006.
- Hunsberger, E.; Eliasmith, C. Spiking deep networks with LIF neurons. arXiv 2015, arXiv:1510.08829. [Google Scholar]
- Chen, Q.; Sun, C.; Gao, C.; Liu, S.C. Epilepsy Seizure Detection and Prediction using an Approximate Spiking Convolutional Transformer. arXiv 2024, arXiv:2402.09424. [Google Scholar]
- Wang, Z.; Fang, Y.; Cao, J.; Zhang, Q.; Wang, Z.; Xu, R. Masked spiking transformer. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 1761–1771.
- Wang, S.; Cheng, T.H.; Lim, M.H. LTMD: Learning Improvement of Spiking Neural Networks with Learnable Thresholding Neurons and Moderate Dropout. Advances in Neural Information Processing Systems 2022, 35, 28350–28362. [Google Scholar]
- Liu, Q.; Furber, S. Noisy softplus: A biology inspired activation function. In Proceedings of the Proceeddings of the International Conference on Neural Information Processing. Springer, 2016, pp. 405–412.
- Zhu, R.J.; Zhao, Q.; Eshraghian, J.K. SpikeGPT: Generative pre-trained language model with spiking neural networks. arXiv 2023, arXiv:2302.13939. [Google Scholar]
- Lian, S.; Shen, J.; Liu, Q.; Wang, Z.; Yan, R.; Tang, H. Learnable surrogate gradient for direct training spiking neural networks. In Proceedings of the Proceedings of the International Joint Conference on Artificial Intelligence, 2023, pp. 3002–3010.
- Hodgkin, A.L.; Huxley, A.F. A quantitative description of membrane current and its application to conduction and excitation in nerve. The Journal of Physiology 1952, 117, 500. [Google Scholar] [CrossRef] [PubMed]
- Rueckauer, B.; Lungu, I.A.; Hu, Y.; Pfeiffer, M.; Liu, S.C. Conversion of continuous-valued deep networks to efficient event-driven networks for image classification. Frontiers in Neuroscience 2017, 11, 682. [Google Scholar] [CrossRef] [PubMed]
- Rastegari, M.; Ordonez, V.; Redmon, J.; Farhadi, A. Xnor-net: Imagenet classification using binary convolutional neural networks. In Proceedings of the Proceedings of the European Conference on Computer Vision. Springer, 2016, pp. 525–542.
- Lin, D.; Talathi, S.; Annapureddy, S. Fixed point quantization of deep convolutional networks. In Proceedings of the Proceedings of the International Conference on Machine Learning, 2016, pp. 2849–2858.
- Izhikevich, E.M.; Desai, N.S.; Walcott, E.C.; Hoppensteadt, F.C. Bursts as a unit of neural information: selective communication via resonance. Trends in Neurosciences 2003, 26, 161–167. [Google Scholar] [CrossRef]
- Brandli, C.; Berner, R.; Yang, M.; Liu, S.C.; Delbruck, T. A 240× 180 130 dB 3 μs latency global shutter spatiotemporal vision sensor. IEEE Journal of Solid-State Circuits 2014, 49, 2333–2341. [Google Scholar] [CrossRef]
- Li, Y.; Deng, S.; Dong, X.; Gu, S. Error-Aware Conversion from ANN to SNN via Post-training Parameter Calibration. International Journal of Computer Vision 2024, pp. 1–24.
- Krizhevsky, A.; Hinton, G. Learning multiple layers of features from tiny images. Master’s thesis, Department of Computer Science, University of Toronto 2009.
- Yan, J.; Liu, Q.; Zhang, M.; Feng, L.; Ma, D.; Li, H.; Pan, G. Efficient spiking neural network design via neural architecture search. Neural Networks 2024, p. 106172.
- Zniyed, Y.; Nguyen, T.P.; et al. Efficient tensor decomposition-based filter pruning. Neural Networks 2024, 178, 106393. [Google Scholar]
- Gerstner, W.; Kistler, W.M. Spiking neuron models: Single neurons, populations, plasticity; Cambridge University Press, 2002.
- Liu, Z.; Shen, Z.; Savvides, M.; Cheng, K.T. Reactnet: Towards precise binary neural network with generalized activation functions. In Proceedings of the Proceedings of the European Conference on Computer Vision. Springer, 2020, pp. 143–159.
- iniLabs. DVS 128.
- Ning, Q.; Hesham, M.; Fabio, S.; Dora, S. A reconfigurable on-line learning spiking neuromorphic processor comprising 256 neurons and 128K synapses. Frontiers in Neuroscience 2015, 9, 141. [Google Scholar]
- Lian, S.; Shen, J.; Wang, Z.; Tang, H. IM-LIF: Improved Neuronal Dynamics With Attention Mechanism for Direct Training Deep Spiking Neural Network. IEEE Transactions on Emerging Topics in Computational Intelligence 2024. [Google Scholar] [CrossRef]
- Zhang, J.; Shen, J.; Wang, Z.; Guo, Q.; Yan, R.; Pan, G.; Tang, H. SpikingMiniLM: Energy-efficient Spiking Transformer for Natural Language Understanding. Science China Information Sciences 2024. [Google Scholar] [CrossRef]
- Jiang, Y.; Hu, K.; Zhang, T.; Gao, H.; Liu, Y.; Fang, Y.; Chen, F. Spatio-Temporal Approximation: A Training-Free SNN Conversion for Transformers. In Proceedings of the Proceedings of the International Conference on Learning Representations; 2024. [Google Scholar]
- Vedaldi, A.; Lenc, K. MatConvNet: Convolutional Neural Networks for Matlab. In Proceedings of the Proceedings of the ACM International Conference on Multimedia, 2015, pp. 689–692.
- Dampfhoffer, M.; Mesquida, T.; Valentian, A.; Anghel, L. Backpropagation-based learning techniques for deep spiking neural networks: A survey. IEEE Transactions on Neural Networks and Learning Systems 2023.
- He, K.; Zhang, X.; Ren, S.; Sun, J. Identity mappings in deep residual networks. In Proceedings of the Proceedings of the European Conference on Computer Vision. Springer, 2016, pp. 630–645.
- Scellier, B.; Bengio, Y. Equilibrium propagation: Bridging the gap between energy-based models and backpropagation. Frontiers in Computational Neuroscience 2017, 11, 24. [Google Scholar] [CrossRef]
- Mahowald, M.A. Silicon retina with adaptive photoreceptors. In Proceedings of the Proceedings of the SPIE/SPSE Symposium on Electronic Science and Technology: from Neurons to Chips, 1991, Vol. 1473, pp. 52–58.
- Delbruck, T.; Mead, C.A. Adaptive photoreceptor with wide dynamic range. In Proceedings of the IEEE International Symposium on Circuits and Systems. IEEE, 1994, Vol. 4, pp. 339–342.
- Qin, H.; Gong, R.; Liu, X.; Shen, M.; Wei, Z.; Yu, F.; Song, J. Forward and backward information retention for accurate binary neural networks. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2020, pp. 2250–2259.
- Zou, S.; Mu, Y.; Zuo, X.; Wang, S.; Cheng, L. Event-based human pose tracking by spiking spatiotemporal transformer. arXiv 2023, arXiv:2303.09681. [Google Scholar]
- Davies, M.; Srinivasa, N.; Lin, T.H.; Chinya, G.; Cao, Y.; Choday, S.H.; Dimou, G.; Joshi, P.; Imam, N.; Jain, S.; et al. Loihi: A neuromorphic manycore processor with on-chip learning. IEEE Micro 2018, 38, 82–99. [Google Scholar] [CrossRef]
- Xu, B.; Geng, H.; Yin, Y.; Li, P. DISTA: Denoising Spiking Transformer with intrinsic plasticity and spatiotemporal attention. arXiv 2023, arXiv:2311.09376. [Google Scholar]
- O’Connor, P.; Gavves, E.; Welling, M. Training a spiking neural network with equilibrium propagation. In Proceedings of the Proceedings of the International Conference on Artificial Intelligence and Statistics, 2019, pp. 1516–1523.
- Hinton, G. The forward-forward algorithm: Some preliminary investigations. arXiv 2022, arXiv:2212.13345. [Google Scholar]
- Rueckauer, B.; Liu, S.C. Conversion of analog to spiking neural networks using sparse temporal coding. In Proceedings of the Proceedings of the IEEE International Symposium on Circuits and Systems. IEEE, 2018, pp. 1–5.
- Lee, J.H.; Delbruck, T.; Pfeiffer, M. Training deep spiking neural networks using backpropagation. Frontiers in Neuroscience 2016, 10, 508. [Google Scholar] [CrossRef]
- Fang, W.; Yu, Z.; Chen, Y.; Masquelier, T.; Huang, T.; Tian, Y. Incorporating learnable membrane time constant to enhance learning of spiking neural networks. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 2661–2671.
- Thorpe, S.; Delorme, A.; Van Rullen, R. Spike-based strategies for rapid processing. Neural Networks 2001, 14, 715–725. [Google Scholar] [CrossRef]
- Park, E.; Ahn, J.; Yoo, S. Weighted-entropy-based quantization for deep neural networks. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2017, pp. 5456–5464.
- Benjamin, B.V.; Gao, P.; McQuinn, E.; Choudhary, S.; Chandrasekaran, A.R.; Bussat, J.M.; Alvarez-Icaza, R.; Arthur, J.V.; Merolla, P.A.; Boahen, K. Neurogrid: A mixed-analog-digital multichip system for large-scale neural simulations. Proceedings of the IEEE 2014, 102, 699–716. [Google Scholar] [CrossRef]
- Shi, X.; Hao, Z.; Yu, Z. SpikingResformer: Bridging ResNet and Vision Transformer in Spiking Neural Networks. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2024, pp. 5610–5619.
- Zniyed, Y.; Nguyen, T.P.; et al. Enhanced network compression through tensor decompositions and pruning. IEEE Transactions on Neural Networks and Learning Systems 2024. [Google Scholar]
- Liu, M.; Tang, J.; Li, H.; Qi, J.; Li, S.; Wang, K.; Wang, Y.; Chen, H. Spiking-PhysFormer: Camera-Based Remote Photoplethysmography with Parallel Spike-driven Transformer. arXiv 2024, arXiv:2402.04798. [Google Scholar] [CrossRef]
- Collobert, R.; Weston, J.; Bottou, L.; Karlen, M.; Kavukcuoglu, K.; Kuksa, P. Natural language processing (almost) from scratch. Journal of Machine Learning Sesearch 2011, 12, 2493–2537. [Google Scholar]
- Intel Corporation. Intel Xeon Platinum 9282 Processor.
- Sengupta, A.; Ye, Y.; Wang, R.; Liu, C.; Roy, K. Going deeper in spiking neural networks: VGG and residual architectures. Frontiers in Neuroscience 2019, 13, 95. [Google Scholar]
- Nair, V.; Hinton, G.E. Rectified linear units improve restricted boltzmann machines. In Proceedings of the Proceedings of the International Conference on Machine Learning, 2010.
- Shen, S.; Zhao, D.; Shen, G.; Zeng, Y. TIM: An Efficient Temporal Interaction Module for Spiking Transformer. arXiv 2024, arXiv:2401.11687. [Google Scholar]
- Schuman, C.D.; Kulkarni, S.R.; Parsa, M.; Mitchell, J.P.; Date, P.; Kay, B. Opportunities for neuromorphic computing algorithms and applications. Nature Computational Science 2022, 2, 10–19. [Google Scholar] [PubMed]
- Zhou, Z.; Che, K.; Fang, W.; Tian, K.; Zhu, Y.; Yan, S.; Tian, Y.; Yuan, L. Spikformer V2: Join the High Accuracy Club on ImageNet with an SNN Ticket. arXiv 2024, arXiv:2401.02020. [Google Scholar]
- Berner, R.; Brandli, C.; Yang, M.; Liu, S.C.; Delbruck, T. A 240× 180 10 mW 12us latency sparse-output vision sensor for mobile applications. In Proceedings of the Symposium on VLSI Circuits. IEEE, 2013, pp. C186–C187.
- Arafa, Y.; ElWazir, A.; ElKanishy, A.; Aly, Y.; Elsayed, A.; Badawy, A.H.; Chennupati, G.; Eidenbenz, S.; Santhi, N. Verified instruction-level energy consumption measurement for NVIDIA GPUs. In Proceedings of the Proceedings of the ACM International Conference on Computing Frontiers, 2020, pp. 60–70.
- Pérez-Carrasco, J.A.; Zhao, B.; Serrano, C.; Acha, B.; Serrano-Gotarredona, T.; Chen, S.; Linares-Barranco, B. Mapping from frame-driven to frame-free event-driven vision systems by low-rate rate coding and coincidence processing–application to feedforward ConvNets. IEEE Transactions on Pattern Analysis and Machine Intelligence 2013, 35, 2706–2719. [Google Scholar]
- Masquelier, T.; Thorpe, S.J. Unsupervised learning of visual features through spike timing dependent plasticity. PLoS Computational Biology 2007, 3, e31. [Google Scholar]
- Lichtsteiner, P.; Posch, C.; Delbruck, T. A 128 × 128 120 dB 15μ s Latency Asynchronous Temporal Contrast Vision Sensor. IEEE Journal of Solid-State Circuits 2008, 43, 566–576. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the Proceedings of the International Conference on Learning Representations, 2015.
- Datta, G.; Liu, Z.; Li, A.; Beerel, P.A. Spiking Neural Networks with Dynamic Time Steps for Vision Transformers. arXiv 2023, arXiv:2311.16456. [Google Scholar]
- Bu, T.; Ding, J.; Yu, Z.; Huang, T. Optimized potential initialization for low-latency spiking neural networks. In Proceedings of the Proceedings of the AAAI Conference on Artificial Intelligence, 2022, Vol. 36, pp. 11–20.
- Eshraghian, J.K.; Ward, M.; Neftci, E.O.; Wang, X.; Lenz, G.; Dwivedi, G.; Bennamoun, M.; Jeong, D.S.; Lu, W.D. Training spiking neural networks using lessons from deep learning. Proceedings of the IEEE 2023, 111. [Google Scholar]
- Rathi, N.; Roy, K. DIET-SNN: A low-latency spiking neural network with direct input encoding and leakage and threshold optimization. IEEE Transactions on Neural Networks and Learning Systems 2021, 34, 3174–3182. [Google Scholar]
- Kim, Y.; Li, Y.; Park, H.; Venkatesha, Y.; Panda, P. Neural architecture search for spiking neural networks. In Proceedings of the Proceedings of the European Conference on Computer Vision, 2022, pp. 36–56.
- Indiveri, G.; Corradi, F.; Qiao, N. Neuromorphic architectures for spiking deep neural networks. In Proceedings of the Proceedings of the IEEE International Electron Devices Meeting, 2015, pp. 4–2.
- Kim, Y.; Chough, J.; Panda, P. Beyond classification: Directly training spiking neural networks for semantic segmentation. Neuromorphic Computing and Engineering 2022. [Google Scholar]
- Calvin, W.H.; Stevens, C.F. Synaptic noise and other sources of randomness in motoneuron interspike intervals. Journal of Neurophysiology 1968, 31, 574–587. [Google Scholar] [PubMed]
- Xing, X.; Gao, B.; Zhang, Z.; Clifton, D.A.; Xiao, S.; Du, L.; Li, G.; Zhang, J. SpikeLLM: Scaling up Spiking Neural Network to Large Language Models via Saliency-based Spiking. arXiv 2024, arXiv:2407.04752. [Google Scholar]
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Advances in neural information processing systems 2012, 25, 1097–1105. [Google Scholar]
- Zhang, W.; Li, P. Temporal spike sequence learning via backpropagation for deep spiking neural networks. Proceeddings of the International Conference on Neural Information Processing 2020, 33, 12022–12033. [Google Scholar]
- Desai, K.; Johnson, J. Virtex: Learning visual representations from textual annotations. In Proceedings of the Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2021, pp. 11162–11173.
- Adrian, E.D.; Zotterman, Y. The impulses produced by sensory nerve endings: Part 3. Impulses set up by Touch and Pressure. The Journal of Physiology 1926, 61, 465. [Google Scholar]
- Furber, S.B.; Galluppi, F.; Temple, S.; Plana, L.A. The SpiNNaker project. Proceedings of the IEEE 2014, 102, 652–665. [Google Scholar]
- Horowitz, M. Energy table for 45nm process. Stanford VLSI wiki 2014. [Google Scholar]
- Chan, V.; Liu, S.C.; van Schaik, A. AER EAR: A matched silicon cochlea pair with address event representation interface. IEEE Transactions on Circuits and Systems 2007, 54, 48–59. [Google Scholar]
- Deng, H.; Zhu, R.; Qiu, X.; Duan, Y.; Zhang, M.; Deng, L. Tensor Decomposition Based Attention Module for Spiking Neural Networks. arXiv 2023, arXiv:2310.14576. [Google Scholar]
- Ma, D.; Jin, X.; Sun, S.; Li, Y.; Wu, X.; Hu, Y.; Yang, F.; Tang, H.; Zhu, X.; Lin, P.; et al. Darwin3: a large-scale neuromorphic chip with a novel ISA and on-chip learning. National Science Review 2024, 11. [Google Scholar]
- VanRullen, R.; Thorpe, S.J. Surfing a spike wave down the ventral stream. Vision Research 2002, 42, 2593–2615. [Google Scholar] [PubMed]
- Merolla, P.A.; Arthur, J.V.; Alvarez-Icaza, R.; Cassidy, A.S.; Sawada, J.; Akopyan, F.; Jackson, B.L.; Imam, N.; Guo, C.; Nakamura, Y.; et al. A million spiking-neuron integrated circuit with a scalable communication network and interface. Science 2014, 345, 668–673. [Google Scholar]
- Wu, J.; Xu, C.; Han, X.; Zhou, D.; Zhang, M.; Li, H.; Tan, K.C. Progressive tandem learning for pattern recognition with deep spiking neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence 2021, 44, 7824–7840. [Google Scholar]
- Brown, T.; Mann, B.; Ryder, N.; Subbiah, M.; Kaplan, J.D.; Dhariwal, P.; Neelakantan, A.; Shyam, P.; Sastry, G.; Askell, A.; et al. Language models are few-shot learners. Advances in neural information processing systems 2020, 33, 1877–1901. [Google Scholar]
- Mukhoty, B.; Bojkovic, V.; de Vazelhes, W.; Zhao, X.; De Masi, G.; Xiong, H.; Gu, B. Direct training of snn using local zeroth order method. Advances in Neural Information Processing Systems 2024, 36, 18994–19014. [Google Scholar]
- Xu, Z.; Lin, M.; Liu, J.; Chen, J.; Shao, L.; Gao, Y.; Tian, Y.; Ji, R. ReCU: Reviving the dead weights in binary neural networks. In Proceedings of the Proceedings of the IEEE/CVF International Conference on Computer Vision, 2021, pp. 5198–5208.
- Kim, Y.; Panda, P. Revisiting batch normalization for training low-latency deep spiking neural networks from scratch. Frontiers in Neuroscience 2021, p. 1638.
- Meng, Q.; Xiao, M.; Yan, S.; Wang, Y.; Lin, Z.; Luo, Z.Q. Training High-Performance Low-Latency Spiking Neural Networks by Differentiation on Spike Representation. In Proceedings of the Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 12444–12453.
- Yi, Z.; Lian, J.; Liu, Q.; Zhu, H.; Liang, D.; Liu, J. Learning rules in spiking neural networks: A survey. Neurocomputing 2023, 531, 163–179. [Google Scholar]
| Neuron Model | Equation Complexity | Biological Plausibility | Computational Cost |
|---|---|---|---|
| Integrate-and-Fire (IF) | Low | Low | Very Low |
| Leaky Integrate-and-Fire (LIF) | Moderate | Medium | Low |
| Hodgkin-Huxley (HH) | High | High | Very High |
| Izhikevich Model | Moderate | High | Medium |
| Adaptive Exponential IF (AdEx) | High | High | High |
| Training Method | Scalability | Biological Plausibility | Computational Cost |
|---|---|---|---|
| Backpropagation with Surrogate Gradients | High | Low | High |
| Spike-Timing-Dependent Plasticity (STDP) | Moderate | High | Low |
| Reward-modulated STDP (R-STDP) | Moderate | High | Medium |
| Hybrid (BP + STDP) | High | Medium | Medium |
| Hardware Platform | Technology | Neuron Capacity | Power Efficiency | Scalability |
|---|---|---|---|---|
| IBM TrueNorth | Digital | neurons | High | Moderate |
| Intel Loihi | Digital | neurons | Very High | High |
| SpiNNaker | Digital | neurons | Moderate | High |
| BrainScaleS | Analog | neurons | Low | Low |
| Tianjic | Hybrid | neurons | High | High |
| Vision System | Energy Consumption | Processing Speed | Data Representation |
|---|---|---|---|
| CNN-based Vision | High | Moderate | Frame-based |
| SNN-based Vision | Low | High | Event-based |
| Training Method | Computational Cost | Energy Efficiency | Scalability |
|---|---|---|---|
| Backpropagation with Surrogate Gradients | High | Moderate | Low |
| STDP-based Learning | Low | High | Medium |
| ANN-to-SNN Conversion | Moderate | High | High |
| Reinforcement Learning in SNNs | High | Low | Low |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the author. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).
