SpikeRWKV:Energy-efficient Large Language Model with Spiking Neural Network
Authors:
Yulu Zhang, Qianzi Shen, and Zijian Wang
Conference:
ICIC 2025 Posters, Ningbo, China, July 26-29, 2025
Pages:
661-674
Keywords:
spiking neural networks , energy efficiency, spike encoding scheme
Abstract
Spiking Neural Networks SNNs , as the third generation of neural networks, hold great promise for enhancing the energy efficiency of large language models LLMs due to their event-driven computation. However, their naive application in large-scale models typically depends on binary spike simulations over long time steps, making it challenging to balance performance and energy consumption. To address this issue, we propose a Multi-head Spike Encoding scheme with three advantages. First, it enables parallel spike processing to accelerate computation Second, it supports precise representation of positive and negative spikes Third, it mitigates energy surges caused by high-frequency spikes through hierarchical spike decomposition. To demonstrate the effectiveness of our encoding scheme, we introduce SpikeRWKV, an SNN-based adaptation of the RWKV language model. Experimental results demonstrate that SpikeRWKV significantly enhances performance on natural language understanding NLU tasks, achieving a $3.15 times$ reduction in energy consumption compared to the baseline, along with an 8.3 lower perplexity and 5.7 improvement in bits-per-character BPC . Furthermore, SpikeRWKV is $3.88 times$ more energy-efficient than its non-spiking counterpart.
BibTeX Citation:
@inproceedings{ICIC2025,
author = {Yulu Zhang, Qianzi Shen, and Zijian Wang},
title = {SpikeRWKV:Energy-efficient Large Language Model with Spiking Neural Network},
booktitle = {Proceedings of the 21st International Conference on Intelligent Computing (ICIC 2025)},
month = {July},
date = {26-29},
year = {2025},
address = {Ningbo, China},
pages = {661-674},
}