Publication Details
Overview
 
 
 

Chapter in Book/ Report/ Conference proceeding

Abstract 

Spiking Neural Networks (SNNs) present strong potential for power-efficient edge applications because of their sparsity and event-driven computation. However, direct training of SNNs is particularly challenging for large and customized transformers because of their non-differentiable spike dynamics. To overcome this challenge, Transformer-to-SNN conversion methods have been proposed to leverage existing architectures, but these approaches often suffer from significant performance degradation and high firing rates, which diminish the efficiency benefits of SNNs. In this paper, we propose a conversion approach inspired by the Few Spikes (FS) method and adapt it to the complex deep unfolding sparse transformer (DUST), enabling the conversion of major transformer components. We validate the proposed spiking DUST on the NVIDIA Jetson Orin Nano platform. Experimental results show that the converted model achieves comparable reconstruction quality to the original DUST while reducing power consumption by approximately 25%, demonstrating its potential for power-efficient edge deployment.

Reference