Late Breaking Results: Quamba-SE: Soft-edge Quantizer for Activations in State Space Models
2 authors
arXiv:2601.09451v1
Authors
Yizhi ChenAhmed Hemani
Abstract
We propose Quamba-SE, a soft-edge quantizer for State Space Model (SSM) activation quantization. Unlike existing methods, using standard INT8 operation, Quamba-SE employs three adaptive scales: high-precision for small values, standard scale for normal values, and low-precision for outliers. This preserves outlier information instead of hard clipping, while maintaining precision for other values. We evaluate on Mamba- 130M across 6 zero-shot benchmarks. Results show that Quamba- SE consistently outperforms Quamba, achieving up to +2.68% on individual benchmarks and up to +0.83% improvement in the average accuracy of 6 datasets.
Paper Information
- arXiv ID:
- 2601.09451v1
- Published:
- Categories:
- cs.LG, cs.AI, cs.AR