ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.10959
44
1

OuroMamba: A Data-Free Quantization Framework for Vision Mamba Models

13 March 2025
Akshat Ramachandran
Mingyu Lee
Huan Xu
Souvik Kundu
Tushar Krishna
    MQ
ArXivPDFHTML
Abstract

We present OuroMamba, the first data-free post-training quantization (DFQ) method for vision Mamba-based models (VMMs). We identify two key challenges in enabling DFQ for VMMs, (1) VMM's recurrent state transitions restricts capturing of long-range interactions and leads to semantically weak synthetic data, (2) VMM activations exhibit dynamic outlier variations across time-steps, rendering existing static PTQ techniques ineffective. To address these challenges, OuroMamba presents a two-stage framework: (1) OuroMamba-Gen to generate semantically rich and meaningful synthetic data. It applies contrastive learning on patch level VMM features generated through neighborhood interactions in the latent state space, (2) OuroMamba-Quant to employ mixed-precision quantization with lightweight dynamic outlier detection during inference. In specific, we present a thresholding based outlier channel selection strategy for activations that gets updated every time-step. Extensive experiments across vision and generative tasks show that our data-free OuroMamba surpasses existing data-driven PTQ techniques, achieving state-of-the-art performance across diverse quantization settings. Additionally, we implement efficient GPU kernels to achieve practical latency speedup of up to 2.36x. Code will be released soon.

View on arXiv
@article{ramachandran2025_2503.10959,
  title={ OuroMamba: A Data-Free Quantization Framework for Vision Mamba Models },
  author={ Akshat Ramachandran and Mingyu Lee and Huan Xu and Souvik Kundu and Tushar Krishna },
  journal={arXiv preprint arXiv:2503.10959},
  year={ 2025 }
}
Comments on this paper