ResearchTrend.AI
  • Communities
  • Connect sessions
  • AI calendar
  • Organizations
  • Join Slack
  • Contact Sales
Papers
Communities
Social Events
Terms and Conditions
Pricing
Contact Sales
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2026 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.17729
494
1
v1v2v3 (latest)

Fast convolution algorithm for state space models

22 November 2024
Gregory Beylkin
ArXiv (abs)PDFHTML
Main:3 Pages
Bibliography:2 Pages
Appendix:1 Pages
Abstract

We present an unconditionally stable algorithm for applying matrix transfer function of a linear time invariant system (LTI) in time domain. The state matrix of an LTI system used for modeling long range dependencies in state space models (SSMs) has eigenvalues close to 111. The standard recursion defining LTI system becomes unstable if the m×mm\times mm×m state matrix has just one eigenvalue with absolute value even slightly greater than 1. This may occur when approximating a state matrix by a structured matrix to reduce the cost of matrix-vector multiplication from O(m2)\mathcal{O}\left(m^{2}\right)O(m2) to O(m)\mathcal{O}\left(m\right)O(m) or O(mlog⁡m).\mathcal{O}\left(m\log m\right).O(mlogm). We introduce an unconditionally stable algorithm that uses an approximation of the rational transfer function in the z-domain by a matrix polynomial of degree 2N+1−12^{N+1}-12N+1−1, where NNN is chosen to achieve any user-selected accuracy. Using a cascade implementation in time domain, applying such transfer function to compute LLL states requires no more than 2L2L2L matrix-vector multiplications (whereas the standard recursion requires LLL matrix-vector multiplications). However, using unconditionally stable algorithm, it is not necessary to assure that an approximate state matrix has all eigenvalues with absolute values strictly less than 1 i.e., within the desired accuracy, the absolute value of some eigenvalues may possibly exceed 111. Consequently, this algorithm allows one to use a wider variety of structured approximations to reduce the cost of matrix-vector multiplication and we briefly describe several of them to be used for this purpose.

View on arXiv
Comments on this paper