Context-Aware Regularization with Markovian Integration for Attention-Based Nucleotide Analysis
Main:10 Pages
8 Figures
Bibliography:3 Pages
21 Tables
Appendix:14 Pages
Abstract
Transformers have revolutionized nucleotide sequence analysis, yet capturing long-range dependencies remains challenging. Recent studies show that autoregressive transformers often exhibit Markovian behavior by relying on fixed-length context windows for next-token prediction. However, standard self-attention mechanisms are computationally inefficient for long sequences due to their quadratic complexity and do not explicitly enforce global transition consistency.
View on arXivComments on this paper
