Blind Video Super-Resolution based on Implicit Kernels
Blind video super-resolution (BVSR) is a low-level vision task which aims to generate high-resolution videos from low-resolution counterparts in unknown degradation scenarios. Existing approaches typically predict blur kernels that are spatially invariant in each video frame or even the entire video. These methods do not consider potential spatio-temporal varying degradations in videos, resulting in suboptimal BVSR performance. In this context, we propose a novel BVSR model based on Implicit Kernels, BVSR-IK, which constructs a multi-scale kernel dictionary parameterized by implicit neural representations. It also employs a newly designed recurrent Transformer to predict the coefficient weights for accurate filtering in both frame correction and feature alignment. Experimental results have demonstrated the effectiveness of the proposed BVSR-IK, when compared with four state-of-the-art BVSR models on three commonly used datasets, with BVSR-IK outperforming the second best approach, FMA-Net, by up to 0.59 dB in PSNR. Source code will be available atthis https URL.
View on arXiv@article{zhu2025_2503.07856, title={ Blind Video Super-Resolution based on Implicit Kernels }, author={ Qiang Zhu and Yuxuan Jiang and Shuyuan Zhu and Fan Zhang and David Bull and Bing Zeng }, journal={arXiv preprint arXiv:2503.07856}, year={ 2025 } }