243
v1v2v3 (latest)

Time-Aware One Step Diffusion Network for Real-World Image Super-Resolution

Main:8 Pages
12 Figures
Bibliography:2 Pages
7 Tables
Appendix:7 Pages
Abstract

Diffusion-based real-world image super-resolution (Real-ISR) methods have demonstrated impressivethis http URLachieve efficient Real-ISR, many works employ Variational Score Distillation (VSD) to distill pre-trained stable-diffusion (SD) model for one-step SR with a fixed timestep. However, since SD will perform different generative priors at different timesteps, a fixed timestep is difficult for these methods to fully leverage the generative priors in SD, leading to suboptimalthis http URLaddress this, we propose a \textbf{T}ime-\textbf{A}ware one-step \textbf{D}iffusion Network for Real-ISR (\textbf{TADSR}). We first introduce a Time-Aware VAE Encoder, which projects the same image into different latent features based onthis http URLjoint dynamic variation of timesteps and latent features, the student model can better align with the input pattern distribution of the pre-trained SD, thereby enabling more effective utilization of SD's generativethis http URLbetter activate the generative prior of SD at different timesteps, we propose a Time-Aware VSD loss that bridges the timesteps of the student model and those of the teacher model, thereby producing more consistent generative prior guidance conditioned on timesteps. Additionally, though utilizing the generative prior in SD at different timesteps, our method can naturally achieve \textbf{controllable trade-offs between fidelity and realism} by changing thethis http URLresults demonstrate that our method achieves both state-of-the-art performance and controllable SR results with only a single step. The source codes are released atthis https URL

View on arXiv
Comments on this paper