ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2411.11282
26
1

Continuous K-space Recovery Network with Image Guidance for Fast MRI Reconstruction

18 November 2024
Yucong Meng
Zhiwei Yang
Minghong Duan
Yonghong Shi
Zhijian Song
ArXivPDFHTML
Abstract

Magnetic resonance imaging (MRI) is a crucial tool for clinical diagnosis while facing the challenge of long scanning time. To reduce the acquisition time, fast MRI reconstruction aims to restore high-quality images from the undersampled k-space. Existing methods typically train deep learning models to map the undersampled data to artifact-free MRI images. However, these studies often overlook the unique properties of k-space and directly apply general networks designed for image processing to k-space recovery, leaving the precise learning of k-space largely underexplored. In this work, we propose a continuous k-space recovery network from a new perspective of implicit neural representation with image domain guidance, which boosts the performance of MRI reconstruction. Specifically, (1) an implicit neural representation based encoder-decoder structure is customized to continuously query unsampled k-values. (2) an image guidance module is designed to mine the semantic information from the low-quality MRI images to further guide the k-space recovery. (3) a multi-stage training strategy is proposed to recover dense k-space progressively. Extensive experiments conducted on CC359, fastMRI, and IXI datasets demonstrate the effectiveness of our method and its superiority over other competitors.

View on arXiv
@article{meng2025_2411.11282,
  title={ Continuous K-space Recovery Network with Image Guidance for Fast MRI Reconstruction },
  author={ Yucong Meng and Zhiwei Yang and Minghong Duan and Yonghong Shi and Zhijian Song },
  journal={arXiv preprint arXiv:2411.11282},
  year={ 2025 }
}
Comments on this paper