论文标题
对大视力模型的稳定优化基于锥形束CT重建中的深层图像
Stable Optimization for Large Vision Model Based Deep Image Prior in Cone-Beam CT Reconstruction
论文作者
论文摘要
大型视力模型(LVM)最近显示出了医学成像任务的巨大潜力,尽管需要大量的训练数据,但仍可以增强稀疏视锥锥束计算机断层扫描(CBCT)的图像。同时,深层图像先验(DIP)有效指导未经训练的神经网络,以生成高质量的CBCT图像,而无需任何训练数据。但是,原始的DIP方法依赖于定义明确的前向模型和大容量的骨干网络,众所周知,这很难收敛。在本文中,我们为稀疏视图CBCT的无前模型,基于LVM的DIP模型提出了一种稳定的优化方法。我们的方法由两个主要特征组成:(1)多尺度感知损失(MSPL),该损失(MSPL)测量了在多个分辨率下参考图像和输出图像之间的相似性而无需任何正向模型,以及(2)稳定MSPL迭代轨迹的重新加权机制。一种射击优化用于同时稳定地重新重新重量MSPL并优化LVM。我们在两个公开可用的数据集上评估了我们的方法:备用和胡桃木。结果显示出图像质量指标和可视化的显着改善,表明条纹伪像减少。源代码可应要求提供。
Large Vision Model (LVM) has recently demonstrated great potential for medical imaging tasks, potentially enabling image enhancement for sparse-view Cone-Beam Computed Tomography (CBCT), despite requiring a substantial amount of data for training. Meanwhile, Deep Image Prior (DIP) effectively guides an untrained neural network to generate high-quality CBCT images without any training data. However, the original DIP method relies on a well-defined forward model and a large-capacity backbone network, which is notoriously difficult to converge. In this paper, we propose a stable optimization method for the forward-model-free, LVM-based DIP model for sparse-view CBCT. Our approach consists of two main characteristics: (1) multi-scale perceptual loss (MSPL) which measures the similarity of perceptual features between the reference and output images at multiple resolutions without the need for any forward model, and (2) a reweighting mechanism that stabilizes the iteration trajectory of MSPL. One shot optimization is used to simultaneously and stably reweight MSPL and optimize LVM. We evaluate our approach on two publicly available datasets: SPARE and Walnut. The results show significant improvements in both image quality metrics and visualization that demonstrates reduced streak artifacts. The source code is available upon request.