Convolutional neural networks (CNNs) have been widely used for low-dose CT (LDCT) image denoising. To alleviate the over-smoothing effect caused by conventional mean squared error, researchers often resort to adversarial training to achieve faithful structural and texture recovery. On one hand, such adversarial training is typically difficult to train and may lead to a potential CT value shift. On the other hand, these CNNs-based denoising models usually generalize poorly to new unseen dose levels. Recently, diffusion models have exhibited higher image quality and stable trainability compared to other generative models. Therefore, we present a Contextual Conditional Diffusion model (CoCoDiff) for low-dose CT denoising, which aims to address the issues of existing denoising models. More specifically, during the training stage, we train a noise estimation network to gradually convert a residual image to a Gaussian distribution based on a Markov chain with a low-dose image as the condition. During the inference stage, the Markov chain is reversed to generate the residual image from random Gaussian noise. In addition, the contextual information of adjacent slices is also utilized for noise estimation to suppress potential structural distortion. Experimental results on Mayo LDCT datasets show that the proposed model can restore faithful structural details and generalize well to other unseen noise levels.
|