PurposeSupervised deep convolutional neural network (CNN)-based methods have been actively used in clinical CT to reduce image noise. The networks of these methods are typically trained using paired high- and low-quality data from a massive number of patients and/or phantom images. This training process is tedious, and the network trained under a given condition may not be generalizable to patient images acquired and reconstructed under different conditions. We propose a self-trained deep CNN (ST_CNN) method for noise reduction in CT that does not rely on pre-existing training datasets.ApproachThe ST_CNN training was accomplished using extensive data augmentation in the projection domain, and the inference was applied to the data itself. Specifically, multiple independent noise insertions were applied to the original patient projection data to generate multiple realizations of low-quality projection data. Then, rotation augmentation was adopted for both the original and low-quality projection data by applying the rotation angle directly on the projection data so that images were rotated at arbitrary angles without introducing additional bias. A large number of paired low- and high-quality images from the same patient were reconstructed and paired for training the ST_CNN model.ResultsNo significant difference was found between the ST_CNN and conventional CNN models in terms of the peak signal-to-noise ratio and structural similarity index measure. The ST_CNN model outperformed the conventional CNN model in terms of noise texture and homogeneity in liver parenchyma as well as better subjective visualization of liver lesions. The ST_CNN may sacrifice the sharpness of vessels slightly compared to the conventional CNN model but without affecting the visibility of peripheral vessels or diagnosis of vascular pathology.ConclusionsThe proposed ST_CNN method trained from the data itself may achieve similar image quality in comparison with conventional deep CNN denoising methods pre-trained on external datasets.
Detection of low contrast liver metastases varies between radiologists. Training may improve performance for lower-performing readers and reduce inter-radiologist variability. We recruited 31 radiologists (15 trainees, eight non-abdominal staff, and eight abdominal staff) to participate in four separate reading sessions: pre-test, search training, classification training, and post-test. In the pre-test, each radiologist interpreted 40 liver CT exams containing 91 metastases, circumscribed suspected hepatic metastases while under eye tracker observation, and rated confidence. In search training, radiologists interpreted a separate set of 30 liver CT exams while receiving eye tracker feedback and after coaching to increase use of coronal reformations, interpretation time, and use of liver windows. In classification training, radiologists interpreted up to 100 liver CT image patches, most with benign or malignant lesions, and compared their annotations to ground truth. Post-test was identical to pre-test. Between pre- and post-test, sensitivity increased by 2.8% (p = 0.01) but AUC did not change significantly. Missed metastases were classified as search errors (<2 seconds gaze time) or classification errors (>2 seconds gaze time) using the eye tracker. Out of 2775 possible detections, search errors decreased (10.8% to 8.1%; p < 0.01) but classification errors were unchanged (5.7% vs 5.7%). When stratified by difficulty, easier metastases showed larger reductions in search errors: for metastases with average sensitivity of 0-50%, 50-90%, and 90-100%, reductions in search errors were 16%, 35%, and 58%, respectively. The training program studied here may be able to improve radiologist performance by reducing errors but not classification errors.
PurposeDeep convolutional neural network (CNN)-based methods are increasingly used for reducing image noise in computed tomography (CT). Current attempts at CNN denoising are based on 2D or 3D CNN models with a single- or multiple-slice input. Our work aims to investigate if the multiple-slice input improves the denoising performance compared with the single-slice input and if a 3D network architecture is better than a 2D version at utilizing the multislice input.ApproachTwo categories of network architectures can be used for the multislice input. First, multislice images can be stacked channel-wise as the multichannel input to a 2D CNN model. Second, multislice images can be employed as the 3D volumetric input to a 3D CNN model, in which the 3D convolution layers are adopted. We make performance comparisons among 2D CNN models with one, three, and seven input slices and two versions of 3D CNN models with seven input slices and one or three output slices. Evaluation was performed on liver CT images using three quantitative metrics with full-dose images as reference. Visual assessment was made by an experienced radiologist.ResultsWhen the input channels of the 2D CNN model increases from one to three to seven, a trend of improved performance was observed. Comparing the three models with the seven-slice input, the 3D CNN model with a one-slice output outperforms the other models in terms of noise texture and homogeneity in liver parenchyma as well as subjective visualization of vessels.ConclusionsWe conclude the that multislice input is an effective strategy for improving performance for 2D deep CNN denoising models. The pure 3D CNN model tends to have a better performance than the other models in terms of continuity across axial slices, but the difference was not significant compared with the 2D CNN model with the same number of slices as the input.
Supervised deep convolutional neural network (CNN)-based methods have been actively used in clinical CT to reduce image noise. The networks of these methods are typically trained using paired high- and low-quality data from a large number of patients and/or phantom images. This training process is tedious, and the network trained under a given condition may not be generalizable to patient images acquired and reconstructed at different conditions. In this paper, we propose a self-trained deep CNN (ST_CNN) method which does not rely on pre-existing training datasets. The training is accomplished using extensive data augmentation through projection domain and the inference is applied to the data itself. Preliminary evaluation on patient images demonstrated that the proposed method could achieve similar image quality in comparison with conventional deep CNN denoising methods pre-trained on external datasets.
Purpose: Radiologists exhibit wide inter-reader variability in diagnostic performance. This work aimed to compare different feature sets to predict if a radiologist could detect a specific liver metastasis in contrast-enhanced computed tomography (CT) images and to evaluate possible improvements in individualizing models to specific radiologists.Approach: Abdominal CT images from 102 patients, including 124 liver metastases in 51 patients were reconstructed at five different kernels/doses using projection domain noise insertion to yield 510 image sets. Ten abdominal radiologists marked suspected metastases in all image sets. Potentially salient features predicting metastasis detection were identified in three ways: (i) logistic regression based on human annotations (semantic), (ii) random forests based on radiologic features (radiomic), and (iii) inductive derivation using convolutional neural networks (CNN). For all three approaches, generalized models were trained using metastases that were detected by at least two radiologists. Conversely, individualized models were trained using each radiologist’s markings to predict reader-specific metastases detection.Results: In fivefold cross-validation, both individualized and generalized CNN models achieved higher area under the receiver operating characteristic curves (AUCs) than semantic and radiomic models in predicting reader-specific metastases detection ability (p < 0.001). The individualized CNN with an AUC of mean (SD) 0.85(0.04) outperformed the generalized one [AUC = 0.78 ( 0.06 ) , p = 0.004]. The individualized semantic [AUC = 0.70 ( 0.05 ) ] and radiomic models [AUC = 0.68 ( 0.06 ) ] outperformed the respective generalized versions [semantic AUC = 0.66 ( 0.03 ) , p = 0.009; radiomic AUC = 0.64 ( 0.06 ) , p = 0.03].Conclusions: Individualized models slightly outperformed generalized models for all three feature sets. Inductive CNNs were better at predicting metastases detection than semantic or radiomic features. Generalized models have implementation advantages when individualized data are unavailable.
Deep convolutional neural network (CNN) based methods have become popular choices for reducing image noise in CT. Some of these methods showed promising results, especially in terms of preserving natural CT noise texture. Early attempts of CNN denoising were based on 2D CNN models with either single-slice or 3-slice input. The 3-slice input was mainly to utilize the existing network architecture that were proposed for natural images with 3 input channels. Multi-slice input has the potential to incorporate spatial information from adjacent slices. However, it remains unknown if this strategy indeed improves the denoising performance compared to a 2D model with a single-slice input and what is the best network architecture to utilize the multi-slice input. Two categories of network architectures can be used for multi-slice input. First, multi-slice low-dose images can be stacked channelwise as multi-channel input to a 2D CNN model. Second, multi-slice images can be employed as the 3D volumetric input to a 3D CNN model, in which the 3D convolution layers are adopted. In this study, we compare the performance of multiple CNN models with 1, 3, and 7 input slices. For the 7-slice input, we also include a comparison between 2D and 3D CNN models. When the input channels of the 2D CNN model increases from 1 to 3 to 7, a trend of improved performance was observed. Comparing the two models with 7-slice input, the 3D model slightly outperforms the 2D model in terms of noise texture and homogeneity in liver parenchyma as well as better subjective visualization of vessels such as intrahepatic portal vein and jejunal artery.
Eye-tracking techniques can be used to understand the visual search process in diagnostic radiology. Nonetheless, most prior eye-tracking studies in CT only involved single cross-sectional images or video playback of the reconstructed volume and meanwhile applied strong constraints to reader-image interactivity, yielding a disconnection between the corresponding experimental setup and clinical reality. To overcome this limitation, we developed an eye-tracking system that integrates eye-tracking hardware with in-house-built image viewing software. This system enabled recording of radiologists’ real-time eye-movement and interactivity with the displayed images in clinically relevant tasks. In this work, the system implementation was demonstrated, and the spatial accuracy of eye-tracking data was evaluated using digital phantom images and patient CT angiography exam. The measured offset between targets and gaze points was comparable to that of many prior eye-tracking systems (The median offset: phantom – visual angle ~0.8°; patient CTA – visual angle ~0.7 – 1.3°). Further, the eye-tracking system was used to record radiologists’ visual search in a liver lesion detection task with contrast-enhanced abdominal CT. From the measured data, several variables were found to correlate with radiologists’ sensitivity, e.g., mean sensitivity of readers with longer interpretation time was higher than that of the others (88 ± 3% vs 78 ± 10%; p < 0.001). In summary, the proposed eye-tracking system has the potential of providing high-quality data to characterize radiologists’ visual-search process in clinical CT tasks.
There is substantial variability in the performance of radiologist readers. We hypothesized that certain readers may have idiosyncratic weaknesses towards certain types of lesions, and unsupervised learning techniques might identify these patterns. After IRB approval, 25 radiologist readers (9 abdominal subspecialists and 16 non-specialists or trainees) read 40 portal phase liver CT exams, marking all metastases and providing a confidence rating on a scale of 1 to 100. We formed a matrix of reader confidence ratings, with rows corresponding to readers, and columns corresponding to metastases, and each matrix entry providing the confidence rating that a reader gave to the metastasis, with zero confidence used for lesions that were not marked. A clustergram was used to permute the rows and columns of this matrix to group similar readers and metastases together. This clustergram was manually interpreted. We found a cluster of lesions with atypical presentation that were missed by several readers, including subspecialists, and a separate cluster of small, subtle lesions where subspecialists were more confident of their diagnosis than trainees. These and other observations from unsupervised learning could inform targeted training and education of future radiologists.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.