Open Access
18 April 2023 Multi-vendor robustness analysis of a commercial artificial intelligence system for breast cancer detection
Author Affiliations +
Abstract

Purpose

Population-based screening programs for the early detection of breast cancer have significantly reduced mortality in women, but they are resource intensive in terms of time, cost, and workload and still have limitations mainly due to the use of 2D imaging techniques, which may cause overlapping of tissues, and interobserver variability. Artificial intelligence (AI) systems may be a valuable tool to assist radiologist when reading and classifying mammograms based on the malignancy of the detected lesions. However, there are several factors that can influence the outcome of a mammogram and thus also the detection capability of an AI system. The aim of our work is to analyze the robustness of the diagnostic ability of an AI system designed for breast cancer detection.

Approach

Mammograms from a population-based screening program were scored with the AI system. The sensitivity and specificity by means of the area under the receiver operating characteristic (ROC) curve were obtained as a function of the mammography unit manufacturer, demographic characteristics, and several factors that may affect the image quality (age, breast thickness and density, compression applied, beam quality, and delivered dose).

Results

The area under the curve (AUC) from the scoring ROC curve was 0.92 (95% confidence interval = 0.89 − 0.95). It showed no dependence with any of the parameters considered, as the differences in the AUC for different interval values were not statistically significant.

Conclusion

The results suggest that the AI system analyzed in our work has a robust diagnostic capability, and that its accuracy is independent of the studied parameters.

1.

Introduction

Breast cancer is the most common cause of cancer death among women worldwide, as well as the cancer with the highest incidence.1 The implementation of population-based screening programs has significantly reduced mortality among women, as early detection of breast cancer increases the likelihood of successful treatment.2 However, these programs are based on 2D imaging diagnostic methods, which may lead to inaccuracies in terms of sensitivity (true positive rate) and specificity (true negative rate) as the overlapping of tissues can cause certain lesions to be hidden, while clusters of healthy tissue are perceived as lesions. To improve this aspect, screening programs recommend obtaining two views per breast: a craniocaudal (CC) and a mediolateral oblique (MLO) view, which increases the time taken by the radiologist to evaluate the examination. Screening programs are resource-intensive, both in terms of the time spent by radiologists reading mammograms and the financial cost.3 To facilitate the reading of mammograms, computer-aided diagnosis (CAD) systems were introduced in 1980. However, the detection capability of these systems relied on prior knowledge of the physical characteristics of breast lesions, which led to biased results, so the benefit of using CAD in screening is still unclear.4 On the contrary, with the introduction of artificial intelligence (AI), and mostly the introduction of deep learning (DL) algorithms, these systems have indeed proven to be an improvement, as several studies show that the diagnosis of the radiologists improves with the help of an AI system, with no need to lengthen the mammography reading time.46 Moreover, AI-based systems are unaffected by fatigue or subjective diagnosis.

Image quality can be defined on the basis of three important factors, including contrast, spatial resolution, and signal-to-noise ratio (SNR). These factors are in turn altered by parameters related to the image acquisition process, such as breast compression and thickness, which aims at minimizing scattered radiation and thus increasing contrast and SNR, positioning, the x-ray beam quality (determined by the anode/filter combination and voltage), and the radiation dose.7,8 Furthermore, these parameters may affect differently in the detection of different findings suggesting breast cancer, which are mainly soft-tissue lesions and calcifications.9,10 In the case of the former, the margins delimiting the lesion and its density are determinants of its malignancy, whereas in the case of calcifications, their number, morphology and distribution predominate.8 Thus, factors, such as contrast, may affect more severely the detection of soft-tissue lesions, whereas spatial resolution or noise may be more decisive in detecting calcifications.11,12

In addition, breast density (i.e., the percentage or absolute amount of fibroglandular tissue) is also known for being an essential factor affecting image quality. There are several studies claiming that the sensitivity of mammography screening is severely affected in women with high breast density, as the presence of heterogeneous or extremely dense tissue patterns may mask suspicious lesions.1315 Moreover, studies have shown that women with high breast density are at higher risk of developing breast cancer and with more aggressive tumor characteristics,16 which highlights the importance of this property. The age of the woman plays also an important role in breast density, as it decreases with age and with the start of the menopause, although it has been shown that it is not an accurate surrogate for breast density.17 The most frequently used model among radiologists and screening examinations to classify breasts according to their density is the breast imaging-reporting and data system (BI-RADS), which classifies breasts into four categories: fatty, medium dense, heterogeneously dense, and extremely dense breasts.

Neural networks are susceptible to image quality-related parameters, such as noise distortions, contrast, blurring, or resolution.18,19 Therefore, just as image quality can affect image interpretation and detection of breast lesions by expert radiologists,7,20,21 it might also be an influential factor in the performance of a commercial AI-based CAD system. Currently, there are studies analyzing the effect of breast density on these systems,22,23 but to the best of our knowledge, the relationship of their performance with other parameters has not yet been investigated. The exponential development of AI systems applied to diagnostic imaging leads to the need to develop methods to evaluate their performance and determine whether there are parameters that may affect it. Consequently, the aim of this study is to evaluate the performance of a commercial AI system on a large cohort of screening exams as a function of different technical and demographic parameters.

2.

Materials and Methods

On this study, the performance of a stand-alone AI-based screening tool is analyzed. For this purpose, the AI predictions on a particular set of screening examinations (inference) are compared against the clinical diagnostic on the same set (ground truth).

2.1.

AI System

In this work, we have evaluated the robustness of the AI system marketed under the name Transpara® (version 1.6.0, ScreenPoint Medical, The Netherlands). The model is composed of two modules that use DL convolutional neural networks, image analysis algorithms, and feature classifiers to detect both calcifications24,25 and soft tissue lesions (nodules or masses, distortions in the architecture of the breast parenchyma, and asymmetries).2628 The soft tissue and calcification findings (showed as contoured and diamond marked regions, respectively) are subsequently combined to determine the suspicious region findings, assigning each region a value between 1 and 100, which represents the level of suspicion that the lesion is malignant (100 being the highest level of suspicion). Finally, dedicated algorithms are used to combine the scores of all regions detected on the right/left breast CC/MLO images into the exam-based score (AI score), which ranges from 1 to 10 (10 representing the higher probability that cancer is present). This score represents the overall probability of cancer on mammography, and it is calibrated so that approximately the same number of exams classified as normal (which is 10% of the total number of normal exams) falls in each AI score level (1 to 10).

The analyzed system works on full-field digital mammograms and is supported by examinations performed by mammography units of different manufacturers. At the moment of this study, the algorithm had been trained on 8800 biopsy-diagnosed cancer-positive exams, 5000 benign exams diagnosed by biopsy or by 1 year of patient follow-up, and 183,000 noncancerous exams, verified by 1 year of follow-up in the case of clinical diagnostic exams, or by 2 years of follow-up in the case of screening programs. These examinations were originated from devices from four different vendors (Siemens, Hologic, General Electric, Philips) and from institutions across Europe, United States, and Asia.29 Since Transpara is a commercial product and is not an open-source code, it is neither possible to provide further details on the model architecture nor on the train-test split that was used to train and validate the model. However, it is worth noting that none of the images that were used in this study had been previously seen by the algorithm during its training, validation, or testing.

2.2.

Characteristics of the Study

This study has been carried out based on the results obtained when applying the AI system to a collection of mammography exams belonging to a population-based screening program, which were acquired consecutively between January and November 2018. Examinations within this time period were retrospectively collected from the picture archiving and communication system (PACS) without applying any additional selection criteria. The screening exams were all acquired at a single European institution with devices from two different vendors: Mammomat Inspiration (Siemens Healthineers, Forchheim, Germany) and Selenia Dimensions (Hologic Inc, Bedford, Massachusetts, United States), roughly 70% acquired with the former and 30% with the latter. To simplify, from now on, we will refer to these mammography units as “Vendor 1” (Siemens) and “Vendor 2” (Hologic). The imaging protocol consisted of the acquisition of two views per breast (CC and MLO) and the double reading of the mammogram by two radiologists independently. In total, six radiologists were involved in the screening reading in this sample. No information is available on interval cancers (cases diagnosed between consecutive screening rounds) or on the follow-up for normal exams.

Each examination was accompanied by the following details: the age of the woman, the thickness of the breast, the compression applied during the examination, the glandular dose, the density of the breast, the mammography unit in which the examination was performed, with its corresponding anode/filter combination, and the clinical diagnostic (ground truth), dichotomized as 0, without cancer, or 1 with cancer, corroborated by biopsy as the gold standard. For each patient, breast density was estimated independently for each of the four views (CC and MLO for each breast) as the volumetric breast percent dense tissue volume (PDV), using a DL algorithm that works on processed digital mammograms, developed by Vanegas et al.30 However, the mean value of both breasts in CC view was taken as the PDV for each patient, since the value in MLO view may be affected by the presence of the pectoral muscle in the images. The population characteristics are shown in Table 1.

Table 1

Overall demographic and examination-related characteristics of the examinations included in the study [median, range and interquartile range (IQR)]. They are shown for all the examinations, and for those obtained with the unit from Vendor 1 and from Vendor 2.

ParameterTotalSiemensHologic
Number of exams17,77712,125 (68.2%)5,652 (31.8%)
Cancer prevalencea114 (6.4/1000)81 (6.7/1000)33 (5.9/1000)
AgeMedian585857
Range48 to 6948 to 6950 to 69
IQR54 to 6354 to 6353 to 63
Thickness (mm)Median595959
Range13 to 11713 to 11716 to 106
IQR49 to 6849 to 6848 to 68
PDV (%)Median10.159.8610.78
Range4.55 to 44.224.55 to 42.345.07 to 44.22
IQR8.11 to 13.557.83 to 13.238.75 to 14.20
Compression (N)Median117115121
Range16 to 27516 to 22025 to 275
IQR100 to 14099 to 138100 to 145
Glandular dose (mGy)Median1.21.11.7
Range0.04 to 5.20.04 to 4.10.5 to 5.2
IQR0.9 to 1.60.9 to 1.41.2 to 2.1

aRatio of positive cancer examinations, corroborated by biopsy as the gold standard.

Regarding beam quality, Vendor 1 keeps the anode/filter combination of W/Rh for all examined breasts, regardless of thickness [Fig. 1(a)], whereas in the units from Vendor 2, the rhodium filter (W/Rh) is preferred for breasts thickness of less than 70  mm, and silver (W/Ag) for greater thicknesses [Fig. 1(b)]. The characteristics of both units (Vendor 1 and Vendor 2) are shown in Table 2.

Fig. 1

Distribution of exams performed with (a) Vendor 1 (Siemens) and (b) Vendor 2 (Hologic) and the anode/filter combination as a function of the breast thickness.

JMI_10_5_051807_f001.png

Table 2

Technical characteristics of the different mammography units involved in the study.31,32

VendorAnodeFilterDetectorDetector sizeDetector pitchPixel array
SiemensW50  μm RhaSe24.0×30.0  cm85  μm2800×3518
HologicW50  μm RhaSe23.3×28.7  cm70  μm3328×4096
50  μm Ag

aAbbreviations: aSe, amorphous selenium; W, tungsten; Rh, rhodium; Ag, silver.

2.3.

Statistical Analysis

The sensitivity and specificity of the AI system has been studied by means of the area under the curve (AUC) of the receiver operating characteristic (ROC) curve. We have evaluated the influence of the following technical and demographic parameters on the results of the AI algorithm: the mammography unit manufacturer and the corresponding anode/filter combination (beam quality), the round of screening in which the woman is examined, the age of the patient, the compression applied, the thickness of the breast, the PDV, and the delivered glandular dose.

To this end, the AUC of the system was obtained for different intervals of these parameters and then the respective AUC values were compared to determine whether there are statistically significant differences between them. These intervals were defined as follows: two intervals separated by the median for the age, compression, thickness, and dose, and four for the density, separated on the basis of the quartiles. In all cases except for the density, the median was chosen as the threshold value to maximize the statistical power of the comparison, avoiding unbalanced intervals in number of exams (especially of cancer-positive tests). Particularly, breast density was divided based on the quartiles of the distribution to show the possibility of making a comparison of performance between four classes, which was inspired by the BI-RADS classification of breast densities. However, this does not imply a one-to-one relationship between the four intervals and BI-RADS category scales. In the case of the vendor, the anode/filter combinations and the round of screening, no intervals have been distinguished, but the comparison has been made between the AUCs obtained for each case (i.e., Vendor 1 versus Vendor 2; W/Rh versus W/Ag, etc.).

Spearman’s rank correlation coefficient (ρ) was used to obtain the correlations between the different parameters, to show that they are independent and therefore require separate analysis, assuming weak correlation for values 0ρ0.40, and strong correlation for values of ρ0.70.33 A Kruskal–Wallis test was performed to determine whether there were statistically significant differences between PDVs obtained from each view (CC and MLO). Bootstrapping was used to calculate the 95% confidence interval (CI) of the AUC. In the case of this study, bootstrapping analysis was also used to obtain the P-value (P) and to analyze whether there were statistically significant differences between two samples by hypothesis testing. For all statistical tests, we assumed a 95% confidence level, so significance was assumed if P<0.05. The libraries SciPy34 and scikit-learn35 from Python 3.8 have been used in the statistic analysis of this study and to obtain the AUC values of the ROC curves.

3.

Results

3.1.

Correlation between Parameters

The values of the Spearman’s rank correlation coefficients (Table 3) indicate a weak correlation between most parameters except for breast thickness and dose, especially in the case of Vendor 2. With the exception of these two parameters, the lack of correlation between the rest indicates that the analysis should be performed independently for each of them.

Table 3

Spearman’s rank correlation coefficient values (ρ) for the different parameter pairs for the total cohort of exams and for the exams performed by both vendors.

ParametersTotalSiemensHologic
Age versus thickness0.020.020.01
Age versus density0.230.020.23
Thickness versus compression0.270.260.3
Thickness versus dose0.660.660.85
Thickness versus density0.570.590.55
Compression versus dose0.240.20.3
Compression versus density0.280.270.35
Dose versus density0.090.090.28

3.2.

AI System’s Overall Performance

The distribution of screening exams according to the AI score is shown in Fig. 2(b). As seen in the image, normal exams are evenly distributed throughout all the categories, with 10% of all the tests being grouped together in each one, which was expected due to the calibration of the system. The AI system classifies about 90% (87.7%) of all positive tests into category 10, which represents the maximum probability of the presence of cancer, and 96.5% into categories 5 to 10, both inclusive. The AI system’s overall AUC resulted in 0.92 (95% CI = 0.89 to 0.95). The ROC curve is represented in Fig. 2(a). For a threshold score of 10.0, the sensitivity and specificity values are obtained as 87.7% and 87.5%, respectively.

Fig. 2

(a) Distribution of true negative, true positive, and total screen exams as a function of the AI system’s score (1 to 10). It is observed that normal tests are evenly distributed across all categories, with approximately 10% of all tests falling into each category, which is to be expected due to the calibration of the system. (b) ROC Curve and AUC value of the AI system evaluated over all exams.

JMI_10_5_051807_f002.png

3.3.

Dependence of the AI System with the Different Parameters

3.3.1.

Mammography unit manufacturer

The diagnostic capability of the system as a function of the mammography equipment (independent of image quality) was determined by the AI score of the examinations performed with each equipment. The median of the AI scores obtained with Vendor 1 (Me = 5.83) and Vendor 2 (Me = 4.86) was found to be significantly different according to the Kruskal–Wallis test (ΔMe=0.97, P<0.001). For a threshold value of 10.0, the sensitivity and specificity of the system with the exams performed by Vendor 1 is 90.1% and 86.4%, respectively, whereas for the examinations performed by Vendor 2 are 81.8% and 89.7%, respectively . However, as seen in Figs. 3 and 4, the distribution of screening exams as a function of the AI score is similar for both vendors. In addition, no significant differences (P>0.05) were found between the AUC of the AI system obtained only with the examinations performed with Vendor 1 (AUC = 0.93 (95% CI = 0.89 to 0.96)) from those with Vendor 2 (0.92 (95% CI = 0.86 to 0.96)), this difference resulting in 0.01 (95% CI=0.05 to 0.07, P=0.48).

Fig. 3

(a) Distribution of true negative, true positive, and total number of exams as a function of the AI score and (b) ROC curve and AUC value of the AI system for the examinations performed with Vendor 1 (Siemens).

JMI_10_5_051807_f003.png

Fig. 4

(a) Distribution of true negative, true positive, and total number of exams as a function of the AI score and (b) ROC curve and AUC value of the AI system for the examinations performed with Vendor 2 (Hologic).

JMI_10_5_051807_f004.png

3.3.2.

Anode/filter combination

In the case of Vendor 2, the AUC of the model was obtained as a function of the anode/filter combination used (W/Ag or W/Rh, as shown in Table 2). The Kruskal–Wallis test indicates that there are not statistically significant differences between the scores of the two samples (P=0.85). In addition, the AUC with both combinations was obtained 0.90 (95% CI = 0.84 to 0.96) with W/Rh and 0.95 (95% CI = 0.93 to 0.98) with W/Ag, resulting in a not significant difference of 0.05 (95% CI=0.005 to 0.104, P=0.47).

3.3.3.

First round of screening

The distribution of screening examinations performed as a function of the age of the woman is shown in Fig. 5. It has been considered that women between 48 and 51 years old are initiated into the program, therefore being examined in their first round of screening. As seen in the figure, this round concentrates the majority of exams (both cancer and non-cancer cases). The AUC of the system was 0.92 (95% CI = 0.86 to 0.97) with the examinations from the first round and 0.92 (95% CI = 0.89 to 0.96) from the rest of the rounds, which resulted in an AUC difference of 0.0001 (95% CI=0.060 to 0.072, P=0.48), considered not statistically significant.

Fig. 5

Histograms of the number of exams performed resulting in (a) cancer cases (truth = 1) and (b) noncancer cases (truth = 0) according to the age of the women considered in the study.

JMI_10_5_051807_f005.png

3.3.4.

Age of the woman

The ROC analysis has been also performed by dividing mammography exams into two intervals according to the age of the woman [Fig. 6(a)]. The first group comprises women between 48 and 58 years old, as it has been previously shown that the first interval does not affect the performance, whereas the second group comprises women between 59 and 69 years old. The AUC of the system accounting for the exams from the first interval results in 0.92 (95% CI = 0.87 to 0.96) and for the second interval is 0.93 (95% CI = 0.89 to 0.97). The difference between both AUCs is not statistically significant, being 0.01 (95% CI=0.05 to 0.08, P=0.51).

Fig. 6

Comparison of the different ROC curves together with the corresponding AUC (legend) obtained for each of the intervals of each parameter, being (a) age of the woman, (b) thickness of the breast, (c) applied compression, (d) density in terms of PDV, (e) glandular dose from Vendor 1, and (f) glandular dose from Vendor 2.

JMI_10_5_051807_f006.png

3.3.5.

Thickness of the breast

The AUC has been computed for two samples according to the thickness of the breast [Fig. 6(b)]. The first sample corresponds to low thickness breasts (from 13 to 59 mm) and the second to high thickness breasts (from 60 to 117 mm). It has been obtained that the AUC of the system for the examinations corresponding to the first interval (0.92, 95% CI = 0.86 to 0.96) does not show significant differences with the AUC from the second interval (0.93, 95% CI = 0.89 to 0.96), being this difference 0.01 (95% CI=0.04 to 0.08, P=0.49).

3.3.6.

Compression applied

The values representing the compression applied to the breast during the examination were also divided into two intervals: low compression values (20 to 117 N) and higher values (118 to 275 N) [Fig. 6(c)]. The difference between the AUC of the AI system with data from the first interval (0.95, 95% CI = 0.92 to 0.98) and from the second interval (0.90, 95% CI = 0.85 to 0.94) was 0.05 (95% CI=0.003 to 0.11, P=0.49), which is not statistically significant.

3.3.7.

Density of the breast

The AUC was obtained according to each of the four intervals into which the PDV was divided, and compared across all [Fig. 6(d)]. The AUC of the system for each interval and their comparison is shown in Table 4, where it can be seen that the difference between AUCs for each pair of intervals is not statistically significant.

Table 4

AUC values for each of the four density intervals and AUC differences (ΔAUC) between each pair of intervals. For visualization purposes, the comparisons between int 1 versus int 3 and int 4 versus int 3 are not shown, although they present AUC differences of 0.03 (95% CI=−0.05 to 0.11) and 0.001 (95% CI=−0.11 to 0.10), respectively, with P>0.05 for both. The P-value higher than 0.05 indicates that there are no statistically significant differences between intervals.

IntervalPDV range (%)AUCCI (95%)IntervalsΔAUCCI (95%)P
14.55 to 8.110.930.88 to 0.97Int 2 versus Int 10.030.03 to 0.070.50
28.11 to 10.150.960.91 to 0.99Int 2 versus Int 40.050.03 to 0.160.48
310.15 to 13.550.900.84 to 0.96Int 2 versus Int 30.060.02 to 0.140.49
413.55 to 44.220.910.81 to 0.98Int 1 versus Int 40.020.07 to 0.130.48

aAbbreviations: ΔAUC, difference between AUC for each interval.

3.3.8.

Delivered dose

We obtained the dependence of the AI system’s performance with the delivered dose according to the manufacturer of the unit. For each vendor, the dose values were divided into two intervals, as shown in Table 5. In the case of Vendor 1, the difference in AUCs from two intervals is 0.01 (95% CI=0.06 to 0.09, P=0.485) [Fig. 6(e)] and in case of Vendor 2 this difference results in 0.11 (95% CI = 0.003 to 0.23, P=0.476) [Fig. 6(f)], both differences being not statistically significant.

Table 5

Dose intervals for both vendors. The range of the interval and the AUC of the system for the examinations within it are shown.

VendorIntervalRange (mGy)AUCCI (95%)
Siemens10.04 to 1.10.920.85 to 0.97
21.2 to 4.10.930.88 to 0.97
Hologic10.5 to 1.70.850.73 to 0.95
21.8 to 5.20.960.93 to 0.98

4.

Discussion

Image quality has a major impact on the diagnostic rate of AI to evaluate breast cancer, as high-quality images may favor the detection and diagnosis of lesions.36 Therefore, in this work, we have evaluated the predictive ability of the AI system is independent of the mammography unit manufacturer (based on two vendors) Transpara and its potential dependency on certain parameters that may affect the image quality and thus, the prediction of the system.

Spearman’s rank correlation coefficients showed weak correlations between most of the analyzed parameters, except for breast thickness and dose, which show stronger positive correlation especially in the case of Vendor 2. This correlation was expected, as showed by other studies,37 as larger thicknesses require higher doses to maintain optimal image quality.10 In addition, thickness and density also showed an inverse weak correlation with the grade of breast density, which has also been showed in the literature.37 However, no pair of parameters showed such a strong correlation as to be excluded from the analysis.

The obtained AUC index of the system, accounting for all the examinations (0.92, 95% CI = 0.89 to 0.95) indicates high classification accuracy. In addition, the sensitivity (87.7%) and specificity (87.5%) values for a threshold score of 10.0, are in the range of the reference values of sensitivity and specificity for mammographic screening exams in the United States, which are 86.9% and 88.9% respectively, obtained from the Breast Cancer Surveillance Consortium (BCSC).38 The high proportion of correctly classified cases (category 10) indicates that the analyzed system is highly sensitive, which is consistent with the value of the AUC obtained from the ROC curve. This value is also consistent with other studies performed with Transpara for different sets of examinations.4,6,22,29,39

The values for sensitivity and specificity of the system for the examinations performed by each vendor are also similar to the reference values from the BCSC.38 Nevertheless, it is important to note that the number of examinations performed with Vendor 1 is considerably higher than with Vendor 2 (68% versus 32%, respectively). This implies that the data are unbalanced, which may explain why a statistically significant difference was found between the medians of the two groups and the lower AUC for Vendor 2 (Figs. 3 and 4). However, the ratio of positive cancer examinations is similar in both groups (~6/1000), which may explain that this difference was found to be not statistically significant. Therefore, we conclude that the diagnostic capability of Transpara, obtained in terms of sensitivity and specificity, does not depend on the mammography equipment used, in this case, Siemens and Hologic. In addition, it was shown that the anode/filter combination in the case of Vendor 2 units neither affects the accuracy of the system. These results are in agreement with other studies that use multicenter and multi-vendor data (Siemens Healthineers, Hologic, Philips, General Electric) to demonstrate the stand-alone breast cancer detection with Transpara.4,29

Breast cancer screening programs are performed at 2-year intervals. In the first round of screening, when women between 48 and 51 years old are initiated into the program, many of the lesions are seen for the first time, so radiologists refer more women to the specialist oncologist and more cancers are detected. That is why most of the examinations are concentrated on this first round, as seen in Fig. 5. Besides, it has been shown that breast cancer prevalence, the cancer detection rate, and all secondary screening mammography performance measures increase substantially with age.40 However, this study showed that neither the age of the woman nor the fact that the examination belongs to the first round of screening affects the accuracy of the system.

Compression, thickness, and glandular dose are also potential parameters that can affect image quality and thus the AI system capability, as they directly impact upon factors, such as image blurring or spatial resolution, which are crucial for lesion detection in the case of radiologists.8 However, we obtained that the AUC of the system was not affected by different values of these parameters. Nevertheless, regarding the glandular dose given with Vendor 2 [Fig. 6(f)], it is worth noting that the ROC curve for interval 2 (higher doses) comprises a larger AUC than for interval 1 (lower doses), in turn, making the difference greater (0.11 (95% CI = 0.003 to 0.23). This result may be due to the fact that the proportion of cases with cancer is lower in the low-dose interval (4.5/1000 versus 7.2/1000 in the first and second interval, respectively), so more cases would be needed to backup this result.

Breast density (PDV) is the only parameter that has been treated differently, dividing its distribution into four intervals. As indicated in previous sections, this classification was inspired by the BI-RADS classification. Initially, it was intended to establish a one-to-one relationship between interval boundaries and BI-RADS values reported by other authors, such as the Volpara density grades.41 However, since our data are not evenly distributed, the classes were not balanced in terms of the number of exams, thus diminishing the statistical power of the comparison. Nevertheless, it was found that the capacity of Transpara, in terms of AUC, does not vary according to the four defined density categories. This result is in agreement with another two studies that analyze the diagnostic capacity of Transpara as a function of density. In the study by Dustler et al.,23 they found no evidence of significant differences in the risk scores assigned to breast cancer cases in different BIRADS density categories, but point out that the system applies systematically higher scores for normal cases in certain BIRADS density categories. Lauritzen et al.22 found that the AI-based screening worked equally well across all breast densities.

Our results suggest that the AI system analyzed in this work (Transpara®, version 1.6.0, ScreenPoint Medical, The Netherlands) has a robust diagnostic capability, and that its accuracy is independent of the mammography equipment used, beam quality, screening round, woman’s age, breast thickness and applied compression, density defined as PDV, and glandular dose. As future work, it would be interesting to improve the statistical power of the study by increasing the proportion of cases with cancer, as well as to include information on the outcome of the system with benign lesions. With respect to breast density, it would also be worth performing the same analysis on the four BI-RADS densities, in addition to studying the dependence of other parameters that have not been considered in this study, such as the positioning, radiographic techniques, the race or ethnicity of the woman, the reproductive history, or history of cancer.

5.

Conclusion

Population-based screening programs play a crucial role in the early detection of breast cancer, and AI-based systems represent a breakthrough in both speed and accuracy of diagnosis. This will enable more cost-effective scenarios in which the role of the human reader will change significantly. However, before being put into practice, it is necessary to thoroughly validate these systems and demonstrate that their diagnostic capability does not depend on factors related to either image quality or the screening population. The AI system analyzed in this work has been shown to be a robust and highly accurate model, whose diagnostic capability has not been affected by the parameters studied.

Disclosures

The authors declare no relevant financial interests to disclose. At the time of the preparation of the study, the author A.R.R. was working at ScreenPoint Medical (Nijmegen, The Netherlands) as a clinical research scientist. ScreenPoint Medical provided us with the screening program data. This work was originally presented as a final master’s thesis project in 2020 and as an oral communication at the 16th International Workshop of Breast Imaging (IWBI), which took place May 22–25, 2022 in Leuven (Belgium). After this presentation, the corresponding conference proceeding was published at the SPIE Digital Library.42

Acknowledgments

The corresponding author would like to acknowledge the rest of the authors for their advice and support. The authors would also like to thank the final master thesis examining board of Complutense University and the scientific committee of the IWBI for the high esteem in which they held this work. Funding sources: Virtual Clinical Trials and preclinical experiments for optimizing new imaging and AI algorithms in breast cancer (VICTORIA) (Funding body: Ministerio de Ciencia e Innovación [2022–2025], Grant No. PID2021-123390OB-C21, Amount granted: 185,130 Eur. IP: Dr. Robert Marti) and “Experimentos preclínicos para el desarrollo y optimización de nuevas modalidades de imagen para cáncer de mama” (PREDICO) (Funding body: Ministerio de Ciencia e Innovación [2022–2025], Grant No. PID2021-123390OB-C22. Amount granted: 164,197 Eur. IP: D García Pinto, Co-IP: M Chevalier, Coordinated Project No. PID2021-123390OB-C21).

Data, Materials, and Code Availability

The data that have been used to perform this study are not available, although the developed statistical models are available upon request.

References

1. 

J. Ferlay et al., Global Cancer Observatory: Cancer Today, International Agency for Research on Cancer, Lyon, France (2018). Google Scholar

2. 

D. R. Youlden et al., “The descriptive epidemiology of female breast cancer: an international comparison of screening, incidence, survival and mortality,” Cancer Epidemiol., 36 (3), 237 –248 https://doi.org/10.1016/j.canep.2012.02.007 CEBPE4 1055-9965 (2012). Google Scholar

3. 

N. England et al., “The breast imaging and diagnostic workforce in the United Kingdom: results of a survey of NHS breast screening programme units and radiology departments,” (2016). Google Scholar

4. 

A. Rodríguez-Ruiz et al., “Detection of breast cancer with mammography: effect of an artificial intelligence support system,” Radiology, 290 (2), 305 –314 https://doi.org/10.1148/radiol.2018181371 RADLAX 0033-8419 (2019). Google Scholar

5. 

S. M. McKinney et al., “International evaluation of an ai system for breast cancer screening,” Nature, 577 (7788), 89 –94 https://doi.org/10.1038/s41586-019-1799-6 (2020). Google Scholar

6. 

A. Rodríguez-Ruiz et al., “Can we reduce the workload of mammographic screening by automatic identification of normal exams with artificial intelligence? A feasibility study,” Eur. Radiol., 29 (9), 4825 –4832 https://doi.org/10.1007/s00330-019-06186-9 (2019). Google Scholar

7. 

S. H. Taplin et al., “Screening mammography: clinical image quality and the risk of interval breast cancer,” Am. J. Roentgenol., 178 (4), 797 –803 https://doi.org/10.2214/ajr.178.4.1780797 AJROAM 0092-5381 (2002). Google Scholar

8. 

J. Boita et al., “How does image quality affect radiologists’ perceived ability for image interpretation and lesion detection in digital mammography?,” Eur. Radiol., 31 5335 –5343 https://doi.org/10.1007/s00330-020-07679-8 (2021). Google Scholar

9. 

A. E. Burgess, “Effect of detector element size on signal detectability in digital mammography,” Proc. SPIE, 5745 232 –242 https://doi.org/10.1117/12.593770 PSISDG 0277-786X (2005). Google Scholar

10. 

D. B. Copans, Breast Imaging, Lippincott Williams & Wilkins( (2007). Google Scholar

11. 

L. M. Warren et al., “Effect of image quality on calcification detection in digital mammography,” Med. Phys., 39 (6Part1), 3202 –3213 https://doi.org/10.1118/1.4718571 MPHYA6 0094-2405 (2012). Google Scholar

12. 

P. Timberg et al., “Visibility of microcalcification clusters and masses in breast tomosynthesis image volumes and digital mammography: a 4AFC human observer study,” Med. Phys., 39 (5), 2431 –2437 https://doi.org/10.1118/1.3694105 (2012). Google Scholar

13. 

D. Puliti et al., “Volumetric breast density and risk of advanced cancers after a negative screening episode: a cohort study,” Breast Cancer Res., 20 (1), 1 –7 https://doi.org/10.1186/s13058-018-1025-8 BCTRD6 (2018). Google Scholar

14. 

J. O. Wanders et al., “The effect of volumetric breast density on the risk of screen-detected and interval breast cancers: a cohort study,” Breast Cancer Res., 19 (1), 1 –13 https://doi.org/10.1186/s13058-017-0859-9 BCTRD6 (2017). Google Scholar

15. 

C. Huo et al., “Mammographic density-a review on the current understanding of its association with breast cancer,” Breast Cancer Res. Treatment, 144 (3), 479 –502 https://doi.org/10.1007/s10549-014-2901-2 BCTRD6 (2014). Google Scholar

16. 

V. A. McCormack and I. dos Santos Silva, “Breast density and parenchymal patterns as markers of breast cancer risk: a meta-analysis,” Cancer Epidemiol. Prevent. Biomark., 15 (6), 1159 –1169 https://doi.org/10.1158/1055-9965.EPI-06-0034 (2006). Google Scholar

17. 

C. M. Checka et al., “The relationship of mammographic density and age: implications for breast cancer screening,” Am. J. Roentgenol., 198 (3), W292 –W295 https://doi.org/10.2214/AJR.10.6049 AJROAM 0092-5381 (2012). Google Scholar

18. 

S. Dodge and L. Karam, “Understanding how image quality affects deep neural networks,” in Eighth Int. Conf. Quality of Multimedia Exp. (QoMEX), 1 –6 (2016). https://doi.org/10.1109/QoMEX.2016.7498955 Google Scholar

19. 

C. F. Sabottke and B. M. Spieler, “The effect of image resolution on deep learning in radiography,” Radiol. Artif. Intell., 2 (1), e190015 https://doi.org/10.1148/ryai.2019190015 (2020). Google Scholar

20. 

S. A. Feig, “Image quality of screening mammography: effect on clinical outcome,” Am. J. Roentgenol., 178 (4), 805 –807 https://doi.org/10.2214/ajr.178.4.1780805 AJROAM 0092-5381 (2002). Google Scholar

21. 

C. V. Ongeval, H. Bosmans and A. Van Steen, “Current challenges of full field digital mammography,” Radiat. Prot. Dosimetry, 117 (1-3), 148 –153 https://doi.org/10.1093/rpd/nci729 RPDODE 0144-8420 (2005). Google Scholar

22. 

A. D. Lauritzen et al., “An artificial intelligence–based mammography screening protocol for breast cancer: outcome and radiologist workload,” Radiology, 304 (1), 41 –49 https://doi.org/10.1148/radiol.210948 RADLAX 0033-8419 (2022). Google Scholar

23. 

M. Dustler et al., “The effect of breast density on the performance of deep learning-based breast cancer detection methods for mammography,” Proc. SPIE, 11513 541 –546 https://doi.org/10.1117/12.2564328 PSISDG 0277-786X (2020). Google Scholar

24. 

A. Bria, N. Karssemeijer and F. Tortorella, “Learning from unbalanced data: a cascade-based approach for detecting clustered microcalcifications,” Med. Image Anal., 18 (2), 241 –252 https://doi.org/10.1016/j.media.2013.10.014 (2014). Google Scholar

25. 

J.-J. Mordang et al., “Reducing false positives of microcalcification detection systems by removal of breast arterial calcifications,” Med. Phys., 43 (4), 1676 –1687 https://doi.org/10.1118/1.4943376 MPHYA6 0094-2405 (2016). Google Scholar

26. 

R. Hupse and N. Karssemeijer, “Use of normal tissue context in computer-aided detection of masses in mammograms,” IEEE Trans. Med. Imaging, 28 (12), 2033 –2041 https://doi.org/10.1109/TMI.2009.2028611 ITMID4 0278-0062 (2009). Google Scholar

27. 

N. Karssemeijer and G. M. te Brake, “Detection of stellate distortions in mammograms,” IEEE Trans. Med. Imaging, 15 (5), 611 –619 https://doi.org/10.1109/42.538938 ITMID4 0278-0062 (1996). Google Scholar

28. 

N. Karssemeijer, “Automated classification of parenchymal patterns in mammograms,” Phys. Med. Biol., 43 (2), 365 https://doi.org/10.1088/0031-9155/43/2/011 PHMBA7 0031-9155 (1998). Google Scholar

29. 

A. Rodriguez-Ruiz et al., “Stand-alone artificial intelligence for breast cancer detection in mammography: comparison with 101 radiologists,” JNCI: J. Natl. Cancer Inst., 111 (9), 916 –922 https://doi.org/10.1093/jnci/djy222 (2019). Google Scholar

30. 

D. Vanegas et al., “A deep learning method for volumetric breast density estimation from processed full field digital mammograms,” Proc. SPIE, 10950 109500F https://doi.org/10.1117/12.2512818 PSISDG 0277-786X (2019). Google Scholar

31. 

C. Strudley, L. Warren and K. Young, “Technical evaluation of the siemens mammomat inspiration digital breast tomosynthesis system,” (2018). Google Scholar

32. 

C. Strudley, P. Looney and K. Young, “Technical evaluation of Hologic Selenia Dimensions digital breast tomosynthesis system (NHSBSP equipment report 1307 version 2),” (2014). Google Scholar

33. 

J. Fowler, L. Cohen and P. Jarvis, Practical Statistics for Field Biology, John Wiley & Sons( (2013). Google Scholar

34. 

P. Virtanen et al., “SciPy 1.0: fundamental algorithms for scientific computing in Python,” Nat. Methods, 17 261 –272 https://doi.org/10.1038/s41592-019-0686-2 1548-7091 (2020). Google Scholar

35. 

F. Pedregosa et al., “Scikit-learn: machine learning in Python,” J. Mach. Learn. Res., 12 2825 –2830 https://doi.org/10.5555/1953048.2078195 (2011). Google Scholar

36. 

Y.-M. Lei et al., “Artificial intelligence in medical imaging of the breast,” Front. Oncol., 11 600557 https://doi.org/10.3389/fonc.2021.600557 FRTOA7 0071-9676 (2021). Google Scholar

37. 

S. U. Shin et al., “Comparative evaluation of average glandular dose and breast cancer detection between single-view digital breast tomosynthesis (DBT) plus single-view digital mammography (DM) and two-view DM: correlation with breast thickness and density,” Eur. Radiol., 25 (1), 1 –8 https://doi.org/10.1007/s00330-014-3399-z (2015). Google Scholar

38. 

Breast Cancer Surveillance Consortium, “the BCSC and its data collection and sharing activities are funded by grants from the National Cancer Institute (P01CA154292, U54CA163303), Patient-Centered Outcomes Research Institute (PCS-1504-30370), and Agency for Health Research and Quality (R01 HS018366-01A1),” (2022). Google Scholar

39. 

A. J. Wanders et al., “Interval cancer detection using a neural network and breast density in women with negative screening mammograms,” Radiology, 303 (2), 269 –275 https://doi.org/10.1148/radiol.210832 RADLAX 0033-8419 (2022). Google Scholar

40. 

J. D. Keen and J. E. Keen, “How does age affect baseline screening mammography performance measures? A decision model,” BMC Med. Inf. Decis. Mak., 8 (1), 1 –16 https://doi.org/10.1186/1472-6947-8-40 (2008). Google Scholar

41. 

D. van der Waal et al., “Comparing visually assessed bi-rads breast density and automated volumetric breast density software: a cross-sectional study in a breast cancer screening setting,” PLoS One, 10 (9), e0136667 https://doi.org/10.1371/journal.pone.0136667 POLNCL 1932-6203 (2015). Google Scholar

42. 

M. Riveira-Martin, A. Rodríguez-Ruiz and M. Chevalier, “Evaluation of an AI system designed for breast cancer detection,” Proc. SPIE, 12286 241 –248 https://doi.org/10.1117/12.2623686 PSISDG 0277-786X (2022). Google Scholar

Biography

Mercedes Riveira-Martin received her BSc in Physics and MSc in biomedical physics from Complutense University of Madrid. She is currently pursuing a PhD in medical physics from the same university. She is currently working in the SINFONIA Project [funded by Euratom research and training programme 2019 to 2020 under (Grant No. 945196), www.sinfonia-appraisal-eu] at the Galicia Sur Health Research Institute in the Meixoeiro Hospital (Vigo, Spain). Her current interests include, among others, artificial intelligence systems for medical applications.

Alejandro Rodríguez-Ruiz: Biography is not available.

Robert Martí is an associate professor in the Department of Computer Architecture and Technology, and member of the Computer Vision Research Group at the Institute for Research in Computer Vision and Robotics (VICOROB). He has been the coordinator and promoter of the Erasmus Mundus MAIA (Medical Imaging and Applications) and VIBOT (Computer Vision and Robotics) master's degrees and is currently the rector's delegate for international master's degrees. His research is based on the application of machine learning to image analysis, especially in medical imaging for diagnostic.

Margarita Chevalier is a professor in the Department of Radiology, Rehabilitation and Physiotherapy at the Complutense University of Madrid Medical School. She leads the research group “920560 Medical Physics: Dosimetry, Radiation Protection and Medical Imaging” in the same university. Her research is based on digital mammography, digital breast tomosynthesis, and micro-CT applications.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Mercedes Riveira-Martin, Alejandro Rodríguez-Ruiz, Robert Martí, and Margarita Chevalier "Multi-vendor robustness analysis of a commercial artificial intelligence system for breast cancer detection," Journal of Medical Imaging 10(5), 051807 (18 April 2023). https://doi.org/10.1117/1.JMI.10.5.051807
Received: 15 September 2022; Accepted: 3 April 2023; Published: 18 April 2023
Lens.org Logo
CITATIONS
Cited by 1 scholarly publication.
Advertisement
Advertisement
KEYWORDS
Artificial intelligence

Breast

Breast density

Mammography

Cancer detection

Breast cancer

Image quality

Back to Top