Open Access
16 November 2023 Sequestration of imaging studies in MIDRC: stratified sampling to balance demographic characteristics of patients in a multi-institutional data commons
Author Affiliations +
Abstract

Purpose

The Medical Imaging and Data Resource Center (MIDRC) is a multi-institutional effort to accelerate medical imaging machine intelligence research and create a publicly available image repository/commons as well as a sequestered commons for performance evaluation and benchmarking of algorithms. After de-identification, approximately 80% of the medical images and associated metadata become part of the open commons and 20% are sequestered from the open commons. To ensure that both commons are representative of the population available, we introduced a stratified sampling method to balance the demographic characteristics across the two datasets.

Approach

Our method uses multi-dimensional stratified sampling where several demographic variables of interest are sequentially used to separate the data into individual strata, each representing a unique combination of variables. Within each resulting stratum, patients are assigned to the open or sequestered commons. This algorithm was used on an example dataset containing 5000 patients using the variables of race, age, sex at birth, ethnicity, COVID-19 status, and image modality and compared resulting demographic distributions to naïve random sampling of the dataset over 2000 independent trials.

Results

Resulting prevalence of each demographic variable matched the prevalence from the input dataset within one standard deviation. Mann–Whitney U test results supported the hypothesis that sequestration by stratified sampling provided more balanced subsets than naïve randomization, except for demographic subcategories with very low prevalence.

Conclusions

The developed multi-dimensional stratified sampling algorithm can partition a large dataset while maintaining balance across several variables, superior to the balance achieved from naïve randomization.

1.

Introduction

Since early 2020, the COVID-19 pandemic has presented an urgent and critical public health crisis. This crisis produced several needs for essential biomedical research and development to address, including early detection and differential diagnosis of COVID-19, prognosis and assessment of response to treatment, and monitoring of the post-COVID patient. All these needs created an important role for artificial intelligence (AI) in medical imaging of the COVID patient and stimulated a rapid research effort in AI model development for COVID-19 applications.13 Many studies have found promising performance of AI models for various applications. However, potential impact of algorithm bias and lack of clinical utility have been noted as major shortcomings in AI developed for COVID-19 medical imaging, particularly algorithms developed early in the pandemic.46 As a response to this urgent public health need, the Medical Imaging and Data Resource Center (MIDRC7) was established in August, 2020, to accelerate medical imaging machine intelligence research for COVID-19.

MIDRC is a multi-institutional research collaboration between the American College of Radiology (ACR®), the Radiological Society of North America (RSNA), and the American Association of Physicists in Medicine (AAPM), created to address critical gaps in resources and technology for AI in medical imaging. Through the work of five technology development projects and 12 collaborating research projects, MIDRC is providing processes for data intake, de-identification, quality assessment, and distributed public access in addition to organizing research challenges and curated datasets to support high-quality research methods. The aim of MIDRC is to accelerate machine intelligence research for COVID-19 and eventually for other diseases that utilize medical imaging in, e.g., detection, diagnosis, or prognosis. One primary component of MIDRC is the development of a publicly available image repository,7 as well as a sequestered database for performance evaluation and benchmarking of algorithms.

While the majority of the de-identified data (both images and metadata) submitted to MIDRC are open, i.e., accessible to the public, approximately 20% of the data are being sequestered from public use for the purpose of machine intelligence algorithm evaluation. These sequestered data will act as a large base from which task-based samples, or “test-sets,” can be drawn to provide an estimate of an algorithm’s performance or generalization ability, without ever releasing the data publicly or giving users direct access to the cases used for testing, thus maintaining the integrity of the test set.8

To assure both datasets are similarly representative of the population, we developed a methodology to balance demographic characteristics, or variables, across the sequestered and public data. This process is implemented for incoming batches of data on an ongoing basis. While MIDRC aims to provide a platform with a wide array of diverse data representative of the population, gaps and biases may inadvertently arise. MIDRC aims to recognize and address these biases throughout the research workflow through multiple bias-reduction strategies and resources for users to further reduce potential bias.9 Providing a public resource of curated COVID-19 data and open-source algorithms for data analysis and cohort building will aid in bias mitigation by increasing availability of multi-institutional data and standardized data labels across many demographic variables.10 Further, purposeful selection of patients for the sequestered database will be a useful tool for the future assessment of the bias of the evolving database itself, as well as the algorithms developed based on the database.

Potential bias and lack of generalizability in AI algorithms have been key shortcomings of the clinical utility of AI, and multiple studies have found demographics to have a profound impact on the performance of medical image classifiers.1015 Biases can also be hidden in data structures that are not explicitly defined or monitored, leading to unintended bias in resulting model performance.1618 These studies highlight the key importance of mindful balancing of demographic distributions within datasets used for machine intelligence applications.14

Balancing multiple variables among subgroups is a widely studied topic in the field of clinical trial development.19,20 Shifting demographic profiles of the study population have also been investigated in the context of clinical trials.21 However, similar approaches have rarely been applied in the field of machine intelligence due to the use of typical train-test splitting in datasets that are often small or moderate in size. In both clinical trials and imaging data, there are similar patient variables, but imaging contributes additional complexity regarding imaging machine type, protocol, etc., as the data collection process is more varied. Here, we apply a developed method of multi-dimensional stratified sampling to separate incoming MIDRC data into open and sequestered commons and evaluate the performance of the sampling method in terms of similarities between the two commons. Our method differs from existing approaches in that it incorporates several demographic variables and is used on an ongoing basis to continually build two demographically matched data commons, at scale. In addition, the methodology developed for this work takes the standard concept of stratification and expands upon it in a script that is shared as a public resource on the MIDRC GitHub website.

The goal of the developed sequestration algorithm is to generate two commons with demographically similar characteristics. Developers can then sample and train on any portion of the publicly available data, and algorithm performance will be assessed with portions of the sequestered commons relevant to the user’s task. Our goal is to create a balance between the open data that developers will use for training algorithms and the sequestered data that will provide a balanced test set. The purpose of this work is to describe this process and provide a brief evaluation of the current state of sequestration of imaging studies within MIDRC.

2.

Methods

Prior to sequestration, de-identified clinical data of patients are submitted to MIDRC through data input portals hosted by ACR and RSNA.7 Quality of submitted medical imaging studies is assessed, and the associated metadata are harmonized for representation within the MIDRC data model at Ref. 7. Subsets of the incoming data undergo separation and are designated as “open” or “sequestered” on an ongoing basis in batches created at regular submission time intervals.

To sequester approximately 20% of an incoming data batch, first, de-identified patient IDs are compared across all previously processed batches. If a patient ID already exists in either the public or sequestered commons, the incoming data for this patient are placed in the relevant commons. This process ensures data are placed into the open or sequestered commons at the patient level, and all images from longitudinal studies of a given patient are contained in only one of the two commons. Following this longitudinal data check, the data of remaining patients in the intake batch are sequentially separated into multiple strata based upon the anonymized clinical site ID, image modality, COVID status (whether a patient ever tested positive for COVID), and reported patient race, age, sex at birth, and ethnicity. These variables were selected from the standard variables included in the Ref. 7 patient metadata by the authors after consideration of the potential impact of each variable on future developed algorithms and importance of balanced representation of that variable between the two commons. The number of included variables is a compromise between balanced representation across many variables but few enough variables that stratification does not creates a majority of bins with such low prevalence that they will not be sufficiently balanced. The selected variables are not a comprehensive list, but represent common variables considered in ML algorithm subgroup analyses and have been shown in the literature to have potential impact on the performance of developed classifiers.1114 A diagram of the sequestration process is shown in Fig. 1. Within each resulting bin or strata, i.e., a group of patients with a particular combination of characteristics, the patients are randomly assigned to the open dataset or the sequestered dataset with proportions of approximately 80% and 20%, respectively. Thus, for n variables of interest, the balance of the n-dimensional distribution of variable combinations can be controlled.

Fig. 1

Diagram of demographic factors used to stratify data for separation into sequestered and open databases. The input data batch is sequentially split into all possible variations of each category until an individual stratum, containing a unique combination of variables, is achieved. This individual stratum is then randomly separated into the open and sequestered commons with proportions of approximately 80:20.

JMI_10_6_064501_f001.png

Since, for any given patient, imaging studies from multiple modalities might be available, such as computed tomographs (CT) or radiographs, the sequestering of patient data by modality was accomplished by first identifying the most prevalent image modalities in an intake data batch. If a patient entry contained more than one modality, the modality most prevalent in the intake data batch was established to be that patient’s “primary” modality, and images from any less prevalent modality were assigned along with the most prevalent modality. Modalities present in this analysis included computed radiograph (CR), CT, digital radiography (DX), and magnetic resonance (MR) images. Patient age data were grouped into categories matching the age group categories provided by the Center for Disease Control COVID-19 database.22 Patient sex at birth, race, and ethnicity were grouped in agreement with the categories defined by the NIH.23,24

To demonstrate the sequestration process in simulation, 5000 patients were randomly selected from the public data commons7 to serve as an example of an input dataset. This number of patients approximately models a typical data submission from a single clinical site to MIDRC from 2021 to 2022. As such, the example input dataset was not separated by contributing clinical site. It is important to note that users will not be training or testing on any single batch of data, rather all cumulative data batches (after being sampled using the developed sequestering method) will be available to users from the public data commons. The developed sequestration algorithm was applied to the remaining demographic variables (age, race, sex at birth, ethnicity, COVID-19 status, and image modality) to achieve a similar distribution of variables between the input dataset and the two subsets, i.e., the open and the sequestered data commons. The variables of age, race, sex at birth, ethnicity, COVID-19 status, and image modality contained 9, 7, 4, 3, 3, and 4 categories, as listed in Table 1, respectively, resulting in a total of 9072 strata. As such, strata used in separating data are multiplicative, and the overall number of categories and strata are relatively large. As more strata are included, it is important to have enough data to adequately fill those strata to generate a balanced separation. There is no known rule on how many strata are too many, but the number should be considered with the size of incoming data and what demographic variables are of key importance to research questions.20

Table 1

Distribution of all balanced variables in the input, open (∼80%), and sequestered (∼20%) datasets following splitting via stratified sampling. Prevalence values are written as the mean percent (standard deviation) over 2000 independent trials. The label of “not reported” was added to variables with blank entries.

Demographic subcategoryInput dataset countInput dataset prevalenceOpen subset prevalenceSequestered subset prevalence
Age group
[0, 18)741.5%1.5% (0.1%)1.4% (0.3%)
[18, 30)3937.9%7.9% (0.1%)7.9% (0.3%)
[30, 40)52910.6%10.6% (0.1%)10.3% (0.4%)
[40, 50)68713.7%13.7% (0.1%)13.9% (0.4%)
[50, 65)143428.7%28.6% (0.1%)29.1% (0.4%)
[65, 75)90918.2%18.2% (0.1%)18.1% (0.4%)
[75, 85)59711.9%11.9% (0.1%)11.9% (0.3%)
[85, 140)2845.7%5.7% (0.1%)5.5% (0.3%)
Not reported931.9%1.8% (0.1%)1.9% (0.2%)
Race
American Indian or Alaska Native170.3%0.3% (0.0%)0.3% (0.2%)
Asian2945.9%5.9% (0.1%)5.9% (0.4%)
Black or African American138627.7%27.8% (0.1%)27.6% (0.4%)
Native Hawaiian or other Pacific Islander150.3%0.3% (0.0%)0.3% (0.2%)
White256851.4%51.2% (0.1%)51.8% (0.6%)
Not reported55411.1%11.1% (0.1%)10.8% (0.5%)
Other1663.3%3.3% (0.1%)3.3% (0.4%)
Sex at birth
Female253350.7%50.6% (0.1%)50.8% (0.5%)
Male246449.3%49.3% (0.1%)49.1% (0.5%)
Other00.0%0.0% (0.0%)0.0% (0.0%)
Not reported30.1%0.1% (0.0%)0.1% (0.1%)
Ethnicity
Hispanic or Latino49910.0%10.0% (0.1%)9.7% (0.6%)
Not Hispanic or Latino444388.9%88.8% (0.2%)89.2% (0.6%)
Not reported581.2%1.2% (0.1%)1.2% (0.3%)
COVID-19 status
No260252.0%52.0% (0.1%)52.2% (0.5%)
Not reported10.0%0.0% (0.0%)0.0% (0.0%)
Yes239747.9%48.0% (0.1%)47.8% (0.5%)
Image modality
CR204941.0%40.9% (0.2%)41.3% (0.8%)
CT91018.2%18.3% (0.2%)18.0% (0.8%)
DX259651.9%52.0% (0.1%)51.9% (0.5%)
MR270.5%0.5% (0.0%)0.5% (0.2%)

To summarize the performance of our developed sequestration algorithm, we applied the algorithm to the input dataset for 2000 independent trials. In each independent trial, a different random seed was used to initiate the splitting. The mean and standard deviation of each demographic category’s prevalence was calculated over all independent trials and compared to the prevalence in the input dataset and between the two subsets.

To compare the performance of the sequestration algorithm to “naïve” separation of the dataset with an overall 80:20 random drawing, where the assignment to open or sequestered dataset is made randomly without considering demographic variables, we sought to evaluate resulting balance in the two subsets. In this paper, “more balanced” is intended to mean that the relative distributions of demographic categories are more closely matched between resulting open and sequestered subsets. For example, a balanced sample would result in both the open and sequestered subsets having a nearly 50:50 split of male and female patients, matching the initial dataset. Conversely, an unbalanced sample may have 60% male patients (and 40% female patients) in the open subset and 40% male patients (and 60% female patients) in the sequestered subset. We applied the naïve algorithm to the input dataset for 2000 independent trials, reviewed the resulting distribution of cases across demographic categories through visual inspection, and evaluated demographic balance quantitatively by comparing the median of each distribution using the one-tailed Mann–Whitney U test.25 The resulting distributions were first compared by creating histograms of the scaled difference from expectation in each category, calculated according to Eq. (1):

Eq. (1)

Scaled difference from expectation=|(f)NTNOpen|(f)NT.

Here, f represents the fraction placed in the open commons; this value is a scaling factor to scale the total number of patients in a given category to the proportion placed in the open commons. In this work, it is equal to 0.8. NT represents the number of patients in a given category in total from the input dataset; this is the incoming number of patients, to be split, in one category. If, within the initial 5000 total patients there were 500 patients with reported race as Asian, NT would be equal to 500 for the race category of Asian. This is the unscaled expected counts. Similarly, NOpen represents the number of patients in a given category in the open dataset, which would be expected to be 80% of NT in size if the dataset split was exact; this is the resulting number of patients, after being split, in one category. If, within the resulting open subset there were 390 patients with reported race as Asian, NOpen would be equal to 390 for the race category of Asian. This is the observed count. The value of this metric from all independent trials was plotted in histograms for a given bin within a category, e.g., in the race categories, all Asian patients. Distributions from our developed stratified sampling algorithm and from naïve random sampling were compared using the one-tailed Mann–Whitney U test. The Mann–Whitney U test is a nonparametric alternative to the t-test to compare central measures of two distributions, which can be employed without any assumption about the shape of the distribution. In this analysis, we are using a one-tailed Mann–Whitney U test with the alternative hypothesis indicating the median of the distribution of the scaled difference from expectation [Eq. (1)] from stratified sampling was less than that for the naïve random sampling. P-values <0.05 were considered significant. Results were adjusted for multiple comparisons using the Holm–Bonferroni multiple comparisons correction.26 In this correction, the calculated p-value must be less than the chosen significance threshold divided by the rank of said p-value, after ordering from least to greatest where the smallest p-value would have a rank of the total number of comparisons. In this analysis, 29 comparisons were evaluated.

3.

Results

Results obtained from splitting the input dataset using our stratified sampling method over 2000 independent trials are shown in Table 1. The mean and standard deviation of each demographic category’s prevalence was calculated over all independent trials and compared to the prevalence in the input dataset and between the two subsets. For all demographic subcategories, the prevalence in the input dataset was matched in both the open and sequestered subsets within one standard deviation. For the category of image modality, the prevalence represents the percentage of patients with a given image modality available, and since many patients have images from multiple modalities, these percentages will not add to 100%.

Histograms of the scaled difference from expectation over 2000 independent trials for the categories of age and race in the “Open” subset (80%) are shown in Figs. 2 and 3. For most categories analyzed, sequestration by stratified sampling provided lower scaled differences from expectation, in general, than from the naïve randomization, as indicated by the narrower distributions for stratified sampling than naïve randomization. However, for some categories with low prevalence, such as race of American Indian or Native American, stratified sampling showed similar variation from expectation as the naïve randomization.

Fig. 2

Histograms of the scaled difference from expectation for the category of age in the “Open” dataset after 2000 independent trials of dataset splitting using naïve random sampling (blue) and our stratified sampling method (yellow). Here, a narrower histogram indicates better balance, and a wider histogram indicates worse balance in a given subcategory over all independent trials.

JMI_10_6_064501_f002.png

Fig. 3

Histograms of the scaled difference from expectation for the category of race in the “Open” dataset after 2000 independent trials of dataset splitting using naïve random sampling (blue) and our stratified sampling method (yellow). Here, a narrower histogram indicates better balance, and a wider histogram indicates worse balance in a given subcategory over all independent trials.

JMI_10_6_064501_f003.png

Results of the statistical comparison of the histograms of the scaled difference from expectation for all demographic subcategories categories in the “Open” subset using the one-tailed Mann–Whitney U test are shown in Table 2. Similar results were found for the “Sequestered” subset, despite having smaller relative sample size but are omitted for brevity. Statistical results support the qualitative summary that sequestration by stratified sampling provided lower differences from expectation, in general, than did the naïve randomization, with the exception of demographic subcategories with very low prevalence. P-values <0.05 were considered significant. Correction for multiple comparisons using Holm–Bonferroni did not change significance between the differences from expectation for any subcategory except the category of having an ethnicity of not Hispanic or Latino. As a result, for most demographic categories, our developed method of stratified sampling significantly outperformed naïve random sampling, providing more balanced demographic distributions on average.

Table 2

Results of the one-tailed Mann–Whitney U test, comparing distributions of the scaled difference from expectation in the “Open” dataset after 2000 independent trials of dataset splitting using naïve random sampling and using our stratified sampling method. P-values shown in bold were considered significant (p<0.05) and p-values with two asterisks were considered significant after Holm–Bonferroni multiple comparison correction. Here, statistically significant indicates the histogram of the scaled difference from expectation from stratified sampling algorithm was significantly narrower than the histogram from naïve random sampling.

Demographic subcategoryInput dataset countMann–Whitney U test result
Age group
[0, 18)74p < 0.01**
[18, 30)393p < 0.01**
[30, 40)529p < 0.01**
[40, 50)687p < 0.01**
[50, 65)1434p < 0.01**
[65, 75)909p < 0.01**
[75, 85)597p < 0.01**
[85, 140)284p < 0.01**
Not reported93p < 0.01**
Race
American Indian or Alaska Native17p = 0.70
Asian294p < 0.01**
Black or African American1386p < 0.01**
Native Hawaiian or other Pacific Islander15p = 0.29
White2568p < 0.01**
Not reported554p < 0.01**
Other166p < 0.01**
Sex at birth
Female2533p < 0.01**
Male2464p < 0.01**
Other0N/A
Not reported3p = 0.52
Ethnicity
Hispanic or Latino499p < 0.01**
Not Hispanic or Latino4443p = 0.01
Not reported58p = 0.57
COVID-19 status
No2602p < 0.01**
Not reported1p = 0.29
Yes2397p < 0.01**
Image modality
CR2049p < 0.01**
CT910p < 0.01**
DX2596p < 0.01**
MR27p < 0.01**

4.

Discussion

We demonstrated, using our proposed method of multi-dimensional stratified sampling, that splitting an input dataset of 5000 COVID-19 patients into an 80% open dataset and a 20% sequestered dataset based on the variables of age, race, sex at birth, ethnicity, COVID-19 status, and image modality resulted in subsets that exhibited distributions very similar to those of the input dataset and each other. The high degree of similarity in the distributions indicates that the sequestration algorithm operated as expected. Moreover, the distributions of the differences from the expected values for the developed stratified sampling algorithm and naïve randomization indicated that the stratified sampling algorithm, in general, outperformed naïve randomization, providing more balanced distributions of demographics versus the demographic distributions obtained from naïve randomization.

It is important to note that while the analysis of the data focused on individual demographic categories, the developed method of multi-dimensional stratified sampling also controls the n-dimensional joint distribution of demographic categories. Thus, for each unique combination of demographic categories (e.g., 30 to 39 years old, Asian, Not Hispanic, COVID-19 positive, females, etc.), the stratified sampling algorithm separates that unique strata into “Open” or “Sequestered,” preserving the joint distributions. Additional work to evaluate the joint distribution balance of the entire MIDRC data commons is discussed in another publication.27

While the high degree of similarity in the distributions of variables across both subsets is promising, indicating that the proposed sampling method worked as intended, the ultimate goal in constructing a sequestered dataset for algorithm evaluation does not aim for perfect symmetry relative to the data going to the open dataset. Sequestration will provide an ongoing method to monitor and maintain a high level of similarity in the variable distributions, but perturbations in the demographics will also be purposely implemented to assure algorithm generalizability. Balanced distributions of demographic variables between open and sequestered commons allow for establishment and maintenance of a standard “MIDRC demographic” across both commons. This demographic profile is useful in evaluation of the bias and diversity of MIDRC data, which is discussed in another publication.28 Balance of the demographic distributions will also allow for users to sample similar data for training that can be later sampled for testing from the sequestered commons. However, all the data separated into each commons do not represent the subset on which training or testing will be directly performed. It instead represents an entire data commons that will need to be sampled by another method for selection of a cohort for a specific task. When an algorithm is tested using data from the sequestered dataset, test samples will be randomly drawn from the sequestered set according to the distributions related to the task (e.g., clinical question, clinical claim, and intended population), that is, the sequestered set in its entirety will not be used in the test. Furthermore, from the algorithm testing using sequestered data, only summary performance information will be reported back and not case-specific results. Methodology for this specific task-based sampling that is more directly related to performance of individual algorithms will be discussed in a future publication.8

The size of 5000 patients for the example set was selected to be approximately representative of a “typical” incoming data batch internal to MIDRC that would be proportionally split between the two commons. The 5000 patients were selected at random from the existing public data commons and are available for download on Ref. 7. Larger subsets of data would be expected to have relatively better achieved balance with stratification and smaller subsets would be expected to have relatively worse achieved balance with stratification, decreasing until it approaches naïve random sampling. However, this is shown within the current dataset within the smaller and larger strata investigated.

Assessment of machine learning algorithm performance is often achieved through methods such as k-fold cross validation or bootstrapping.29 These methods sample a limited dataset many times to test the algorithm on a variety of sample characteristics. More commonly, sampling for these methods is performed using naïve randomization. As such, naïve random sampling was an intuitive choice for a standard of comparison with the developed method of multi-step stratified sampling. Stratified randomization is an existing process used in separating training and testing datasets, but generally only allows for stratification across a single variable. However, balancing of multiple variables across public and sequestered datasets, from which cases cannot be made known, or used and replaced, is a task not typically considered in machine intelligence applications. Using the presented process, which sequentially steps through each branch until a single multi-variable stratum is obtained, balance across all possible combinations of the selected variables can be controlled. Similar processes are used in the construction of case and control populations in clinical trials, but these processes are typically conducted once, after collection of the entire population. Our process is implemented on each incoming MIDRC data batch, which are received on an ongoing basis. To the knowledge of the authors, this is the first application of multi-step stratified sampling to generate continually growing machine intelligence datasets.15,19,20

Stratification across a large number of variables also necessitates a relatively large initial dataset to be split to achieve its intended effects, i.e., achieving better balance than naïve randomization. This is evidenced in this work in the relatively lower difference in performance between random guessing and stratified sampling in subgroups of low prevalence. Once a subgroup population is too low to split 80:20 evenly, the process of assigning a patient to a specified subgroup is random with probability of 80:20. This is a known limitation for clinical trial stratification schemes.20 Having large data batches, on the order of thousands of patients, enables the stratification across multiple variables. Oakden-Rayner et al. provide an example of why stratification across many variables may be a beneficial approach since fewer variables are “hidden” and balance across many subgroups is monitored within MIDRC.17 The developed sequestration algorithm will provide an open-source tool for avoiding hidden stratification, through the use of a type of variable population or schema completion. Other, more sophisticated methods of sampling, such as minimization, could also be used for comparison to stratified sampling and may indeed outperform stratified sampling in certain circumstances, particularly when the number of patients relative to the number of strata is low. However, using a method of multi-step stratified sampling is intuitive for users to understand and may be simply modified to accommodate new stratification variables, increasing the general utility of an open-source tool.

With the explosion of machine intelligence algorithm development, validation of algorithm performance and generalizability has become increasingly important. As a result, many academic journals now encourage authors to make datasets and algorithms public when publishing. The creation of a large, common sequestered dataset that can be sampled for specific task-based algorithm performance evaluations could provide a new gold standard in the field of machine intelligence. Further, a sequestered database for algorithm testing could allow for expedited clinical implementation of algorithms developed for medical decision making if accepted by regulating bodies. This work outlines the process by which such a database has been developed for use in a multi-institutional data commons.

It is also important to discuss a few limitations of this study. The ability of stratified sampling to achieve a much higher degree of balance than simple randomization is highly dependent on the incoming dataset size and prevalence of a given demographic subcategory. Subgroups that are very rare in incoming data batches are likely to be rare across the entire data commons, and the ability to sample rare cases will always be limited by their availability in each data commons. Clinical trials that use similar stratification methods also note this be to be a limitation.20 While the currently developed algorithm samples based on the demographic categories of race, age, ethnicity, sex at birth, COVID-19 status, image modality, and clinical site, the algorithm will be adapted in the future to additional needs, including body site or patient location information. The authors also acknowledge that specific labels of race, ethnicity, or sex at birth may not adequately describe all populations or provide a clear correlate to genetic ancestry, which emphasizes the importance of oversight and monitoring of AI algorithms for equity in their applications to healthcare. As additional information and understanding of case descriptors are acquired, the sequestration algorithm will be modified to accommodate. In addition, while ensuring similar distributions of demographic characteristics may help to reduce potential algorithm bias, certain biases will still persist and must be acknowledged.

5.

Conclusion

The creation of a sequestered dataset within MIDRC will allow for algorithm performance evaluation using large, verifiable multi-institutional data. This work outlines the process by which the sequestered dataset will be sampled from incoming data batches to ensure balanced variables in both the public and sequestered data. Using an example input dataset, both subsets were found to match the distribution of variables in the original data with a high degree of similarity.

Disclosures

MLG is a stockholder in R2 technology/Hologic and QView, receives royalties from Hologic and Samsung, and was a cofounder in Quantitative Insights (now Qlarity Imaging). It is the University of Chicago Conflict of Interest Policy that investigators disclose publicly actual or potential significant financial interest that would reasonably appear to be directly and significantly affected by the research activities. The mention of commercial products, their sources, or their use in connection with material reported herein is not to be construed as either an actual or implied endorsement of such products by the Department of Health and Human Services. KD receives royalties from Hologic. UCLA has a Master Research Agreement with Siemens Healthineers; MMG is a grant recipient of Siemens Healthineers. GJK is a research consultant of MedQIA LLC.

Code and Data Availability

The software used in this paper, including MATLAB and Python translation of the algorithm, are publicly available on https://github.com/MIDRC/Stratified_Sampling/. The data used for this paper are maintained publicly at data.MIDRC.org. Additional information on the selection of the 5000 patients used in this study will be available on GitHub.

Acknowledgments

The authors would like to thank Dr. Gene Pennello of the FDA for his expertise and assistance and the team members at Gen3 Data Commons for their help in facilitating this project. This research was funded through The Medical Imaging Data Resource Center (MIDRC), which is funded by the National Institute of Biomedical Imaging and Bioengineering (NIBIB) of the National Institutes of Health under Contract Nos. 75N92020C00008 and 75N92020C00021. This paper expands upon the previously published SPIE conference proceedings paper by Baughan et al.30

References

1. 

F. Shi et al., “Review of artificial intelligence techniques in imaging data acquisition, segmentation, and diagnosis for COVID-19,” IEEE Rev. Biomed. Eng, 14 4 –15 https://doi.org/10.1109/RBME.2020.2987975 (2021). Google Scholar

2. 

J. D. Fuhrman et al., “A review of explainable and interpretable AI with applications in COVID-19 imaging,” Med. Phys., 49 (1), 1 –14 https://doi.org/10.1002/mp.15359 MPHYA6 0094-2405 (2022). Google Scholar

3. 

Q. Hu, K. Drukker and M. L. Giger, “Role of standard and soft tissue chest radiography images in deep-learning-based early diagnosis of COVID-19,” J. Med. Imaging, 8 (Suppl. 1), 014503 https://doi.org/10.1117/1.JMI.8.S1.014503 JMEIET 0920-5497 (2021). Google Scholar

4. 

M. Roberts et al., “Common pitfalls and recommendations for using machine learning to detect and prognosticate for COVID-19 using chest radiographs and CT scans,” Nat. Mach. Intell., 3 (3), 199 –217 https://doi.org/10.1038/s42256-021-00307-0 (2021). Google Scholar

5. 

I. El Naqa et al., “Lessons learned in transitioning to AI in the medical imaging of COVID-19,” J. Med. Imaging, 8 (Suppl. 1), 010902 https://doi.org/10.1117/1.JMI.8.S1.010902 JMEIET 0920-5497 (2021). Google Scholar

6. 

H. Estiri et al., “An objective framework for evaluating unrecognized bias in medical AI models predicting COVID-19 outcomes,” J. Am. Med. Inf. Assoc., 29 (8), 1334 –1341 https://doi.org/10.1093/jamia/ocac070 (2022). Google Scholar

7. 

“MIDRC : medical imaging and data resource center,” https://www.midrc.org/ (). Google Scholar

8. 

N. Baughan et al., “Task-based sampling of the MIDRC sequestered data commons for algorithm performance evaluation,” Med. Phys., 49 (6), E257 –E258 (2022). Google Scholar

9. 

K. Drukker et al., “Toward fairness in artificial intelligence for medical image analysis: identification and mitigation of potential biases in the roadmap from data collection to model deployment,” J. Med. Imaging, 10 (6), 061104 https://doi.org/10.1117/1.JMI.10.6.061104 JMEIET 0920-5497 (2023). Google Scholar

10. 

N. Norori et al., “Addressing bias in big data and AI for health care: a call for open science,” Patterns, 2 (10), 100347 https://doi.org/10.1016/j.patter.2021.100347 (2021). Google Scholar

11. 

A. J. Larrazabal et al., “Gender imbalance in medical imaging datasets produces biased classifiers for computer-aided diagnosis,” Proc. Natl. Acad. Sci. U. S. A., 117 (23), 12592 –12594 https://doi.org/10.1073/pnas.1919012117 (2020). Google Scholar

12. 

J. W. Gichoya et al., “AI recognition of patient race in medical imaging: a modelling study,” Lancet Digital Health, 4 (6), e406 –e414 https://doi.org/10.1016/S2589-7500(22)00063-2 (2022). Google Scholar

13. 

J. Adleberg et al., “Predicting patient demographics from chest radiographs with deep learning,” J. Am. Coll. Radiol., 19 (10), 1151 –1161 https://doi.org/10.1016/j.jacr.2022.06.008 (2022). Google Scholar

14. 

B. Glocker et al., “Algorithmic encoding of protected characteristics in chest X-ray disease detection models,” eBioMedicine, 89 104467 https://doi.org/10.1016/j.ebiom.2023.104467 (2023). Google Scholar

15. 

R. Schwartz et al., Towards a Standard for Identifying and Managing Bias in Artificial Intelligence, National Institute of Standards and Technology( (2022). Google Scholar

16. 

V. Azimi and M. A. Zaydman, “Optimizing equity: working towards fair machine learning algorithms in laboratory medicine,” J. Appl. Lab. Med., 8 (1), 113 –128 https://doi.org/10.1093/jalm/jfac085 (2023). Google Scholar

17. 

L. Oakden-Rayner et al., “Hidden stratification causes clinically meaningful failures in machine learning for medical imaging,” in Proc. ACM Conf. Health Inference Learn., 151 –159 (2020). Google Scholar

18. 

L. Seyyed-Kalantari et al., “Underdiagnosis bias of artificial intelligence algorithms applied to chest radiographs in under-served patient populations,” Nat. Med., 27 (12), 2176 –2182 https://doi.org/10.1038/s41591-021-01595-0 1078-8956 (2021). Google Scholar

19. 

O. Sverdlov, “Modern adaptive randomized clinical trials: statistical and practical aspects,” Int. Stat. Rev., 84 (2), 321 –322 https://doi.org/10.1111/insr.12188 (2016). Google Scholar

20. 

Y. Lin, M. Zhu and Z. Su, “The pursuit of balance: an overview of covariate-adaptive randomization techniques in clinical trials,” Contemp. Clin. Trials, 45 21 –25 https://doi.org/10.1016/j.cct.2015.07.011 (2015). Google Scholar

21. 

S.-C. Chow, M. Chang and A. Pong, “Statistical consideration of adaptive methods in clinical development,” J. Biopharm. Stat., 15 (4), 575 –591 https://doi.org/10.1081/BIP-200062277 JBSTEL 1054-3406 (2005). Google Scholar

22. 

Centers for Disease Control and Prevention, “COVID-19 death data and resources: weekly updates by select demographic and geographic characteristics,” (2021). https://www.cdc.gov/nchs/nvss/vsrr/covid_weekly/index.htm#SexAndAge Google Scholar

23. 

National Institutes of Health (NIH), “Consideration of sex as a biological variable in NIH-funded research: NOT-OD-15-102,” (01 December 2021). https://grants.nih.gov/grants/guide/notice-files/not-od-15-102.html Google Scholar

24. 

National Institutes of Health (NIH), “Racial and ethnic categories and definitions for NIH diversity programs and for other reporting purposes: NOT-OD-15-089,” (01 December 2021). https://grants.nih.gov/grants/guide/notice-files/NOT-OD-15-089.html Google Scholar

25. 

H. B. Mann and D. R. Whitney, “On a test of whether one of two random variables is stochastically larger than the other,” Ann. Math. Stat., 18 50 –60 https://doi.org/10.1214/aoms/1177730491 AASTAD 0003-4851 (1947). Google Scholar

26. 

S. Holm, “A simple sequentially rejective multiple test procedure,” Scand. J. Stat., 6 (2), 65 –70 (1979). Google Scholar

27. 

N. Baughan et al., “Sequestration methodology in practice through evaluation of joint demographic distributions of 54,185 patients in the Medical Imaging and Data Resource Center (MIDRC) data commons,” Proc. SPIE, 12469 1246909 https://doi.org/10.1117/12.2654247 (2023). Google Scholar

28. 

H. M. Whitney et al., “Longitudinal assessment of demographic representativeness in the Medical Imaging and Data Resource Center open data commons,” J. Med. Imag., 10 (6), 061105 https://doi.org/10.1117/1.JMI.10.6.061105 (2023). Google Scholar

29. 

M. Budka and B. Gabrys, “Density-preserving sampling: robust and efficient alternative to cross-validation for error estimation,” IEEE Trans. Neural Networks Learn. Syst., 24 (1), 22 –34 https://doi.org/10.1109/TNNLS.2012.2222925 (2013). Google Scholar

30. 

N. Baughan et al., “Sequestration of imaging studies in MIDRC: a multi-institutional data commons,” Proc. SPIE, 12035 120350F https://doi.org/10.1117/12.2610239 PSISDG 0277-786X (2022). Google Scholar

Biography

Natalie Baughan is a recent PhD graduate from the University of Chicago Graduate Program in Medical Physics and a medical physics resident at the University of North Carolina at Chapel Hill. After receiving her BS degree from the University of Michigan in 2019, her research has focused on breast cancer risk assessment in mammography and statistical methods for AI. She is a student member of both the American Association of Physicists in Medicine and SPIE.

Heather M. Whitney, PhD, is a research assistant professor of radiology at the University of Chicago. Her experience in quantitative medical imaging has ranged from polymer gel dosimetry to radiation damping in nuclear magnetic resonance to radiomics. She is interested in investigating the effects of the physical basis of imaging on radiomics, the repeatability and robustness of radiomics, the development of methods for task-based distribution, and bias and diversity of medical imaging datasets. She is a member of SPIE.

Karen Drukker, PhD, is a research associate professor at the University of Chicago. Her research interests include AI/machine learning in medical imaging and algorithm performance assessment. She is a fellow of SPIE and AAPM.

Berkman Sahiner, PhD, is a senior biomedical research scientist with the Division of Imaging, Diagnostics and Software Reliability, Center for Devices and Radiological Health, U.S. Food and Drug Administration. He has a PhD in electrical engineering and computer science from the University of Michigan, Ann Arbor, USA. His research is focused on the evaluation of medical imaging and computer-assisted diagnosis devices, including devices that incorporate machine learning and artificial intelligence. He is a fellow of SPIE and AIMBE.

Michael McNitt-Gray, PhD, is a professor in the Department of Radiological Sciences at the David Geffen School of Medicine at UCLA. His research interests include computed tomography (CT) imaging and the use of AI/machine learning algorithms in the detection, diagnosis, and evaluation of disease. He is a fellow of AAPM and ACR.

Maryellen L. Giger, PhD, is the A. N. Pritzker Distinguished Service Professor of Radiology, the Committee on Medical Physics, and the College at the University of Chicago. Her research includes computer-aided diagnosis/machine learning/deep learning in medical imaging on cancer for risk assessment, diagnosis, prognosis, and therapeutic response, as well as the analysis of COVID-19 on CT and chest radiographs. She is the contact PI on the NIBIB-funded Medical Imaging and Data Resource Center (MIDRC), which, as August 2022, has ingested more than 150,000 medical imaging studies for use by AI investigators. She is a fellow of SPIE as well as a past president of SPIE.

Biographies of the other authors are not available.

CC BY: © The Authors. Published by SPIE under a Creative Commons Attribution 4.0 International License. Distribution or reproduction of this work in whole or in part requires full attribution of the original publication, including its DOI.
Natalie Baughan, Heather M. Whitney, Karen Drukker, Berkman Sahiner, Tingting Hu, Hyun J. Grace Kim, Michael F. McNitt-Gray, Kyle J. Myers, and Maryellen L. Giger "Sequestration of imaging studies in MIDRC: stratified sampling to balance demographic characteristics of patients in a multi-institutional data commons," Journal of Medical Imaging 10(6), 064501 (16 November 2023). https://doi.org/10.1117/1.JMI.10.6.064501
Received: 25 January 2023; Accepted: 25 October 2023; Published: 16 November 2023
Advertisement
Advertisement
KEYWORDS
Algorithm development

COVID 19

Medical imaging

Evolutionary algorithms

Histograms

Artificial intelligence

Databases

Back to Top