1.IntroductionBreast cancer is one of the leading causes of death in women. Magnetic resonance imaging (MRI) has had an integral role in improving breast cancer diagnoses and potentially reducing biopsies, in tumor staging, and in monitoring treatment response.1–3 Dynamic contrast-enhanced (DCE) MRI involves the acquisition of time series images after injection of a contrast agent; typically, one precontrast timepoint and multiple postcontrast timepoint images are acquired in 60 to 90 s intervals.3,4 Artificial intelligence methods, including detection, diagnosis, or segmentation tasks, have been developed to support radiologists in their interpretation decision-making process. The quality of segmentation needed, e.g., an approximate outline or a detailed contour, depends on the subsequent task.5 Precise lesion segmentation is required to extract relevant tumor features to be used in the classification components in computer-aided diagnosis (CADx) systems.4,6 A well-established and clinically used algorithm for breast lesion segmentation on DCE-MRI is a technique based on the fuzzy c-means (FCM) clustering algorithm, which analyzes the contrast uptake over time and yields volumetric segmentations.4 An alternative segmentation method is U-Net, a deep learning convolutional neural network, which produces segmentations based on a single timepoint.7 Without the requirement of using information from an entire dynamic time series, the U-Net has the potential to produce accurate segmentations from a variety of imaging sequences, including regular and abbreviated DCE-MRI acquisitions.8 The U-Net architecture can be designed to accept either 2D image slices or 3D image volumes.9 Several studies have been conducted to assess the performance of using 2D and 3D U-Nets for lesion segmentation from breast DCE-MRI.10–14 These methods have been developed using different DCE timepoints, datasets sizes, or unique ensembles of modified U-Nets. The evaluation criteria for these studies have been reported across a wide range, demonstrating the complexity of our task.10–14 The U-Nets used in our study were trained to segment masses and nonmass enhancing lesions from either first or second postcontrast subtraction images, i.e., subtraction images between the first or second postcontrast image and the precontrast image. Using the 2D U-Net, quasi-3D lesion segmentations can be obtained by stacking slice-by-slice segmentations; however, the lack of vertical (out-of-slice) continuity obtained by this “quasi-3D” U-Net may be a potential source of error that a fully 3D U-Net avoids. In this study, we investigated the potential of using U-Nets in breast lesion segmentation on DCE-MRI by comparing the performances of 2D and 3D U-Nets relative to FCM. 2.MethodsThe viability of using U-Nets in breast lesion segmentation on DCE-MRI was assessed by comparing the performances of 2D and 3D U-Nets in four evaluations. First, in comparison A, quasi-3D and 3D U-Nets were compared to FCM, which served as a surrogate reference standard.15 Second, in comparison B, the 2D U-Net, 3D U-Net, and FCM segmentations were compared to 2D radiologist delineations on lesion center slices for a subset of 71 lesions.15 Next, in comparison C, segmentations from first postcontrast subtraction images were compared to second postcontrast subtraction images for quasi-3D and 3D U-Nets. Finally, in comparison D, the segmentation performance of each method was evaluated for mass versus nonmass enhancing lesions. 2.1.DatasetThe dataset consisted of DCE-MRIs of 994 unique breast lesions (724 malignant and 270 benign) from 689 patients aged 23 to 89 years. The deidentified data were retrospectively collected at the University of Chicago over a span of 8 years (from 2005 to 2013) under Health Insurance Portability and Accountability Act-compliant Institutional Review Board protocols. Routine bilateral breast MRIs were acquired using a Philips Achieva scanner with either 1.5 T () or 3 T () magnet strength. The breast DCE-MRI protocol included a fat-saturated 3D weighted spoiled gradient-echo sequence that was used to acquire precontrast and postcontrast images with a temporal resolution of 60 to 75 s (TE = 2.2 to 2.8 ms, TR = 4.5 to 7.5 ms, flip angle = 10 deg to 20 deg, in-plane resolution = 0.5 to 1.0 mm, FOV = 28.0 to 44.1 cm, matrix = 320 to to 525, slice thickness = 1 to 3.5 mm, and interslice gap = 0.8 to 2.5 mm). Table 1 contains the clinical characteristics of the data obtained from pathology and radiology reports, including pathological truth (benign or malignant) and lesion type (mass or nonmass enhancement). A subset of 71 lesions was manually selected for radiologist delineations so that the distribution of pathological truth and lesion type within this subset was similar to the distribution in the overall group (Table 1). Table 2 presents size distributions of the lesions. Table 1Summary of the DCE-MRI dataset by lesion type. Lesions were categorized by pathological truth and enhancement type. Lesions that were not marked as either mass or nonmass enhancing were labeled “unknown.”
Table 2Summary of the DCE-MRI dataset by lesion size. Lesions were categorized by effective diameter (mm), defined by 2*√(A/π), where A is the area of the lesion in the center slice of the FCM segmentation in mm2.
2.2.Establishment of Reference Standards and PreprocessingEach lesion had previously been segmented using a well-established, in-house, automated 3D FCM approach that yielded, as a surrogate “reference standard,” a 3D binary lesion segmentation.4 FCM segmentation was performed within a region defined by a human operator’s selection of a rectangular bounding box about the lesion in a middle slice along with an indication of the first and last slices in which the lesion appeared.4 The bounding-box volume of interest (VOI) for the FCM segmentation of each lesion was also used as input for subsequent U-Net segmentations of postcontrast subtraction images. Second postcontrast subtraction images were primarily used as inputs for the U-Net; however, first postcontrast subtraction images were introduced for evaluation in comparisons C and D of this study. In addition, an expert radiologist (7 years of experience in breast imaging) manually delineated the lesion within the center slice of the second postcontrast subtraction VOI for the subset of 71 lesions. Here the radiologist segmentations were used as the “reference standard” for comparison B of this study. Since radiologist segmentations were only available for a limited set of center slices and FCM segmentations are used in an FDA-approved clinical breast MRI workstation,3 FCM segmentations served as a reasonable surrogate reference standard to train the U-Net architectures. 2.3.U-Net ArchitecturesTwo different U-Net architectures were evaluated in this study. The first was a 2D U-Net.7 We found that the top and bottom slices of lesions were most difficult to segment, so those two slices were excluded from each lesion in training (though they remained in the test set lesions). The image slices of each lesion’s VOI were resized, by interpolation with a preserved pixel value range, to prior to input into the 2D U-Net. The probability map outputs, with values ranging from 0 to 1, were converted to binary segmentation images based on a threshold of 0.25. The 2D U-Net only processes one image slice at a time, so “quasi-3D” lesion segmentations were produced by stacking the 2D slice-by-slice segmentations obtained by the 2D U-Net to form a 3D volume. Hence, in this paper, “quasi-3D U-Net” refers to the volumetric segmentation produced by the 2D U-Net architecture. The second architecture evaluated in this study was a 3D U-Net.9 This network is similar to the structure of the 2D U-Net, but it is modified with the added third dimension. Prior to input into the 3D U-Net, the lesion VOIs were resized, by interpolation with a preserved pixel value range, to ( is the number of slices in the lesion). The network produced probability map outputs, with values ranging from 0 to 1, which were converted to binary segmentation volumes based on a threshold of 0.23. The threshold for the binary conversion was selected from a range of values between 0.14 and 0.30 to produce the greatest mean Dice similarity coefficient (DSC) calculated from the resulting segmentations during training, hence the slight difference in threshold used for 2D U-Net and 3D U-Net. 2.4.Training and Statistical Analysis of Segmentation PerformancesFivefold cross-validation by lesion ( lesions) was conducted to train and evaluate the U-Net models. The folds were partitioned such that each fold contained a similar distribution based on pathological truth (malignant or benign), lesion enhancement type, and lesion size. Additionally, since adjacent slices within the same lesion VOI often are very similar in appearance, all slices belonging to a given lesion were always allocated to the same fold. Training and test folds were allocated by lesion, i.e., not by slice or patient. The base U-Net models were trained using the Adam optimizer and a binary cross-entropy loss function; training was allowed to run for up to 200 epochs. DSC and Hausdorff distance (HD) were used to evaluate the performances of the different segmentation methods relative to the specific reference standard.16,17 DSC is a measure of how well the areas of the two regions overlap, and HD is a measure of how well the margins of the two regions agree. Note that, throughout, better segmentation performance is indicated by higher DSCs and lower HDs. Predictions from the quasi-3D and 3D U-Nets were resized to their original lesion VOI dimensions before DSCs and HDs were calculated between the predictions and the reference standards. HDs were calculated for each slice, and the median HD for each lesion was used for 3D performance comparisons. To assess statistical significance of difference in performance, the Wilcoxon signed-rank test was used for matched cases in comparisons A, B, and C, and the Mann–Whitney U-test was used in comparison D due the analysis of unmatched cases.16–19 The Bonferroni correction was used to correct -values for multiple comparisons in comparisons B, C, and D.20 2.5.Comparison A: Comparing Quasi-3D U-Net to 3D U-Net Using FCM as the Surrogate Reference StandardThe volumetric segmentations from quasi-3D and 3D U-Nets were compared and as previously noted, FCM segmentations served as the surrogate reference standard for the 994 lesions (Table 1). The Wilcoxon signed-rank test was used to assess statistically significant differences between quasi-3D and 3D U-Net segmentation performances (Fig. 1). 2.6.Comparison B: Comparing FCM, Quasi-3D U-Net, and 3D U-Net Using Radiologist-Delineations as the Reference StandardNext, FCM, quasi-3D U-Net, and 3D U-Net center slice segmentations were compared using the radiologist references available for the subset of 71 lesions. For each of the three segmentation methods, DSCs and HDs were calculated on the center slice with respect to the radiologist reference. Statistically significant differences between quasi-3D U-Net, 3D U-Net, and FCM segmentations were assessed using the Wilcoxon signed-rank test including a Bonferroni correction (Fig. 2). 2.7.Comparison C: Comparing Segmentation Across Postcontrast Timepoints (First Versus Second Postcontrast)The segmentations obtained in comparison A using second postcontrast subtraction images as input were compared to those using the first postcontrast subtraction images for the quasi-3D and 3D U-Nets. Wilcoxon signed-rank tests were used to assess statistical significance between the results after a Bonferroni correction. 2.8.Comparison D: Comparing Segmentation Across Lesion Enhancement Types (Mass Versus Nonmass Enhancement)Finally, the segmentation performances on mass and nonmass enhancing lesions were compared. The segmentations resulting from the first and second postcontrast subtraction inputs to the quasi-3D and 3D U-Nets evaluated in comparison C were compared based on lesion enhancement type. For each comparison, a Mann–Whitney U-test including a Bonferroni correction for statistical significance was used to compare the segmentation performances of the set of mass lesions to the set of nonmass lesions. 3.Results3.1.Comparison A: Comparing Quasi-3D U-Net to 3D U-Net Using FCM as the Surrogate Reference StandardSegmentation performance was assessed by comparing the medians of DSC and HD (Table 3). Note that better segmentation performance is indicated by higher DSCs and lower HDs. Of the 994 lesions in the dataset, the 3D U-Net failed to segment 6 lesions (from 3 unique patients) that were in effective diameter and had an unknown enhancement type. Without prediction volumes available to compare to the reference standard, DSC was essentially zero and it was impossible to calculate HDs, therefore, these lesions were excluded from HD statistical comparisons for the 3D U-Net. The results of the Wilcoxon signed-rank test show that the quasi-3D U-Net statistically significantly outperformed the 3D U-Net in terms of DSC () and HD () for lesion segmentation from second postcontrast subtraction VOIs. Table 3Comparison A: summary statistics of the performance metrics of quasi-3D and 3D U-Nets as compared to FCM reference standards for volume segmentation. U-Nets were trained and tested using fivefold cross validation by lesion. Minimum, maximum, and median values of DSC and HD metrics of all cases are shown. Parenthetical values represent 95% confidence intervals.
3.2.Comparison B: Comparing FCM, Quasi-3D U-Net, and 3D U-Net Using Radiologist-Delineations as the Reference StandardBased on the segmentation results for the subset of 71 lesions, we found that the center slices from each lesion segmentation produced by FCM, quasi-3D U-Net, and 3D U-Net had good agreement with the radiologist-segmented reference standard (Table 4). The statistical comparisons of performance between each segmentation method’s agreement with the reference standard is shown in Table 5. The results indicate that quasi-3D U-Net outperformed both 3D U-Net and FCM for lesion segmentation on second postcontrast subtraction center slices. Table 4Comparison B: summary statistics of the performance metrics of FCM, quasi-3D U-Net, and 3D U-Net, as compared to radiologist reference standard for center slice segmentation. U-Nets were trained and tested using fivefold cross validation by lesion. Minimum, maximum, and median DSC and HD metrics of all cases are shown. Parenthetical values represent 95% confidence intervals (N=71).
Note: DSC, Dice similarity coefficient and HD, Hausdorff distance. Table 5Comparison B: statistical comparisons between the median performance metrics in Table 4 from FCM, quasi-3D U-Net, and 3D U-Net center slice predictions using radiologist-delineations as the reference standard. U-Nets were trained and tested using fivefold cross validation by lesion. Raw, uncorrected p-values from the Wilcoxon signed-rank test are reported; statistical significance was assessed after correcting for three comparisons (N=71).
Note: DSC, Dice similarity coefficient and HD, Hausdorff distance. We observed improved U-Net segmentation agreement with the radiologist reference as lesion size increased, and Fig. 3 shows how quasi-3D U-Net yielded greater DSC values than 3D U-Net, relative to radiologist delineations, across all lesion sizes. 3.3.Comparison C: Comparing Segmentation Across Postcontrast Timepoints (First Versus Second Postcontrast)An example of the segmentations produced by the 2D U-Net, 3D U-Net, FCM, and radiologist for a mass and nonmass enhancing lesion is shown in Fig. 4. In the second postcontrast subtraction images, more lesion enhancement is provided to the U-Net, which as expected, tended to result in segmentations that more closely resembled FCM than segmentations from the first postcontrast subtraction segmentation inputs. Also as expected, the radiologist delineations acquired on the central slice of the second postcontrast subtraction image tended to resemble the center slice of the second postcontrast subtraction segmentation from the 2D U-Net. The performance metrics calculated for the segmentations produced by the U-Nets from first and second postcontrast subtraction inputs are included in Table 6. The statistical comparisons between the resulting DSC and HD metrics for each method are shown in Tables 7 and 8. The results show statistically significantly greater performance from the second postcontrast subtraction inputs than from the first postcontrast subtraction inputs using the quasi-3D and 3D U-Nets, except in the case of nonmass enhancing lesions using the 3D U-Net. The results from both the first and second postcontrast subtraction inputs support the results found in comparisons A and B. The quasi-3D U-Net statistically significantly outperformed the 3D U-Net for the combined lesion types based on DSC, however HD failed to show statistically significant differences between quasi-3D and 3D U-Net for the first postcontrast subtraction input. Table 6Comparisons C and D: summary statistics of the performance metrics of quasi-3D U-Net and 3D U-Net, as compared to FCM surrogate reference standard. U-Nets were trained and tested using fivefold cross validation by lesion (Nmass=687 and Nnonmass=224).
Note: DSC, Dice similarity coefficient and HD, Hausdorff distance. Table 7Comparisons C and D: statistical results for comparisons between input image type and lesion type using the Dice similarity coefficients of U-Net segmentations against FCM reference standard. U-Nets were trained and tested using fivefold cross validation by lesion. Raw, uncorrected p-values from the Wilcoxon signed-rank or Mann–Whitney U-test are reported. However, statistical significance was assessed after including a Bonferroni correction for four comparisons per sample (Nmass=687 and Nnonmass=224).
Table 8Comparisons C and D: statistical results for comparisons between input image type and lesion type using the Hausdorff distance of U-Net segmentations against FCM reference standard. U-Nets were trained and tested using fivefold cross validation by lesion. Raw, uncorrected p-values from the Wilcoxon signed-rank or Mann–Whitney U-test are reported. However, statistical significance was assessed after including a Bonferroni correction for four comparisons per sample (Nmass=687 and Nnonmass=224).
3.4.Comparison D: Comparing Segmentation Across Lesion Enhancement Types (Mass Versus Nonmass Enhancement)The results in Tables 6Table 7–8 demonstrate that, relative to the FCM reference standard, mass lesion segmentation statistically significantly outperformed nonmass enhancing lesion segmentation using first and second postcontrast subtraction image inputs to both the quasi-3D and 3D U-Nets. For nonmass enhancing lesions, quasi-3D U-Net always statistically significantly outperformed the 3D U-Net (as in comparisons A and B). For mass lesions, the DSC results indicate that quasi-3D U-Net statistically significantly outperformed the 3D U-Net (as in comparison B), however, the HD results from the first postcontrast subtraction inputs showed that the 3D U-Net statistically significantly outperformed quasi-3D U-Net. 4.DiscussionA crucial component of artificial intelligence systems is proper segmentation of lesions and other breast regions before subsequent extraction of quantitative values for clinically significant quantities. This study explored the performance of volumetric segmentations obtained with a 2D U-Net (quasi-3D U-Net) and a 3D U-Net. Segmentation performance was assessed against a well-established FCM method, which served as a surrogate reference standard, or radiologist reference segmentation. There were several limitations of this study. First, the segmentation performances were evaluated within bounding-box VOIs; the inputs to the U-Net were based on the FCM volume dimensions. Future investigations could be focused on identifying lesions from the whole breast, rather than from a predefined region of interest. However, use of the bounding-box VOIs does mimic clinical practice where a radiologist may roughly indicate the region about a lesion as input to automatic characterization and CADx. Also there were a limited number of radiologist segmentations available, each acquired for the second postcontrast subtraction center slices ( lesions); this could have influenced the results of the comparisons performed in comparison B. Additionally, there were no radiologist volume segmentations available for full lesion volumes, so FCM segmentations were used as surrogate reference standards. Finally, the 3D U-Net architecture may not be considered fully 3D since many lesions had too few slices to properly pool in the axial dimension. Future work may include an investigation of U-Net performance for breast lesion segmentation by exploring segmentation from abbreviated DCE-MRI sequences. Also, U-Nets may be trained with attention gating, which could potentially improve segmentation performance by focusing the network on the lesions and drawing attention away from the background tissue. This study found that there were statistically significant differences in performance between U-Net and FCM segmentation methods, relative to each other and to a radiologist reference segmentation. In the task of segmenting breast lesions from second postcontrast subtraction DCE-MRI VOIs, the quasi-3D U-Net statistically significantly outperformed the 3D U-Net in segmenting volumes (). Additionally, the comparison between center slices from FCM, quasi-3D U-Net, and 3D U-Net relative to the radiologist reference suggested that 2D U-Net outperforms FCM and 3D U-Net (). Although the vertical (out-of-slice) context was an assumed advantage for the fully 3D U-Net, our results suggest that using the quasi-3D U-Net, which performs a series of 2D convolutions, max pooling, and upsampling operations, can accurately capture the lesion context and enable more precise localization of lesion pixels on a slice-by-slice basis. Another advantage of using the series of 2D convolutions, over 3D ones, is that less training is required due to the reduced complexity of the 2D net. Relative to FCM volumes, U-Net segmentations of second postcontrast subtraction inputs were statistically significantly greater than first postcontrast subtraction inputs, and segmentation of mass lesions statistically significantly outperformed nonmass lesion segmentation. Although improved segmentation using second postcontrast subtraction inputs were found, the 2D U-Net statistically significantly outperformed the 3D U-Net for the first postcontrast subtraction inputs; this could provide a potential benefit to abbreviated MRI applications. The results of this study suggest that using a 2D U-Net to yield quasi-3D U-Net segmentation of breast lesions from postcontrast subtraction DCE-MRIs is feasible and thus could be an effective alternative to more complex segmentation techniques. Ultimately, this work has the potential to encourage future incorporation of the quasi-3D U-Net method into artificial intelligence algorithms designed to improve the efficiency and efficacy of clinical workflows that include breast DCE-MRI. DisclosuresM.L.G. is a stockholder in R2 technology/Hologic and QView, receives royalties from multiple medical imaging companies via licenses through the Polsky Center for Entrepreneurship and Innovation at the University of Chicago and was a cofounder in Quantitative Insights. K.D. receives royalties from Hologic. L.D., R.B., J.F., Q.H., A.E., and D.S. declare no competing interest. It is the University of Chicago Conflict of Interest Policy that investigators disclose publicly actual or potential significant financial interest that would reasonably appear to be directly and significantly affected by the research activities. The corresponding authors had full access to all the data in the study and had final responsibility for the decision to submit for publication. Code and Data AvailabilityThe data used for this manuscript, including DCE-MRIs and ROIs, are not publicly available due to patient privacy and data sharing agreements. AcknowledgmentsThe authors would like to acknowledge the support received from fellow members of the Giger Lab, Department of Radiology, The University of Chicago, Chicago, Illinois, United States. This work was funded in part by the C3. AI Digital Transformation Institute, the University of Chicago Department of Radiology Pilot Fund, and National Institute of Biomedical Imaging and Bioengineering (NIBIB) COVID-19 (Contract No. 75N92020D00021). It was also supported by the NIBIB of the National Institutes of Health (NIH) (Grant No. T32 EB002103). Partial funding for this work was also provided by the NIH Shared Instrument Grant (Grant No. S10-OD025081) and NIH award (Grant No. S10-RR021039). ReferencesD. Saslow et al.,
“American cancer society guidelines for breast screening with MRI as an adjunct to mammography,”
CA Cancer J. Clin., 57
(2), 75
–89 https://doi.org/10.3322/canjclin.57.2.75 CAMCAM 0007-9235
(2007).
Google Scholar
R. M. Mann, N. Cho and L. Moy,
“Breast MRI: state of the art,”
Radiology, 292
(3), 520
–536 https://doi.org/10.1148/radiol.2019182947 RADLAX 0033-8419
(2019).
Google Scholar
Y. Jiang, A. V. Edwards and G. M. Newstead,
“Artificial intelligence applied to breast MRI for improved diagnosis,”
Radiology, 298
(1), 38
–46 https://doi.org/10.1148/radiol.2020200292 RADLAX 0033-8419
(2021).
Google Scholar
W. Chen, M. L. Giger and U. Bick,
“A fuzzy c-means (FCM)-based approach for computerized segmentation of breast lesions in dynamic contrast-enhanced MR images,”
Acad. Radiol., 13
(1), 63
–72 https://doi.org/10.1016/j.acra.2005.08.035
(2006).
Google Scholar
M. L. Giger, N. Karssemeijer and J. A. Schnabel,
“Breast image analysis for risk assessment, detection, diagnosis, and treatment of cancer,”
Annu. Rev. Biomed. Eng., 15 327
–357 https://doi.org/10.1146/annurev-bioeng-071812-152416 ARBEF7 1523-9829
(2013).
Google Scholar
B. Sahiner et al.,
“Computer-aided characterization of mammographic masses: accuracy of mass segmentation and its effects on characterization,”
IEEE Trans. Med. Imaging, 20
(12), 1275
–1284 https://doi.org/10.1109/42.974922 ITMID4 0278-0062
(2001).
Google Scholar
O. Ronneberger, P. Fischer and T. Brox,
“U-Net: convolutional networks for biomedical image segmentation,”
Lect. Notes Comput. Sci., 9351 234
–241 https://doi.org/10.1007/978-3-319-24574-4_28 LNCSD9 0302-9743
(2015).
Google Scholar
D. Leithner et al.,
“Abbreviated MRI of the breast: does it provide value?,”
J. Magn. Reson. Imaging, 49
(7), e85
–e100 https://doi.org/10.1002/jmri.26291
(2019).
Google Scholar
Ö. Çiçek et al.,
“3D U-Net: learning dense volumetric segmentation from sparse annotation,”
Lect. Notes Comput. Sci., 9901 424
–432 https://doi.org/10.1007/978-3-319-46723-8_49 LNCSD9 0302-9743
(2016).
Google Scholar
R. Khaled et al.,
“A U-Net ensemble for breast lesion segmentation in DCE MRI,”
Comput. Biol. Med., 140 105093 https://doi.org/10.1016/j.compbiomed.2021.105093 CBMDAW 0010-4825
(2022).
Google Scholar
J. Zhang et al.,
“Hierarchical convolutional neural networks for segmentation of breast tumors in MRI with application to radiogenomics,”
IEEE Trans. Med. Imaging, 38
(2), 435
–447 https://doi.org/10.1109/TMI.2018.2865671 ITMID4 0278-0062
(2019).
Google Scholar
G. Piantadosi et al.,
“DCE-MRI breast lesions segmentation with a 3TP U-Net deep convolutional neural network,”
in IEEE 32nd Int. Symp. Comput.-Based Med. Syst. (CBMS),
628
–633
(2019). https://doi.org/10.1109/CBMS.2019.00130 Google Scholar
M. Adoui et al.,
“MRI breast tumor segmentation using different encoder and decoder CNN architectures,”
Computers, 8
(3), 52 https://doi.org/10.3390/computers8030052
(2019).
Google Scholar
L. Zhang et al.,
“Deep-learning method for tumor segmentation in breast DCE-MRI,”
Proc. SPIE, 10954 109540F https://doi.org/10.1117/12.2513090 PSISDG 0277-786X
(2019).
Google Scholar
R. Bhattacharjee et al.,
“Comparison of 2D and 3D U-Net breast lesion segmentations on DCE-MRI,”
Proc. SPIE, 11597 115970D https://doi.org/10.1117/12.2581846 PSISDG 0277-786X
(2021).
Google Scholar
L. R. Dice,
“Measures of the amount of ecologic association between species,”
Ecology, 26
(3), 297
–302 https://doi.org/10.2307/1932409 ECGYAQ 0094-6621
(1945).
Google Scholar
F. Hausdorff, Grundzüge der Mengenlehre (“Basics of Set Theory”), Leipzig Viet(
(1914). Google Scholar
F. Wilcoxon,
“Individual comparisons by ranking methods,”
Biometr. Bull., 1
(6), 80 https://doi.org/10.2307/3001968
(1945).
Google Scholar
H. B. Mann and D. R. Whitney,
“On a test of whether one of two random variables is stochastically larger than the other,”
Ann. Math. Stat., 18
(1), 50
–60 https://doi.org/10.1214/aoms/1177730491 AASTAD 0003-4851
(1947).
Google Scholar
C. E. Bonferroni,
“Teoria statistica delle classi e calcolo delle probabilità (statistical class theory and probability calculus),”
3
–62
(1936). Google Scholar
BiographyLindsay Douglas is a diagnostic imaging physics resident at Oregon Health & Science University. She earned her PhD in medical physics from the University of Chicago and her BS degree in physics from the University of Tulsa. Her research interests include machine learning applications to medical imaging, particularly breast cancer screening. She is a member of SPIE. Roma Bhattacharjee is pursuing her undergraduate degree in computer science at Princeton University, with minors in applied and computational mathematics and optimization and quantitative decision science. She is a student member of SPIE. Jordan Fuhrman is a research scientist at The University of Chicago. His research interests primarily lie in the development of innovative AI techniques for CT scan evaluation of both head and chest conditions, including improvements in clinical translatability through interpretable AI as well as improved stratification based on AI predictions. Additionally, she is a collaborating member of the Medical Imaging and Data Resource Center and is a member of AAPM, RSNA, and SPIE. Karen Drukker received her PhD in physics from the University of Amsterdam. She is a research associate professor of radiology at The University of Chicago, where she has been involved in medical imaging research for 20+ years. Her research interests include machine learning applications in the detection, diagnosis, and prognosis of disease, focusing on rigorous training/testing protocols, generalizability, performance evaluation, and bias and fairness of AI. She is a fellow of SPIE and AAPM. Qiyuan Hu is a machine learning scientist at Tempus Labs. She received her PhD in medical physics from The University of Chicago in 2021 and BA degrees in physics and mathematics from Carleton College. Her research interests include machine learning methodologies for medical image analysis. She was a student member of SPIE and an officer of The University of Chicago SPIE Student Chapter. Maryellen Giger is the A.N. Pritzker Distinguished Service Professor at the University of Chicago. Her research involves computer-aided diagnosis/machine learning in medical imaging. She is a contact PI on the NIBIB-funded Medical Imaging and Data Resource Center, a member of the National Academy of Engineering, a recipient of the AAPM Coolidge Gold Medal, the SPIE Harrison H. Barrett Award, and the RSNA Outstanding Researcher Award, and a fellow of AAPM, AIMBE, SPIE, and IEEE. |
Image segmentation
Breast
3D image processing
3D imaging standards
Magnetic resonance imaging
Education and training
Cross validation