TY - JOUR
T1 - Deep learning in breast cancer risk assessment
T2 - Evaluation of convolutional neural networks on a clinical dataset of full-field digital mammograms
AU - Li, Hui
AU - Giger, Maryellen L.
AU - Huynh, Benjamin Q.
AU - Antropova, Natalia O.
N1 - Publisher Copyright:
© 2017 Society of Photo-Optical Instrumentation Engineers (SPIE).
PY - 2017/10/1
Y1 - 2017/10/1
N2 - To evaluate deep learning in the assessment of breast cancer risk in which convolutional neural networks (CNNs) with transfer learning are used to extract parenchymal characteristics directly from full-field digital mammographic (FFDM) images instead of using computerized radiographic texture analysis (RTA), 456 clinical FFDM cases were included: a "high-risk" BRCA1/2 gene-mutation carriers dataset (53 cases), a "high-risk" unilateral cancer patients dataset (75 cases), and a "low-risk dataset" (328 cases). Deep learning was compared to the use of features from RTA, as well as to a combination of both in the task of distinguishing between high- and low-risk subjects. Similar classification performances were obtained using CNN [area under the curve (AUC)=0.83; standard error (SE)=0.03] and RTA (AUC=0.82; SE=0.03) in distinguishing BRCA1/2 carriers and low-risk women. However, in distinguishing unilateral cancer patients and low-risk women, performance was significantly greater with CNN (AUC=0.82; SE=0.03) compared to RTA (AUC=0.73; SE=0.03). Fusion classifiers performed significantly better than the RTA-alone classifiers with AUC values of 0.86 and 0.84 in differentiating BRCA1/2 carriers from low-risk women and unilateral cancer patients from low-risk women, respectively. In conclusion, deep learning extracted parenchymal characteristics from FFDMs performed as well as, or better than, conventional texture analysis in the task of distinguishing between cancer risk populations.
AB - To evaluate deep learning in the assessment of breast cancer risk in which convolutional neural networks (CNNs) with transfer learning are used to extract parenchymal characteristics directly from full-field digital mammographic (FFDM) images instead of using computerized radiographic texture analysis (RTA), 456 clinical FFDM cases were included: a "high-risk" BRCA1/2 gene-mutation carriers dataset (53 cases), a "high-risk" unilateral cancer patients dataset (75 cases), and a "low-risk dataset" (328 cases). Deep learning was compared to the use of features from RTA, as well as to a combination of both in the task of distinguishing between high- and low-risk subjects. Similar classification performances were obtained using CNN [area under the curve (AUC)=0.83; standard error (SE)=0.03] and RTA (AUC=0.82; SE=0.03) in distinguishing BRCA1/2 carriers and low-risk women. However, in distinguishing unilateral cancer patients and low-risk women, performance was significantly greater with CNN (AUC=0.82; SE=0.03) compared to RTA (AUC=0.73; SE=0.03). Fusion classifiers performed significantly better than the RTA-alone classifiers with AUC values of 0.86 and 0.84 in differentiating BRCA1/2 carriers from low-risk women and unilateral cancer patients from low-risk women, respectively. In conclusion, deep learning extracted parenchymal characteristics from FFDMs performed as well as, or better than, conventional texture analysis in the task of distinguishing between cancer risk populations.
KW - breast cancer risk assessment
KW - convolutional neural network
KW - deep learning
KW - full-field digital mammogram
KW - mammographic parenchymal patterns
KW - radiographic texture analysis
KW - transfer learning
UR - http://www.scopus.com/inward/record.url?scp=85029820268&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85029820268&partnerID=8YFLogxK
U2 - 10.1117/1.JMI.4.4.041304
DO - 10.1117/1.JMI.4.4.041304
M3 - Article
AN - SCOPUS:85029820268
SN - 2329-4302
VL - 4
JO - Journal of Medical Imaging
JF - Journal of Medical Imaging
IS - 4
M1 - 041304
ER -