Computerized Medical Imaging and Graphics
Computerized Medical Imaging and Graphics
a r t i c l e i n f o a b s t r a c t
Article history: Breast cancer is the second leading cause of cancer death among women worldwide. Nevertheless, it is
Received 15 April 2018 also one of the most treatable malignances if detected early. Screening for breast cancer with full field
Received in revised form 28 August 2018 digital mammography (FFDM) has been widely used. However, it demonstrates limited performance
Accepted 13 September 2018
for women with dense breasts. An emerging technology in the field is contrast-enhanced digital mam-
mography (CEDM), which includes a low energy (LE) image similar to FFDM, and a recombined image
Keywords:
leveraging tumor neoangiogenesis similar to breast magnetic resonance imaging (MRI). CEDM has shown
Deep learning
better diagnostic accuracy than FFDM. While promising, CEDM is not yet widely available across medical
Image synthesis
Breast tumor
centers. In this research, we propose a Shallow-Deep Convolutional Neural Network (SD-CNN) where
Digital mammography a shallow CNN is developed to derive “virtual” recombined images from LE images, and a deep CNN is
CEDM employed to extract novel features from LE, recombined or “virtual” recombined images for ensemble
Classification models to classify the cases as benign vs. cancer. To evaluate the validity of our approach, we first develop
a deep-CNN using 49 CEDM cases collected from Mayo Clinic to prove the contributions from recombined
images for improved breast cancer diagnosis (0.85 in accuracy, 0.84 in AUC using LE imaging vs. 0.89 in
accuracy, 0.91 in AUC using both LE and recombined imaging). We then develop a shallow-CNN using
the same 49 CEDM cases to learn the nonlinear mapping from LE to recombined images. Next, we use
89 FFDM cases from INbreast, a public database to generate “virtual” recombined images. Using FFDM
alone provides 0.84 in accuracy (AUC = 0.87), whereas SD-CNN improves the diagnostic accuracy to 0.90
(AUC = 0.92).
© 2018 Elsevier Ltd. All rights reserved.
https://doi.org/10.1016/j.compmedimag.2018.09.004
0895-6111/© 2018 Elsevier Ltd. All rights reserved.
54 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62
tal mammography using the intra-venous injection of an iodinated models are established by experienced researchers using publicly
contrast agent in conjunction with a mammography examination. labeled image datasets. For a specific task, the model is often treated
Two techniques have been developed to perform CEDM exami- as a feature generator to extract features describing the images
nations: the temporal subtraction technique with acquisition of from abstract level to detailed levels. One can then develop clas-
high-energy images before and after contrast medium injection sification models (SVMs, ANNs, etc.) using the derived features.
and the dual energy technique with acquisition of a pair of low and Promising results have been reported in several medical appli-
high-energy images only after contrast medium injection. During cations, such as chest pathology identification (Bar et al., 2015),
the exam, a pair of low and high-energy images is obtained after breast mass detection and classification (Samala et al., 2016), just
the administration of a contrast medium agent. The two images are to name a few. While exciting, earlier CNN models such as AlexNet
combined to enhance contrast uptake areas and the recombined (Krizhevsky et al., 2012), GoogLeNet (Simonyan and Zisserman,
image is then generated (Fallenberg et al., 2014). In CEMD, it has 2014) and VGGNet (Szegedy et al., 2014) are known to suffer from
low energy (LE) imaging, which is comparable to routine FFDM and gradient vanishing when the number of layers increases signifi-
recombined imaging similar to breast MRI. Comparing to breast cantly. A newer model, ResNet (He et al., 2014) with a “short-cut”
MRI, CEDM exam is about 4 times faster with only about 1/6 the architecture is recently proposed to address the issue. The imag-
cost (Patel et al., 2017). In addition, CEDM imaging has 10 times ing competition results show the ResNet outperforms other CNN
the spatial resolution of breast MRI. Therefore, CEDM can be used models by at least 44% in classification accuracy.
to more sensitively detect small residual foci of tumor, including The potentials CNN brings to medical imaging research are not
calcified Ductal Carcinoma in Situ (DCIS), than using MRI (Patel limited to deep CNN for imaging feature extraction. A second area
et al., 2017). Several studies including prospective clinical trials that medical research can benefit is indeed using CNN for syn-
conducted at Mayo Clinic have indicated that CEDM is a promising thetic image rendering. Here an image is divided into a number of
imaging modality that overcomes tissue overlapping (“masking”) smaller patches fed into a CNN (e.g., 4-layer CNN in this research)
occurred in FFDM, provides tumor neovascularity related func- as the input and the output is a synthetic image. The CNN is trained
tional information similar to MRI, while maintaining high image to learn the non-linear mapping between the input and output
resolution of FFDM (Cheung et al., 2014; Fallenberg et al., 2014; images. Several successful applications have been reported, such as
Gillman et al., 2014; Luczyńska et al., 2014). Unfortunately, CEDM synthesizing positron emission tomography (PET) imaging (Li et al.,
as a new modality is yet widely available in many other medical 2014) or CT image (Han, 2017; Nie et al., 2017) from MRI image, and
centers or breast cancer screening facilities in the U.S. and/or across from regular X-ray to bone-suppressed recombined X-ray (Yang
the world limiting its broad clinical impacts. et al., 2017).
In clinical breast imaging (US, MRI, FFDM and CEDM), reading Motivated by this two-fold applicability of CNN, this research
and interpreting the images remains a difficult task for radiologists. proposes a Shallow-Deep CNN (SD-CNN) as a new CAD scheme to
Currently, breast cancer screening has high false positive recall tackle the unique problem stemmed from the novel imaging modal-
rate (i.e., ≥10%). Computer-aided detection (CADe) and diagnosis ity, CEDM, for breast cancer diagnosis. Our first hypothesis is that
(CADx) schemes (Tan et al., 2011; Carneiro et al., 2017; Gao et al., applying a deep CNN to CEDM is capable of taking advantage of
2016; Muramatsu et al., 2016) have been developed and demon- recombined imaging for improved breast lesion classification due
strated the clinical potentials to be used as “the second reader” to to the contribution from the tumor functional image features. Sec-
help improve radiologists’ performance in the diagnosis. In order ond, in order to expand the advantages of CEDM imaging modality
to overcome the limitation of lower accessibility to CEDM systems to the regular FFDM modality, we hypothesize that a shallow CNN is
and help radiologists more accurately conduct the diagnosis, this capable to discover the nonlinear mapping between LE and recom-
research proposes the development and validation of a new CADx bined images to synthesize the “virtual” recombined images. As a
scheme, termed Shallow-Deep Convolutional Neural Network (SD- result, traditional FFDM can be enriched with the “virtual” recom-
CNN). SD-CNN combines image processing and machine learning bined images. The objective of this study is to validate these two
techniques to improve the malignancy diagnosis using FFDM by hypotheses by using a unique study procedure and two imaging
taking advantages of information available from the CEDM. datasets of both CEDM and FFDM images. The details of the study
CNN is a feed-forward artificial neural network that has been procedures and experimental results are reported in the following
successfully implemented in the broad computer vision areas for sections of this paper.
decades (Lecun et al., 2015; LeCun et al., 1998). As it evolves,
different CNN models have been developed and implemented. 2. Materials
The computational resource and devices available in recent years
make the training of CNN with large number of layers (namely, In this research, two separate datasets are used, which include
the deep CNN) possible. Applying deep CNNs in image recogni- a dataset acquired from tertiary medical center (Mayo Clinic Ari-
tion was probably first demonstrated in ImageNet competition zona), and a public dataset from INbreast (Moreira et al., 2012).
(Russakovsky et al., 2015) back in 2012. Since then, it has become
a popular model for various applications ranging from natural lan- 2.1. Institutional dataset from Mayo Clinic Arizona
guage processing, image segmentation to medical imaging analysis
(Cha et al., 2016; Tajbakhsh et al., 2016). The main power of a Based on Institutional Review Board (IRB) approved study and
deep CNN lies in the tremendous trainable parameters in differ- data collection protocol, we reviewed CEDM examinations per-
ent layers (Eigen et al., 2013; Zeiler and Fergus, 2014). These are formed using the Hologic Imaging system (Bedford, MA, USA)
used to extract discriminative features at different level of abstrac- between August 2014 and December 2015. All patients undertaken
tion (Tajbakhsh et al., 2016). However, training a deep CNN often CEDM had a BI-RADS (Breast Imaging Reporting and Data Systems)
requires a large volume of labeled training data, which may not be (Liberman and Menell, 2002) rating of 4 and 5 in their original FFDM
easily available in medical applications. Secondly, training a deep screening images. Due to the detection of highly suspicious breast
CNN requires massive computational resources, as well as rigorous lesions, CEDM was offered as an adjunct test to biopsy in a clini-
research in architecture design and hyper-parameters tuning. To cal trial environment. All CEDM tests were performed prior to the
address these challenges, a promising solution is transfer learning biopsies. In summary, the patient cohort in this clinical trial had
(Banerjee et al., 2017), that is, a deep CNN model is trained fol- the following criteria: 1) the diagnostic mammogram was rated BI-
lowed by a task-specific parameter fine-tuning process. The trained RADS 4 or 5, and 2) histopathology test result was available from
F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62 55
Fig. 1. Example of breast images (Cancer and Benign) for LE and recombined (Rec) images with 2 views (CC and MLO) (Lesions are highlighted with green circle). (For
interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article).
surgical or image-guided biopsy. We limited the cohort to BIRADS system with pixel size of 70 mm (microns), and 14-bit contrast res-
4 and 5 lesions because the analysis required the gold standard of olution. For each subject, both CC and MLO view were available.
lesion pathology. 49 cases were identified that met the above inclu- For each image, the annotations of region of interests (ROIs) were
sion criteria, which include 23 benign and 26 cancer biopsy-proven made by a specialist in the field, and validated by a second special-
lesions. We analyzed one lesion per patient. If a patient had mul- ist. The masks of ROIs were also made available. In this research, a
tiple enhancing lesions, the annotating radiologist used the largest dataset of 89 subjects was extracted by including subjects that have
lesion to ensure best feature selection. In CEDM, there are cranial- BI-RADS scores of 1, 2, 5 and 6. Subjects with BI-RADS 1 and 2 are
caudal (CC) and mediolateral-oblique (MLO) views for both LE and regarded as benign tumor, and subjects with BI-RADS 5 and 6 are
recombined images. Fig. 1 illustrates the example views on the LE regarded as cancer. For each subject, images of CC and MLO view
and recombined images, respectively. are used for feature extraction. Examples from INbreast dataset is
For the 49 cases, all CEDM images with DICOM format were iden- shown in Fig. 2.
tified and transferred from the clinical PACS to a research database
and loaded into the open source image processing tool OsiriX 3. Methodology
(OsiriX foundation, Geneva, Switzerland) (Rosset et al., 2004).
DICOM images were anonymized and prepared for blinded read- To fully explore the advantages of CNNs and CEDM in breast
ing by a radiologist. A fellowship trained breast radiologist with cancer research, a Shallow-Deep CNN (SD-CNN) is proposed
over 8 years of imaging experience interpreted the mammogram (Fig. 3). First, we develop a Shallow-CNN from CEDM to discover
independently and used the OsiriX tool to outline lesion contours. the relationships between LE images and recombined images.
Contours were drawn on recombined images (both CC and MLO This Shallow-CNN is then applied to FFDM to render “virtual”
views) for each patient on recombined images. These contours were recombined images. Together with FFDM, a trained Deep-CNN is
then cloned onto LE images. All lesions were visible on both CC and introduced for feature extraction followed by classification models
MLO views. This information is further used in the imaging pre- for diagnosis. Note for CEDM, we can start the workflow with the
processing (see details in methodology section). Some examples Deep-CNN directly.
of LE and recombined images are shown in Fig. 1. As observed, LE
images are not as easy as recombined images to visualize the lesions 3.1. Image pre-processing
for both cancerous and benign cases.
Before the Deep-CNN and Shallow-CNN are employed, a four-
2.2. INbreast public dataset step imaging pre-processing procedure is launched. First, for each
image we identify a minimum-area bounding box that contains the
This dataset was obtained from INbreast, an online accessible tumor region. Specifically, for each tumor, we have a list of bound-
full-field digital mammographic database (Moreira et al., 2012). ary points with coordinates in pair (x,y) available. The bounding
INbreast was established by the researchers from the Breast Center box is decided using the (xmin , ymin ) and (xmax , ymax ) as the two
in CHJKS, Porto, under the permission of both the Hospital’s Ethics diagonal corner points to ensure the box covers the whole tumor
Committee and the National Committee of Data Protection. The area. Note we have CC and MLO views for FFDM and we have CC
FFDM images were acquired from the MammoNovation Siemens and MLO views for both LE and recombined images for CEDM. As
56 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62
Fig. 2. Example of breast images for FFDM images from INbrease dataset with 2 views (CC on left and MLO on right) (Lesions are highlighted with green circle). (For
interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article).
a result, there are two images from FFDM and four images from imal preprocessing. Individual neurons respond to stimuli only in
CEDM. The bounding box size varies case by case due to differ- a restricted region of the visual field known as the receptive field.
ent sizes of tumors (ranging from 65 × 79 to 1490 × 2137 in this This process is simulated through different layers (convolutional,
study). Next, an enlarged rectangle that is 1.44 times (1.2 times in pooling, fully connected). A CNN’s capability is hidden behind the
width and 1.2 times in height) the size of bounding box is obtained. large amount trainable parameters which can be learned iteratively
The enlarged bounding box approach is to include sufficient neigh- through gradient descent algorithms. In this research, a 4-layer CNN
borhood information proved to increase the classification accuracy (Fig. 4) is implemented to model the latent relationship between
(Lévy and Jain, 2016). In the second step, this ‘enlarged’ rectangle the LE images (patches) and recombined images (patches). The
is extracted and saved as one image. The third step is to normalize model is then used to render “virtual” recombined images (patches)
the image intensity to be between 0 and 1 using the max-min nor- from FFDM images (patches).
malization. In the last step, the normalized images are resized to
224 × 224 to fully take advantage of trained ResNet model. Here we
3.3. Deep-CNN: feature generation
take the patches that contain tumor instead of the whole image as
input. This is because the focus of the study is on tumor diagnosis
ResNet is a trained deep CNN developed in 2015 with a revolu-
and we believe the features generated by the deep-CNN from the
tionary architecture using the “short-cut” concept in the building
tumor region shall better characterize the tumor, especially for the
block. As seen in Fig. 5, the output of building blocks takes both final
cases where the tumor region is small.
classification results and the initial inputs (the short-cut) when
updating the parameters. As a result, it outperforms traditional
3.2. Shallow-CNN: virtual image rendering deep-CNNs which are known to suffer from higher testing error
since gradient tends to vanish as the number of layers increases
Inspired by the biological processes (Elmore et al., 2005), CNNs (He et al., 2014). ResNet has different versions with 50, 101 and 152
use a variation of multilayer perceptions designed to require min- layers but all based on the same building blocks. In the ImageNet
F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62 57
Boosting is a machine learning ensemble meta-algorithm aim- The workflow of our first experiment is shown in Fig. 7. Using
ing to reduce bias and variance (Bauer et al., 1999). It converts 49 CEDM cases collected from Mayo Clinic Arizona, we first con-
weak learners to strong ones by weighing each training sample duct the experiments using LE from CEDM images. For each subject
inversely correlated to the performance of previous weak learners. in the dataset, LE images (both CC and MLO views) are processed
58 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62
Fig. 6. Architecture of ResNet [37] (Red star are placed in layers where features are extracted; Dotted shortcuts increase feature dimensions by zero-padding; based on the
output dimension of building blocks, the ResNet is divided into 4 different building blocks (BBs), they are shown with different colors in the figure (BB 1: blue (Fig. 5 Right),
BB 2: orange, BB 3: purple, BB 4: green). Different version of ResNets vary in the number of BBs, for instance, the 50-layer version ResNet has 3 BB 1 s, 4 BB 2 s, 6 BB 3 s and
3 BB 4 s). (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article).
Here we first develop a 4-layer shallow CNN that learns the non-
linear mapping between the LE and recombined images using the
same 49 CEDM dataset. CC and MLO view images are regarded as
separate training data, so a total of 98 images are used, in which 5
subjects (10 images) are selected as validation material, and the rest
44 subjects (88 images) are sued as training material. By randomly
extracting 2500 pair of training samples within masked tumor from
each LE (input) and recombined image (output), a training dataset
of 220,000 (88 × 2500) samples is generated. The input samples
for the CNN are 15 × 15 patches from LE images, the same input
size as in (Li et al., 2014). Considering the relatively small receptive
field and complexity of a shallow CNN, we set the output samples
size as 3 × 3, it is our intention to explore the impact of the differ-
ent output patch size for the breast cancer diagnosis as one of our
future tasks. The output patches from recombined image are cen-
tered in the same position as input patches from the LE image. The
input and output samples are fed into the CNN framework imple-
mented with package of “Keras”. The CNN has 2 hidden layers, with
10 7 × 7 filters in each layer. There are 5 K trainable parameters
through backpropagation with mini-batch gradient decent algo-
rithm to increase learning speed. Batch size is set to be 128. The
Fig. 8. Receiver operating characteristic curve for the model using FFDM image only learning rate is set to be 0.01, ReLu activation function is used in
vs. FFDM and recombined image.
all layers except the output layer, where activation function is not
used. Other parameters are set to be default by “Keras” package.
in the modeling. The features from the recombined images help Finally, with the trained CNN and patches extracted from avail-
improve the accuracy of breast cancer diagnosis from 0.85 to 0.89. able modality, we can construct a “virtual” recombined image by
assembling predicted patches into a whole image.
We use mean squared error (MSE) to evaluate the similarity
4.2. Experiment II: validating the value of “Virtual” recombined between the “virtual” recombined image and the true recombined
imaging in breast cancer diagnosis on FFDM using SD-CNN image for the 10 images in validation dataset. MSE measures the
pairwise squared difference in intensity as:
The improved performance by adding the features from recom-
bined images motivates us to study the validity of constructing and 1 N 2
MSE = TRecombined (i) − VRecombined (i)
using the “virtual” recombined images from FFDM images for breast N i=1
cancer diagnosis.
Fig. 9. Sample images of LE image, true recombined image and its corresponding virtual recombined in dataset I. (from left to right: benign, cancer, cancer, cancer).
60 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62
Fig. 10. Sample images of FFDM in dataset II and its corresponding “virtual” recombined (Two cases on left are cancerous with BI-RADS = 5, two cases on right are benign
with BI-RADS = 2).
Table 5
Contribution of features from different image sources.
LE image 67 22.33%
Virtual 87 77.67%
Recombined image
Table 6
Classification performance for using FFDM feature alone and using features from FFDM and “virtual” recombined and other state-of-the-art methods using INbreast dataset.
Random Forest on features from CNN with pre-training 0.95 ± 0.05 0.91 ± 0.12
(Dhungel et al., 2017)
CNN + hand crafted features pre-training 0.91 ± 0.06 0.87 ± 0.06
(Dhungel et al., 2017)
Random Forest + hand crafted features pre-training 0.90 ± 0.02 0.80 ± 0.15
(Dhungel et al., 2017)
CNN without hand crafted features pre-training 0.72 ± 0.16 0.82 ± 0.07
(Dhungel et al., 2017)
Multilayer perceptron (Sasikala and Ezhilarasi, 2016) 0.88 0.89
Lib SVM (Diz et al., 2016) 0.89 0.90
Multi-kernel classifier (Augusto, 2014) NA 0.87
Linear Discriminant analysis (Domingues et al., 2012) 0.89 NA
Our proposed approach on FFDM only 0.84 ± 0.09 0.87 ± 0.12
Our proposed approach on both FFDM and “virtual” Recombined Image 0.90 ± 0.06 0.92 ± 0.14
specificities (1 – false positive rate) vary greatly. We want to high- we propose a SD-CNN (Shallow-Deep CNN) to study the two-fold
light the importance of specificity as breast cancer screening has applicability of CNN to improve the breast cancer diagnosis. One
high false positive recall rate (i.e., ≥ 10%). One known fact is that contribution of this study is to investigate the advantages of recom-
the probability that a woman will have at least one false positive bined images from CEDM in helping the diagnosis of breast lesions
diagnosis at 10 years screening program is 61.3% with annual and using a Deep-CNN method. CEDM is a promising imaging modal-
41.6% with biennial screening (Michaelson et al., 2016). This will ity providing information from standard FFDM combined with
lead to additional MR exams (extra cost) and even biopsy. Another enhancement characteristics related to neoangiogenesis (similar
side effect is the negative psychological impacts. In this research, to MRI). Based on our review of literature, no existing study has
the use of recombined images (“virtual” recombined images) shows investigated the extent of CEDM imaging potentials using the deep-
the great potential to address these challenges by improving the CNN. Using the state-of-art trained ResNet as a feature generator for
specificity. In Table 4, we summarize the model performances in classification modeling, our experiment shows the features from
terms of accuracy, sensitivity and specificity (threshold is set to be LE images can achieve accuracy of 0.85 and AUC of 0.84, adding
0.75). While we observe that the model on FFDM vs. the model on the recombined imaging features, model performance improves to
FFDM and “virtual” recombined image show no significant differ- accuracy of 0.89 with AUC of 0.91.
ences on accuracy, sensitivity and even AUC, the performance on Our second contribution lies in addressing the limited accessi-
specificity shows significant improvements (p < 0.05). bility of CEDM and developing SD-CNN to improve the breast cancer
In looking into the contributions from the features (Table 5), diagnosis using FFDM in general. This the first study to develop a
the use of “virtual” recombined imaging features improves the 4-layer shallow CNN to discover the nonlinear association between
performances in terms of both accuracy and AUC. Calculation of LE and recombined images from CEDM. The 4-layer shallow-CNN
contribution follows the same procedure as experiment I and is can be applied to render “virtual” recombined images from FFDM
conducted inside each cross-validation loop. Among all the 154 fea- images to fully take advantage of the CEDM in improved breast can-
tures used in this experiment, 87 are from the “virtual” recombined cer diagnosis. Our experiment on 89 FFDM dataset using the same
image, which contribute 77.67% of the total impurity reduction. trained ResNet achieves accuracy of 0.84 with AUC of 0.87. With
The rest 67 features are from LE images, and they contributed the the “virtual” recombined imaging features, the model performance
rest 22.33% impurity reduction. It is interesting to observe from is improved to accuracy of 0.90 with AUC of 0.92.
this experiment that the contributions from “virtual” recombined While promising, there is room for future work. First of all,
images are higher than the contributions from the true recombined the trained ResNet is a black-box feature generator, the features
images from the first experiment. One reason may be the second extracted may not be easy to be interpreted by the physicians. It
dataset has more denser tissue cases and it is believed recombined is our intention to discover possible clinical interpretations from
images shall be more useful in diagnosing the dense breast cases. the features as one of our future tasks. For example, as the ResNet
This is yet to be confirmed with the radiologists which is our imme- goes deeper, the initial layers of the features may represent the
diate next step. raw imaging characteristics as the first order statistics, the deeper
We further explore the state-of-the-art algorithms using the layer of the features may represent the morphological characteris-
same INbreast dataset and compare our methods against the eight tics (e.g., shape). This is yet to be explored. A second future work is
methods from the literature (Table 6). As seen, our approach using related to the patch sizes. We plan to assess impacts of the differ-
“virtual” recombined image outperforms six algorithms in terms of ent sized patches for both input and output images on the breast
both accuracy and AUC. We want to highlight that one of papers by cancer diagnosis.
Dhungel et al. (2017) proposes four approaches. Among the four,
the best performer has a 0.95 in accuracy and 0.91 in AUC, and the
Acknowledgements
second performer has a 0.91 in accuracy and 0.87 in AUC. We con-
clude our approach has better AUC (0.92) comparing to both while
We would like to acknowledge the ASU/Mayo Clinic collab-
inferior in accuracy (0.90). We contend that indeed, AUC is a more
orative research seed grant program for financial support and
robust metric in the medical research and it is considered to be
resources required for this research project.
more consistent and have better discriminatory power comparing
to accuracy (Huang and Ling, 2005).
References
5. Discussion and conclusion
Augusto, G.B., 2014. Multiple Kernel Learning for Breast Cancer Classification.
Bar, Y., Diamant, I., Wolf, L., Greenspan, H., 2015. Deep Learning with Non-medical
Differentiating benign cases from malignant lesions is one of Training Used for Chest Pathology Identification 94140V., http://dx.doi.org/10.
the remaining challenges of breast cancer diagnosis. In this study, 1117/12.2083124.
62 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62
Bauer, E., Kohavi, R., Chan, P., Stolfo, S., Wolpert, D., 1999. An empirical comparison Li, R., Zhang, W., Suk, H., Wang, L., Li, J., Shen, D., Ji, S., 2014. Deep Learning Based
of voting classification algorithms: bagging, boosting, and variants. Mach. Imaging Data Completion for Improved Brain Disease Diagnosis., pp. 305–312.
Learn. 36, 105–139, http://dx.doi.org/10.1023/a:1007515423169. Liberman, Laura, Menell, Jennifer H., 2002. Breast imaging reporting and data
Banerjee, Debrup, et al., 2017. A deep transfer learning approach for improved system (BI-RADS). Radiol. Clin. 40 (3), 409–430.
post-traumatic stress disorder diagnosis. In: Data Mining (ICDM), 2017, IEEE Luczyńska, E., Heinze-Paluchowska, S., Dyczek, S., Blecharz, P., Rys, J., Reinfuss, M.,
International Conference on, IEEE. 2014. Contrast-enhanced spectral mammography: comparison with
Breastcancer.org. (2018). U.S. Breast Cancer Statistics | Breastcancer.org. [online] conventional mammography and histopathology in 152 women. Korean J.
Available at: https://www.breastcancer.org/symptoms/understand bc/ Radiol. 15, 689–696, http://dx.doi.org/10.3348/kjr.2014.15.6.689.
statistics [Accessed 25 Sep. 2018]. Michaelson, J.S., Shih, Y.T., Walter, L.C., Church, T.R., Flowers, C.R., Lamonte, S.J.,
Carneiro, G., Nascimento, J., Bradley, A.P., 2017. Automated analysis of Wolf, M.D., Desantis, C., Lortet-tieulent, J., Andrews, K., Manassaram-baptiste,
unregistered multi-view mammograms with deep learning. IEEE Trans. Med. D., Saslow, D., Smith, R.A., Otis, W., 2016. Guideline Update from the American
Imaging 36, 2355–2365, http://dx.doi.org/10.1109/TMI.2017.2751523. Cancer Society, 314., pp. 1599–1614, http://dx.doi.org/10.1001/jama.2015.
Cha, K.H., Hadjiiski, L., Samala, R.K., Chan, H.-P., Caoili, E.M., Cohan, R.H., 2016. 12783.Breast.
Urinary bladder segmentation in CT urography using deep-learning Moreira, I.C., Amaral, I., Domingues, I., Cardoso, A., Cardoso, M.J., Cardoso, J.S., 2012.
convolutional neural network and level sets. Med. Phys. 43, 1882–1896, http:// INbreast: toward a full-field digital mammographic database. Acad. Radiol. 19,
dx.doi.org/10.1118/1.4944498. 236–248, http://dx.doi.org/10.1016/j.acra.2011.09.014.
Cheung, Y.C., Lin, Y.C., Wan, Y.L., Yeow, K.M., Huang, P.C., Lo, Y.F., Tsai, H.P., Ueng, Muramatsu, C., Hara, T., Endo, T., Fujita, H., 2016. Breast mass classification on
S.H., Chang, C.J., 2014. Diagnostic performance of dual-energy mammograms using radial local ternary patterns. Comput. Biol. Med. 72,
contrast-enhanced subtracted mammography in dense breasts compared to 43–53, http://dx.doi.org/10.1016/j.compbiomed.2016.03.007.
mammography alone: interobserver blind-reading analysis. Eur. Radiol. 24, Nie, Dong, Cao, Xiaohuan, Gao, Yaozong, Wang, Li., SD, 2017. Estimating CT Image
2394–2403, http://dx.doi.org/10.1007/s00330-014-3271-1. from MRI Data Using 3D Fully. Convolutional Net. 1, 170–178, http://dx.doi.
Dhungel, N., Carneiro, G., Bradley, A.P., 2017. A deep learning approach for the org/10.1007/978-3-319-46976-8.
analysis of masses in mammograms with minimal user intervention. Med. Patel, B.K., Garza, S.A., Eversman, S., Lopez-Alvarez, Y., Kosiorek, H., Pockaj, B.A.,
Image Anal. 37, 114–128, http://dx.doi.org/10.1016/j.media.2017.01.009. 2017. Assessing tumor extent on contrast-enhanced spectral mammography
Diz, J., Marreiros, G., Freitas, A., 2016. Applying data mining techniques to improve versus full-field digital mammography and ultrasound. Clin. Imag. 46, 78–84,
breast Cancer diagnosis. J. Med. Syst., 40, http://dx.doi.org/10.1007/s10916- http://dx.doi.org/10.1016/j.clinimag.2017.07.001.
016-0561-y. Rokach, L., Oded, Z., 2008. Maimon. Data mining with decision trees: theory and
Domingues, I., Sales, E., Pereira, W.C.A., Tec, I., Faculdade, P., Porto, U., Engenharia, applications, 69. World scientific.
P.De, Coppe, B., Janeiro, R.De, 2012. Inbreast-Database Masses Rosenquist, C.J., Lindfors, K.K., 1998. Screening mammography beginning at age 40
Characterization., pp. 1–5. years: a reappraisal of cost-effectiveness. Cancer 82, 2235–2240, http://dx.doi.
Eigen, D., Rolfe, J., Fergus, R., LeCun, Y., 2013. Understanding Deep Architectures org/10.1002/(SICI)1097-0142(19980601)82:11<2235::AID-CNCR19>3.0.CO;2-
Using a Recursive Convolutional Network., pp. 1–9. V.
Elmore, J.G., Armstrong, K., Lehman, C.D., Fletcher, S.W., 2005. CLINICIAN’ S Rosset, A., Spadola, L., Ratib, O., 2004. OsiriX: an open-source software for
CORNER screening for breast Cancer. J. Am. Med. Assoc. 293, 1245–1256. navigating in multidimensional DICOM images. J. Digit. Imaging 17, 205–216,
Fallenberg, E.M., Dromain, C., Diekmann, F., Engelken, F., Krohn, M., Singh, J.M., http://dx.doi.org/10.1007/s10278-004-1014-6.
Ingold-Heppner, B., Winzer, K.J., Bick, U., Renz, D.M., 2014. Contrast-enhanced Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy,
spectral mammography versus MRI: initial results in the detection of breast A., Khosla, A., Bernstein, M., Berg, A.C., Fei-Fei, L., 2015. ImageNet large scale
cancer and assessment of tumour size. Eur. Radiol. 24, 256–264, http://dx.doi. visual recognition challenge. Int. J. Comput. Vis. 115, 211–252, http://dx.doi.
org/10.1007/s00330-013-3007-7. org/10.1007/s11263-015-0816-y.
Gao, F., Zhang, M., Wu, T., Bennett, K.M., 2016. 3D small structure detection in Samala, R.K., Chan, H.-P., Hadjiiski, L., Helvie, M.A., Wei, J., Cha, K., 2016. Mass
medical image using texture analysis. 2016 38th Annu. Conf. Proc. IEEE Eng. detection in digital breast tomosynthesis: deep convolutional neural network
Med. Biol. Soc., 6433–6436, http://dx.doi.org/10.1109/EMBC.2016.7592201. with transfer learning from mammography. Med. Phys. 43, 6654–6666, http://
Gillman, J., Toth, H.K., Moy, L., 2014. The role of dynamic contrast-enhanced dx.doi.org/10.1118/1.4967345.
screening breast MRI in populations at increased risk for breast cancer. Sasikala, S., Ezhilarasi, M., 2016. Combination of Mammographic Texture Feature
Women’s Heal. 10, 609–622, http://dx.doi.org/10.2217/WHE.14.61. Descriptors for Improved Breast Cancer Diagnosis. Asian J. Infor.Technol. 15
Han, X., 2017. MR-based synthetic CT generation using a deep convolutional neural (20), 4054–4062.
network method. Med. Phys. 44, 1408–1419, http://dx.doi.org/10.1002/mp. Simonyan, K., Zisserman, A., 2014. Very Deep Convolutional Networks for
12155. Large-Scale Image Recognition., pp. 1–14, http://dx.doi.org/10.1016/j.infsof.
He, K., Zhang, X., Ren, S., Sun, J., https://doi.org/10.1.1.725.4861 2014. Delving Deep 2008.09.005.
Into Rectifiers: Surpassing Human-level Performance on ImageNet Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D.,
Classification. Vanhoucke, V., Rabinovich, A., arXiv:1409.4842 2014. Going Deeper With
Huang, Jin, Ling, Charles X., 2005. Using AUC and accuracy in evaluating learning Convolutions., http://dx.doi.org/10.1109/CVPR.2015.7298594.
algorithms. IEEE Trans. Knowl. Data Eng. 17 (3), 299–310. Tajbakhsh, N., Shin, J.Y., Gurudu, S.R., Hurst, R.T., Kendall, C.B., Gotway, M.B., Liang,
Krizhevsky, A., Sutskever, I., Hinton, G.E., 2012. ImageNet classification with deep J., 2016. Convolutional neural networks for medical image analysis: full
convolutional neural networks. Adv. Neural Inf. Process. Syst., 1–9, http://dx. training or fine tuning? IEEE Trans. Med. Imaging 35, 1299–1312, http://dx.doi.
doi.org/10.1016/j.protcy.2014.09.007. org/10.1109/TMI.2016.2535302.
Lecun, Y., Bengio, Y., Hinton, G., 2015. Deep learning. Nature 521, 436–444, http:// Tan, Maxine, et al., 2011. A novel computer-aided lung nodule detection system for
dx.doi.org/10.1038/nature14539. CT images. Med. Phys. 38 (10), 5630–5645.
LeCun, Y., Bottou, L., Bengio, Y., Haffner, P., 1998. Gradient-based learning applied Warner, E., Plewes, D.B., Hill, K.A., Causer, P.A., Jong, R.A., Cutrara, M.R., Yaffe, M.J.,
to document recognition. Proc. IEEE 86, 2278–2323, http://dx.doi.org/10.1109/ Messner, S.J., Meschino, W.S., Piron, C.A., Narod, S.A., 2004. Surveillance of
5.726791. BRCA1 and BRCA2 mutation. JAMA 292, 1317–1325.
Lehrer, D., Jong, R.A., Pisano, E.D., Barr, R.G., Mahoney, M.C., Iii, W.P.E., Larsen, L.H., Yang, W., Chen, Y., Liu, Y., Zhong, L., Qin, G., Lu, Z., Feng, Q., Chen, W., 2017. Cascade
Morton, M.J., Mendelson, E.B., Farria, D.M., Cormack, J.B., Marques, H.S., Adams, of multi-scale convolutional neural networks for bone suppression of chest
A., https://doi.org/307(13):1394-1404 2012. Detection of Breast Cancer with radiographs in gradient domain. Med. Image Anal. 35, 421–433, http://dx.doi.
Addition of Annual Screening Ultrasound or a Single ScreeningMRI org/10.1016/j.media.2016.08.004.
toMammography inWomen With Elevated Breast Cancer Risk 307. Zeiler, M.D., Fergus, R., 2014. Visualizing and understanding convolutional
Lévy, D., Jain, A., 2016. Breast Mass Classification from Mammograms Using Deep networks arXiv:1311.2901v3 [cs.CV] 28 nov 2013. Comput. Vision–ECCV 8689,
Convolutional Neural Networks. 818–833, http://dx.doi.org/10.1007/978-3-319-10590-1 53.