0% found this document useful (0 votes)
100 views10 pages

Computerized Medical Imaging and Graphics

4
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
100 views10 pages

Computerized Medical Imaging and Graphics

4
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 10

Computerized Medical Imaging and Graphics 70 (2018) 53–62

Contents lists available at ScienceDirect

Computerized Medical Imaging and Graphics


journal homepage: www.elsevier.com/locate/compmedimag

SD-CNN: A shallow-deep CNN for improved breast cancer diagnosis


Fei Gao a , Teresa Wu a,∗ , Jing Li a , Bin Zheng b , Lingxiang Ruan c , Desheng Shang c ,
Bhavika Patel d
a
School of Computing, Informatics, and Decision Systems Engineering, Arizona State University, Tempe, AZ, 85281, USA
b
School of Electrical and Computer Engineering, College of Engineering, University of Oklahoma, Norman, OK, 73019, USA
c
The First Affiliated Hospital of Medical School of Zhejiang University, Hangzhou, China
d
Department of Radiology, Mayo Clinic in Arizona, Scottsdale, AZ, 85259, USA

a r t i c l e i n f o a b s t r a c t

Article history: Breast cancer is the second leading cause of cancer death among women worldwide. Nevertheless, it is
Received 15 April 2018 also one of the most treatable malignances if detected early. Screening for breast cancer with full field
Received in revised form 28 August 2018 digital mammography (FFDM) has been widely used. However, it demonstrates limited performance
Accepted 13 September 2018
for women with dense breasts. An emerging technology in the field is contrast-enhanced digital mam-
mography (CEDM), which includes a low energy (LE) image similar to FFDM, and a recombined image
Keywords:
leveraging tumor neoangiogenesis similar to breast magnetic resonance imaging (MRI). CEDM has shown
Deep learning
better diagnostic accuracy than FFDM. While promising, CEDM is not yet widely available across medical
Image synthesis
Breast tumor
centers. In this research, we propose a Shallow-Deep Convolutional Neural Network (SD-CNN) where
Digital mammography a shallow CNN is developed to derive “virtual” recombined images from LE images, and a deep CNN is
CEDM employed to extract novel features from LE, recombined or “virtual” recombined images for ensemble
Classification models to classify the cases as benign vs. cancer. To evaluate the validity of our approach, we first develop
a deep-CNN using 49 CEDM cases collected from Mayo Clinic to prove the contributions from recombined
images for improved breast cancer diagnosis (0.85 in accuracy, 0.84 in AUC using LE imaging vs. 0.89 in
accuracy, 0.91 in AUC using both LE and recombined imaging). We then develop a shallow-CNN using
the same 49 CEDM cases to learn the nonlinear mapping from LE to recombined images. Next, we use
89 FFDM cases from INbreast, a public database to generate “virtual” recombined images. Using FFDM
alone provides 0.84 in accuracy (AUC = 0.87), whereas SD-CNN improves the diagnostic accuracy to 0.90
(AUC = 0.92).
© 2018 Elsevier Ltd. All rights reserved.

1. Introduction cancer screening due to its relatively low detection sensitivity in


many subgroups of women. For example, although FFDM screening
Although about 1 in 8 U.S. women (∼12%) will develop inva- has an overall cancer detection accuracy of 0.75 to 0.85 in the gen-
sive breast cancer over the course of her lifetime (Breastcancer.org, eral population, its accuracy in several subgroups of the high-risk
2018), breast cancer death rate has been steadily and/or signif- women including those with positive BRCA (BReast CAncer) muta-
icantly decreasing since the implementation of the population- tion or dense breasts decreases to 0.30–0.50 (Elmore et al., 2005).
based breast cancer screening program in late 1970s due to the On the other hand, using dynamic contrast enhanced breast MRI can
early cancer detection and the improved cancer treatment meth- yield significantly higher cancer detection performance due to its
ods (Rosenquist and Lindfors, 1998). Among the existing imaging ability to detect tumor angiogenesis through contrast enhancement
modalities, full field digital mammography (FFDM) is the only clin- and exclude suspicious dense tissues(Warner et al., 2004). Yet, its
ically acceptable imaging modality for the population-based breast substantially higher cost, lower accessibility and longer imaging
cancer screening, while Ultrasound (US) and Magnetic Resonance scanning time forbids breast MRI being used as a primary imag-
Imaging (MRI) are also used as adjunct imaging modalities to mam- ing modality in breast cancer screening and detection. In addition,
mography for certain special subgroups of women (Lehrer et al., lower image resolution of breast MRI is a disadvantage as compar-
2012). However, using FFDM is not an optimal approach in breast ing to FFDM.
In order to combine the advantages of both FFDM and MRI, a new
novel imaging modality namely, contrast-enhanced digital mam-
mography (CEDM) emerges and starts to attract broad research and
∗ Corresponding author.
clinical application interest. CEDM is a recent development of digi-
E-mail address: [email protected] (T. Wu).

https://doi.org/10.1016/j.compmedimag.2018.09.004
0895-6111/© 2018 Elsevier Ltd. All rights reserved.
54 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62

tal mammography using the intra-venous injection of an iodinated models are established by experienced researchers using publicly
contrast agent in conjunction with a mammography examination. labeled image datasets. For a specific task, the model is often treated
Two techniques have been developed to perform CEDM exami- as a feature generator to extract features describing the images
nations: the temporal subtraction technique with acquisition of from abstract level to detailed levels. One can then develop clas-
high-energy images before and after contrast medium injection sification models (SVMs, ANNs, etc.) using the derived features.
and the dual energy technique with acquisition of a pair of low and Promising results have been reported in several medical appli-
high-energy images only after contrast medium injection. During cations, such as chest pathology identification (Bar et al., 2015),
the exam, a pair of low and high-energy images is obtained after breast mass detection and classification (Samala et al., 2016), just
the administration of a contrast medium agent. The two images are to name a few. While exciting, earlier CNN models such as AlexNet
combined to enhance contrast uptake areas and the recombined (Krizhevsky et al., 2012), GoogLeNet (Simonyan and Zisserman,
image is then generated (Fallenberg et al., 2014). In CEMD, it has 2014) and VGGNet (Szegedy et al., 2014) are known to suffer from
low energy (LE) imaging, which is comparable to routine FFDM and gradient vanishing when the number of layers increases signifi-
recombined imaging similar to breast MRI. Comparing to breast cantly. A newer model, ResNet (He et al., 2014) with a “short-cut”
MRI, CEDM exam is about 4 times faster with only about 1/6 the architecture is recently proposed to address the issue. The imag-
cost (Patel et al., 2017). In addition, CEDM imaging has 10 times ing competition results show the ResNet outperforms other CNN
the spatial resolution of breast MRI. Therefore, CEDM can be used models by at least 44% in classification accuracy.
to more sensitively detect small residual foci of tumor, including The potentials CNN brings to medical imaging research are not
calcified Ductal Carcinoma in Situ (DCIS), than using MRI (Patel limited to deep CNN for imaging feature extraction. A second area
et al., 2017). Several studies including prospective clinical trials that medical research can benefit is indeed using CNN for syn-
conducted at Mayo Clinic have indicated that CEDM is a promising thetic image rendering. Here an image is divided into a number of
imaging modality that overcomes tissue overlapping (“masking”) smaller patches fed into a CNN (e.g., 4-layer CNN in this research)
occurred in FFDM, provides tumor neovascularity related func- as the input and the output is a synthetic image. The CNN is trained
tional information similar to MRI, while maintaining high image to learn the non-linear mapping between the input and output
resolution of FFDM (Cheung et al., 2014; Fallenberg et al., 2014; images. Several successful applications have been reported, such as
Gillman et al., 2014; Luczyńska et al., 2014). Unfortunately, CEDM synthesizing positron emission tomography (PET) imaging (Li et al.,
as a new modality is yet widely available in many other medical 2014) or CT image (Han, 2017; Nie et al., 2017) from MRI image, and
centers or breast cancer screening facilities in the U.S. and/or across from regular X-ray to bone-suppressed recombined X-ray (Yang
the world limiting its broad clinical impacts. et al., 2017).
In clinical breast imaging (US, MRI, FFDM and CEDM), reading Motivated by this two-fold applicability of CNN, this research
and interpreting the images remains a difficult task for radiologists. proposes a Shallow-Deep CNN (SD-CNN) as a new CAD scheme to
Currently, breast cancer screening has high false positive recall tackle the unique problem stemmed from the novel imaging modal-
rate (i.e., ≥10%). Computer-aided detection (CADe) and diagnosis ity, CEDM, for breast cancer diagnosis. Our first hypothesis is that
(CADx) schemes (Tan et al., 2011; Carneiro et al., 2017; Gao et al., applying a deep CNN to CEDM is capable of taking advantage of
2016; Muramatsu et al., 2016) have been developed and demon- recombined imaging for improved breast lesion classification due
strated the clinical potentials to be used as “the second reader” to to the contribution from the tumor functional image features. Sec-
help improve radiologists’ performance in the diagnosis. In order ond, in order to expand the advantages of CEDM imaging modality
to overcome the limitation of lower accessibility to CEDM systems to the regular FFDM modality, we hypothesize that a shallow CNN is
and help radiologists more accurately conduct the diagnosis, this capable to discover the nonlinear mapping between LE and recom-
research proposes the development and validation of a new CADx bined images to synthesize the “virtual” recombined images. As a
scheme, termed Shallow-Deep Convolutional Neural Network (SD- result, traditional FFDM can be enriched with the “virtual” recom-
CNN). SD-CNN combines image processing and machine learning bined images. The objective of this study is to validate these two
techniques to improve the malignancy diagnosis using FFDM by hypotheses by using a unique study procedure and two imaging
taking advantages of information available from the CEDM. datasets of both CEDM and FFDM images. The details of the study
CNN is a feed-forward artificial neural network that has been procedures and experimental results are reported in the following
successfully implemented in the broad computer vision areas for sections of this paper.
decades (Lecun et al., 2015; LeCun et al., 1998). As it evolves,
different CNN models have been developed and implemented. 2. Materials
The computational resource and devices available in recent years
make the training of CNN with large number of layers (namely, In this research, two separate datasets are used, which include
the deep CNN) possible. Applying deep CNNs in image recogni- a dataset acquired from tertiary medical center (Mayo Clinic Ari-
tion was probably first demonstrated in ImageNet competition zona), and a public dataset from INbreast (Moreira et al., 2012).
(Russakovsky et al., 2015) back in 2012. Since then, it has become
a popular model for various applications ranging from natural lan- 2.1. Institutional dataset from Mayo Clinic Arizona
guage processing, image segmentation to medical imaging analysis
(Cha et al., 2016; Tajbakhsh et al., 2016). The main power of a Based on Institutional Review Board (IRB) approved study and
deep CNN lies in the tremendous trainable parameters in differ- data collection protocol, we reviewed CEDM examinations per-
ent layers (Eigen et al., 2013; Zeiler and Fergus, 2014). These are formed using the Hologic Imaging system (Bedford, MA, USA)
used to extract discriminative features at different level of abstrac- between August 2014 and December 2015. All patients undertaken
tion (Tajbakhsh et al., 2016). However, training a deep CNN often CEDM had a BI-RADS (Breast Imaging Reporting and Data Systems)
requires a large volume of labeled training data, which may not be (Liberman and Menell, 2002) rating of 4 and 5 in their original FFDM
easily available in medical applications. Secondly, training a deep screening images. Due to the detection of highly suspicious breast
CNN requires massive computational resources, as well as rigorous lesions, CEDM was offered as an adjunct test to biopsy in a clini-
research in architecture design and hyper-parameters tuning. To cal trial environment. All CEDM tests were performed prior to the
address these challenges, a promising solution is transfer learning biopsies. In summary, the patient cohort in this clinical trial had
(Banerjee et al., 2017), that is, a deep CNN model is trained fol- the following criteria: 1) the diagnostic mammogram was rated BI-
lowed by a task-specific parameter fine-tuning process. The trained RADS 4 or 5, and 2) histopathology test result was available from
F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62 55

Fig. 1. Example of breast images (Cancer and Benign) for LE and recombined (Rec) images with 2 views (CC and MLO) (Lesions are highlighted with green circle). (For
interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article).

surgical or image-guided biopsy. We limited the cohort to BIRADS system with pixel size of 70 mm (microns), and 14-bit contrast res-
4 and 5 lesions because the analysis required the gold standard of olution. For each subject, both CC and MLO view were available.
lesion pathology. 49 cases were identified that met the above inclu- For each image, the annotations of region of interests (ROIs) were
sion criteria, which include 23 benign and 26 cancer biopsy-proven made by a specialist in the field, and validated by a second special-
lesions. We analyzed one lesion per patient. If a patient had mul- ist. The masks of ROIs were also made available. In this research, a
tiple enhancing lesions, the annotating radiologist used the largest dataset of 89 subjects was extracted by including subjects that have
lesion to ensure best feature selection. In CEDM, there are cranial- BI-RADS scores of 1, 2, 5 and 6. Subjects with BI-RADS 1 and 2 are
caudal (CC) and mediolateral-oblique (MLO) views for both LE and regarded as benign tumor, and subjects with BI-RADS 5 and 6 are
recombined images. Fig. 1 illustrates the example views on the LE regarded as cancer. For each subject, images of CC and MLO view
and recombined images, respectively. are used for feature extraction. Examples from INbreast dataset is
For the 49 cases, all CEDM images with DICOM format were iden- shown in Fig. 2.
tified and transferred from the clinical PACS to a research database
and loaded into the open source image processing tool OsiriX 3. Methodology
(OsiriX foundation, Geneva, Switzerland) (Rosset et al., 2004).
DICOM images were anonymized and prepared for blinded read- To fully explore the advantages of CNNs and CEDM in breast
ing by a radiologist. A fellowship trained breast radiologist with cancer research, a Shallow-Deep CNN (SD-CNN) is proposed
over 8 years of imaging experience interpreted the mammogram (Fig. 3). First, we develop a Shallow-CNN from CEDM to discover
independently and used the OsiriX tool to outline lesion contours. the relationships between LE images and recombined images.
Contours were drawn on recombined images (both CC and MLO This Shallow-CNN is then applied to FFDM to render “virtual”
views) for each patient on recombined images. These contours were recombined images. Together with FFDM, a trained Deep-CNN is
then cloned onto LE images. All lesions were visible on both CC and introduced for feature extraction followed by classification models
MLO views. This information is further used in the imaging pre- for diagnosis. Note for CEDM, we can start the workflow with the
processing (see details in methodology section). Some examples Deep-CNN directly.
of LE and recombined images are shown in Fig. 1. As observed, LE
images are not as easy as recombined images to visualize the lesions 3.1. Image pre-processing
for both cancerous and benign cases.
Before the Deep-CNN and Shallow-CNN are employed, a four-
2.2. INbreast public dataset step imaging pre-processing procedure is launched. First, for each
image we identify a minimum-area bounding box that contains the
This dataset was obtained from INbreast, an online accessible tumor region. Specifically, for each tumor, we have a list of bound-
full-field digital mammographic database (Moreira et al., 2012). ary points with coordinates in pair (x,y) available. The bounding
INbreast was established by the researchers from the Breast Center box is decided using the (xmin , ymin ) and (xmax , ymax ) as the two
in CHJKS, Porto, under the permission of both the Hospital’s Ethics diagonal corner points to ensure the box covers the whole tumor
Committee and the National Committee of Data Protection. The area. Note we have CC and MLO views for FFDM and we have CC
FFDM images were acquired from the MammoNovation Siemens and MLO views for both LE and recombined images for CEDM. As
56 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62

Fig. 2. Example of breast images for FFDM images from INbrease dataset with 2 views (CC on left and MLO on right) (Lesions are highlighted with green circle). (For
interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article).

Fig. 3. Architecture of Shallow-Deep CNN.

a result, there are two images from FFDM and four images from imal preprocessing. Individual neurons respond to stimuli only in
CEDM. The bounding box size varies case by case due to differ- a restricted region of the visual field known as the receptive field.
ent sizes of tumors (ranging from 65 × 79 to 1490 × 2137 in this This process is simulated through different layers (convolutional,
study). Next, an enlarged rectangle that is 1.44 times (1.2 times in pooling, fully connected). A CNN’s capability is hidden behind the
width and 1.2 times in height) the size of bounding box is obtained. large amount trainable parameters which can be learned iteratively
The enlarged bounding box approach is to include sufficient neigh- through gradient descent algorithms. In this research, a 4-layer CNN
borhood information proved to increase the classification accuracy (Fig. 4) is implemented to model the latent relationship between
(Lévy and Jain, 2016). In the second step, this ‘enlarged’ rectangle the LE images (patches) and recombined images (patches). The
is extracted and saved as one image. The third step is to normalize model is then used to render “virtual” recombined images (patches)
the image intensity to be between 0 and 1 using the max-min nor- from FFDM images (patches).
malization. In the last step, the normalized images are resized to
224 × 224 to fully take advantage of trained ResNet model. Here we
3.3. Deep-CNN: feature generation
take the patches that contain tumor instead of the whole image as
input. This is because the focus of the study is on tumor diagnosis
ResNet is a trained deep CNN developed in 2015 with a revolu-
and we believe the features generated by the deep-CNN from the
tionary architecture using the “short-cut” concept in the building
tumor region shall better characterize the tumor, especially for the
block. As seen in Fig. 5, the output of building blocks takes both final
cases where the tumor region is small.
classification results and the initial inputs (the short-cut) when
updating the parameters. As a result, it outperforms traditional
3.2. Shallow-CNN: virtual image rendering deep-CNNs which are known to suffer from higher testing error
since gradient tends to vanish as the number of layers increases
Inspired by the biological processes (Elmore et al., 2005), CNNs (He et al., 2014). ResNet has different versions with 50, 101 and 152
use a variation of multilayer perceptions designed to require min- layers but all based on the same building blocks. In the ImageNet
F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62 57

Fig. 4. Architecture of 4-layer shallow-CNN for “virtual” recombined image rendering.

Gradient boosting trees (GBT) is one of the most powerful boost-


ing ensemble decision trees used in regression and classification
tasks (Yang et al., 2017). It builds the model in a stage-wise fash-
ion, and it generalizes them by allowing optimization of an arbitrary
differentiable loss function. The nature of GBT makes it robust to
overfitting by measuring the criterion it used when splitting the
tree nodes. In addition, it provides the importance of each feature
in the regression/classification for the ease of interpretation which
is desirable in the medical applications. In GBT, the feature impor-
tance is related to the concept of Gini impurity (Rokach and Oded,
2008). To compute Gini impurity (IG (p)) for a set of items with J
Fig. 5. Building blocks for traditional CNNs (left) and ResNet (right) (He et al., 2014).
classes, suppose i ∈ {1, 2, . . ., J}, and let pi be the fraction of items
labeled with i class in the set, we have:
Table 1
Number of features from each layer for one image.
J
IG (p) = 1 − p2i
i=1
Layer # 10 22 40 49
# of features 256 512 1024 2048 When constructing each decision tree in the boosting classifier,
a feature is used to divide the parent node into two children nodes
based on a threshold. Since the decision tree is constructed with the
competition, ResNet-50, ResNet-101 and ResNet-152 have compa- goal being to minimize the overall Gini impurity, the post-splitting
rable performances (top 5 error: 5.25% vs. 4.60% vs. 4.49%), but with Gini impurity shall be smaller than the pre-splitting Gini impurity.
quite different numbers of parameter (0.85 M vs. 1.7 M vs. 25.5 M). The reduced Gini impurity thus can be used to as a measure of the
For the consideration of balance between computation efficiency contribution from the feature in the process of splitting the tree. The
and accuracy, especially for the limited computation resources, we training procedure is to identify the optimal splitting features that
adopt ResNet-50 in this research. offer the maximum impurity reduction (Yang et al., 2017) among
In general, ResNet consists of four types of buildings blocks. The the whole feature set. The process of building trees serves as feature
CNN structures and the number of features for each block are shown selection and classification.
in Fig. 6. We mark them with different colors. For simplicity, let
blue for block type 1, orange for block type 2, purple for block type 4. Experiments and results
3 and green for block type 4. ResNet-50 is defined as [3, 4, 6, 3]
meaning that it has 3 type 1 blocks, 4 type 2 blocks, 6 type 3 blocks The overall objective of this research is to demonstrate the
and 3 type 4 blocks. The output features are extracted from the clinical utility of our novel SD-CNN approach for breast cancer
final layer of each block type, that is, layer 10, 22, 40 and 49. Since diagnosis. Therefore, we conduct two sets of experiments. The first
we have no prior knowledge about the feature performance, we experiment is to validate the values from recombined images for
decide to take the features from all four layers (10, 22, 40 and 49) improved breast cancer diagnosis. Deep CNN, ResNet is applied. The
for the classification model development. For each feature map, the second experiment is to investigate the feasibility of applying SD-
mean value is calculated and used to represent the whole feature CNN to enrich the traditional FFDM for improved diagnosis. A public
map. The number of features extracted from each layer is listed in FFDM dataset from INbreast is used and the results are compared
Table 1. For each view, we have 3840 (256 + 512 + 1024 + 2048) total with eight state-of-the-art algorithms.
features.
4.1. Experiment I: validating the improved accuracy in breast
3.4. Classification cancer diagnosis on CEDM using deep-CNN

Boosting is a machine learning ensemble meta-algorithm aim- The workflow of our first experiment is shown in Fig. 7. Using
ing to reduce bias and variance (Bauer et al., 1999). It converts 49 CEDM cases collected from Mayo Clinic Arizona, we first con-
weak learners to strong ones by weighing each training sample duct the experiments using LE from CEDM images. For each subject
inversely correlated to the performance of previous weak learners. in the dataset, LE images (both CC and MLO views) are processed
58 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62

Fig. 6. Architecture of ResNet [37] (Red star are placed in layers where features are extracted; Dotted shortcuts increase feature dimensions by zero-padding; based on the
output dimension of building blocks, the ResNet is divided into 4 different building blocks (BBs), they are shown with different colors in the figure (BB 1: blue (Fig. 5 Right),
BB 2: orange, BB 3: purple, BB 4: green). Different version of ResNets vary in the number of BBs, for instance, the 50-layer version ResNet has 3 BB 1 s, 4 BB 2 s, 6 BB 3 s and
3 BB 4 s). (For interpretation of the references to colour in this figure legend, the reader is referred to the web version of this article).

Fig. 7. Workflow of Experiment I.

through pre-processing procedure described in Section 3.1, after Table 2


Classification Performance of Experiment Using LE Images vs. LE and Recombined
which 2 patches (224 × 224) are extracted. They are fed into the
Images.
trained ResNet. As features from different layers of ResNet describe
the image from different scales and aspects, in this research, we LE images LE and Recombined images
have all the features fed into the GBT to classify the case as cancer Accuracy 0.85 0.89
vs. benign. The procedures are implemented with a python library Sensitivity 0.89 0.93
named “sklearn”. Different settings to prevent the model from over- Specificity 0.80 0.86
AUC 0.84 0.91
fitting are used. For example, we set maximum depth of individual
tree to be 3, use early stopping strategy by setting number of deci-
sion tree to be 21, max number of features to be searched for each
√ Table 3
split is N (N is the number of features), the minimal number of Contribution of features from different image sources.
samples falling in each leaf node is 2. Other settings are set to be
Image Source Number of features Contribution of
default.
impurity reduction
Next, we study the added values from recombined images for
improved diagnosis. Specifically, CC and MLO view from recom- LE image 56 76.84%
Recombined image 43 23.16%
bined images are fed into the same pre-processing and feature
generating procedure. The combination of LE and recombined
image features are used in the classification model. Performance
is measured based on leave-one-out cross validation to fully use image for each feature. The feature’s importance score is measured
the training dataset which is limited in size. Performance metrics through calculating the total impurity reduction when building the
are accuracy, sensitivity and specificity, and area under receiver ensemble trees. (Note that the feature importance is calculated
operating characteristic curve (AUC) (see Table 2). The ROC curves inside each leave-one-out loop, and the final result is the aver-
for two models are shown in Fig. 8. By using all the LE features gen- age for each feature among the loops). Table 3 summarizes the
erated by ResNet, we obtain the accuracy of 0.85 (Sensitivity = 0.89 importance scores for the features from different sources (LE vs.
Specificity = 0.80) and 0.84 for AUC. With additional features from Recombined Image). Here the scores are normalized by dividing
recombined image, the model accuracy is improved to 0.89 (Sensi- individual score with summation of all scores. From Table 3, we
tivity = 0.93 Specificity = 0.86) and AUC to 0.91. observe among all the 99 features used in the model, 56 are from
To explore the features contributing to the classification model, LE images which contribute 76.84% of the impurity reduction, 43
we calculated the contribution of each feature, and track the source features are from recombined images which contribute to 23.16%
F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62 59

Here we first develop a 4-layer shallow CNN that learns the non-
linear mapping between the LE and recombined images using the
same 49 CEDM dataset. CC and MLO view images are regarded as
separate training data, so a total of 98 images are used, in which 5
subjects (10 images) are selected as validation material, and the rest
44 subjects (88 images) are sued as training material. By randomly
extracting 2500 pair of training samples within masked tumor from
each LE (input) and recombined image (output), a training dataset
of 220,000 (88 × 2500) samples is generated. The input samples
for the CNN are 15 × 15 patches from LE images, the same input
size as in (Li et al., 2014). Considering the relatively small receptive
field and complexity of a shallow CNN, we set the output samples
size as 3 × 3, it is our intention to explore the impact of the differ-
ent output patch size for the breast cancer diagnosis as one of our
future tasks. The output patches from recombined image are cen-
tered in the same position as input patches from the LE image. The
input and output samples are fed into the CNN framework imple-
mented with package of “Keras”. The CNN has 2 hidden layers, with
10 7 × 7 filters in each layer. There are 5 K trainable parameters
through backpropagation with mini-batch gradient decent algo-
rithm to increase learning speed. Batch size is set to be 128. The
Fig. 8. Receiver operating characteristic curve for the model using FFDM image only learning rate is set to be 0.01, ReLu activation function is used in
vs. FFDM and recombined image.
all layers except the output layer, where activation function is not
used. Other parameters are set to be default by “Keras” package.
in the modeling. The features from the recombined images help Finally, with the trained CNN and patches extracted from avail-
improve the accuracy of breast cancer diagnosis from 0.85 to 0.89. able modality, we can construct a “virtual” recombined image by
assembling predicted patches into a whole image.
We use mean squared error (MSE) to evaluate the similarity
4.2. Experiment II: validating the value of “Virtual” recombined between the “virtual” recombined image and the true recombined
imaging in breast cancer diagnosis on FFDM using SD-CNN image for the 10 images in validation dataset. MSE measures the
pairwise squared difference in intensity as:
The improved performance by adding the features from recom-
bined images motivates us to study the validity of constructing and 1 N  2
MSE = TRecombined (i) − VRecombined (i)
using the “virtual” recombined images from FFDM images for breast N i=1
cancer diagnosis.

Fig. 9. Sample images of LE image, true recombined image and its corresponding virtual recombined in dataset I. (from left to right: benign, cancer, cancer, cancer).
60 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62

Fig. 10. Sample images of FFDM in dataset II and its corresponding “virtual” recombined (Two cases on left are cancerous with BI-RADS = 5, two cases on right are benign
with BI-RADS = 2).

Table 5
Contribution of features from different image sources.

Image Source Number of features Contribution of


impurity reduction

LE image 67 22.33%
Virtual 87 77.67%
Recombined image

With this trained shallow-CNN, we used the 89 FFDM cases from


INbreast dataset to render the “virtual” recombined images. Specif-
ically, for each subject, we slide the 15 × 15 window from left to
right, top to bottom (step size = 1) in FFDM image, to get the input
patches. The input patches are fed into the trained 4-layer CNN,
from which we get the predicted virtual recombined image patches
(3 × 3) as outputs. The small patches are placed at the same position
as their corresponding input patches in the “virtual” recombined
images. For the position with overlapping pixels, the values are
replaced with mean value for all overlapping pixels. At last, the
“virtual” recombined images are rendered. Fig. 10 illustrates some
Fig. 11. Receiver operating characteristic curve for the model using FFDM image example FFDM images and their corresponding “virtual” recom-
only verse FFDM and virtual recombined. bined images. One clinical advantage of recombined image is it
filters out dense tissues which often lead to false positive diagno-
Table 4 sis. As seen from Fig. 10, the “virtual” recombined images preserve
Classification Performance of Experiment Using FFDM Imaging vs. this advantage. Specifically, dense tissues surrounding tumors are
FFDM + Recombined imaging. excluded in “virtual” recombined images, making the core region
FFDM FFDM P value easier to be identified (left two cases in Fig. 10). For within benign
+ Virtual Recombined cases on the right, the suspicious mass (is mainly composed of
Accuracy 0.84 ± 0.09 0.90 ± 0.06 0.14
dense tissues) is mostly filtered out.
Sensitivity 0.81 ± 0.16 0.83 ± 0.16 0.91 Next, following the same procedure as the first experiment,
Specificity 0.85 ± 0.12 0.94 ± 0.04 < 0.05 we apply the ResNet on the FFDM alone, and on both FFDM and
AUC 0.87 ± 0.12 0.92 ± 0.14 0.28 “virtual” recombined images together. ResNet is used for feature
extraction followed by the GBT ensemble classifiers. The parame-
ter for GBT settings is further tuned since the training dataset is
Where N is total number of pixel in the selected patches, TRe- slightly imbalanced (benign: cancer = 30: 59). The training weights
combined(i) and VRecombined (i) are the intensity values for the for benign and cancer are set to be 1 and 0.5. Numbers of trees
same position in patches from the true recombined image and cor- set to be 31. Other parameter settings remain the same as the first
responding virtual recombined image. experiment and 10-fold cross validation is used. Fig. 11 shows the
For the 10 validation images, the MSE is 0.031 (standard devi- mean ROC curves for the model on FFDM alone vs. the model on
ation is 0.021). For illustration purpose, we choose four samples FFDM and the “virtual” recombined image. The mean AUC for the
to demonstrate the resulting “virtual” recombined images vs. the classifier using FFDM features is 0.87 ± 0.12, while after adding the
true recombined images (Fig. 9). As seen, the abstract features (e.g., features from virtual recombined image, the AUC is increased to
shape) and some details of the tumor from true recombined images 0.92 ± 0.14. It is interesting to observe from Fig. 11 that sensitivities
are restored by the “virtual” recombined images. (true positive rate) of the two models have similar performance, the
F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62 61

Table 6
Classification performance for using FFDM feature alone and using features from FFDM and “virtual” recombined and other state-of-the-art methods using INbreast dataset.

Methods Acc. AUC

Random Forest on features from CNN with pre-training 0.95 ± 0.05 0.91 ± 0.12
(Dhungel et al., 2017)
CNN + hand crafted features pre-training 0.91 ± 0.06 0.87 ± 0.06
(Dhungel et al., 2017)
Random Forest + hand crafted features pre-training 0.90 ± 0.02 0.80 ± 0.15
(Dhungel et al., 2017)
CNN without hand crafted features pre-training 0.72 ± 0.16 0.82 ± 0.07
(Dhungel et al., 2017)
Multilayer perceptron (Sasikala and Ezhilarasi, 2016) 0.88 0.89
Lib SVM (Diz et al., 2016) 0.89 0.90
Multi-kernel classifier (Augusto, 2014) NA 0.87
Linear Discriminant analysis (Domingues et al., 2012) 0.89 NA
Our proposed approach on FFDM only 0.84 ± 0.09 0.87 ± 0.12
Our proposed approach on both FFDM and “virtual” Recombined Image 0.90 ± 0.06 0.92 ± 0.14

specificities (1 – false positive rate) vary greatly. We want to high- we propose a SD-CNN (Shallow-Deep CNN) to study the two-fold
light the importance of specificity as breast cancer screening has applicability of CNN to improve the breast cancer diagnosis. One
high false positive recall rate (i.e., ≥ 10%). One known fact is that contribution of this study is to investigate the advantages of recom-
the probability that a woman will have at least one false positive bined images from CEDM in helping the diagnosis of breast lesions
diagnosis at 10 years screening program is 61.3% with annual and using a Deep-CNN method. CEDM is a promising imaging modal-
41.6% with biennial screening (Michaelson et al., 2016). This will ity providing information from standard FFDM combined with
lead to additional MR exams (extra cost) and even biopsy. Another enhancement characteristics related to neoangiogenesis (similar
side effect is the negative psychological impacts. In this research, to MRI). Based on our review of literature, no existing study has
the use of recombined images (“virtual” recombined images) shows investigated the extent of CEDM imaging potentials using the deep-
the great potential to address these challenges by improving the CNN. Using the state-of-art trained ResNet as a feature generator for
specificity. In Table 4, we summarize the model performances in classification modeling, our experiment shows the features from
terms of accuracy, sensitivity and specificity (threshold is set to be LE images can achieve accuracy of 0.85 and AUC of 0.84, adding
0.75). While we observe that the model on FFDM vs. the model on the recombined imaging features, model performance improves to
FFDM and “virtual” recombined image show no significant differ- accuracy of 0.89 with AUC of 0.91.
ences on accuracy, sensitivity and even AUC, the performance on Our second contribution lies in addressing the limited accessi-
specificity shows significant improvements (p < 0.05). bility of CEDM and developing SD-CNN to improve the breast cancer
In looking into the contributions from the features (Table 5), diagnosis using FFDM in general. This the first study to develop a
the use of “virtual” recombined imaging features improves the 4-layer shallow CNN to discover the nonlinear association between
performances in terms of both accuracy and AUC. Calculation of LE and recombined images from CEDM. The 4-layer shallow-CNN
contribution follows the same procedure as experiment I and is can be applied to render “virtual” recombined images from FFDM
conducted inside each cross-validation loop. Among all the 154 fea- images to fully take advantage of the CEDM in improved breast can-
tures used in this experiment, 87 are from the “virtual” recombined cer diagnosis. Our experiment on 89 FFDM dataset using the same
image, which contribute 77.67% of the total impurity reduction. trained ResNet achieves accuracy of 0.84 with AUC of 0.87. With
The rest 67 features are from LE images, and they contributed the the “virtual” recombined imaging features, the model performance
rest 22.33% impurity reduction. It is interesting to observe from is improved to accuracy of 0.90 with AUC of 0.92.
this experiment that the contributions from “virtual” recombined While promising, there is room for future work. First of all,
images are higher than the contributions from the true recombined the trained ResNet is a black-box feature generator, the features
images from the first experiment. One reason may be the second extracted may not be easy to be interpreted by the physicians. It
dataset has more denser tissue cases and it is believed recombined is our intention to discover possible clinical interpretations from
images shall be more useful in diagnosing the dense breast cases. the features as one of our future tasks. For example, as the ResNet
This is yet to be confirmed with the radiologists which is our imme- goes deeper, the initial layers of the features may represent the
diate next step. raw imaging characteristics as the first order statistics, the deeper
We further explore the state-of-the-art algorithms using the layer of the features may represent the morphological characteris-
same INbreast dataset and compare our methods against the eight tics (e.g., shape). This is yet to be explored. A second future work is
methods from the literature (Table 6). As seen, our approach using related to the patch sizes. We plan to assess impacts of the differ-
“virtual” recombined image outperforms six algorithms in terms of ent sized patches for both input and output images on the breast
both accuracy and AUC. We want to highlight that one of papers by cancer diagnosis.
Dhungel et al. (2017) proposes four approaches. Among the four,
the best performer has a 0.95 in accuracy and 0.91 in AUC, and the
Acknowledgements
second performer has a 0.91 in accuracy and 0.87 in AUC. We con-
clude our approach has better AUC (0.92) comparing to both while
We would like to acknowledge the ASU/Mayo Clinic collab-
inferior in accuracy (0.90). We contend that indeed, AUC is a more
orative research seed grant program for financial support and
robust metric in the medical research and it is considered to be
resources required for this research project.
more consistent and have better discriminatory power comparing
to accuracy (Huang and Ling, 2005).
References
5. Discussion and conclusion
Augusto, G.B., 2014. Multiple Kernel Learning for Breast Cancer Classification.
Bar, Y., Diamant, I., Wolf, L., Greenspan, H., 2015. Deep Learning with Non-medical
Differentiating benign cases from malignant lesions is one of Training Used for Chest Pathology Identification 94140V., http://dx.doi.org/10.
the remaining challenges of breast cancer diagnosis. In this study, 1117/12.2083124.
62 F. Gao et al. / Computerized Medical Imaging and Graphics 70 (2018) 53–62

Bauer, E., Kohavi, R., Chan, P., Stolfo, S., Wolpert, D., 1999. An empirical comparison Li, R., Zhang, W., Suk, H., Wang, L., Li, J., Shen, D., Ji, S., 2014. Deep Learning Based
of voting classification algorithms: bagging, boosting, and variants. Mach. Imaging Data Completion for Improved Brain Disease Diagnosis., pp. 305–312.
Learn. 36, 105–139, http://dx.doi.org/10.1023/a:1007515423169. Liberman, Laura, Menell, Jennifer H., 2002. Breast imaging reporting and data
Banerjee, Debrup, et al., 2017. A deep transfer learning approach for improved system (BI-RADS). Radiol. Clin. 40 (3), 409–430.
post-traumatic stress disorder diagnosis. In: Data Mining (ICDM), 2017, IEEE Luczyńska, E., Heinze-Paluchowska, S., Dyczek, S., Blecharz, P., Rys, J., Reinfuss, M.,
International Conference on, IEEE. 2014. Contrast-enhanced spectral mammography: comparison with
Breastcancer.org. (2018). U.S. Breast Cancer Statistics | Breastcancer.org. [online] conventional mammography and histopathology in 152 women. Korean J.
Available at: https://www.breastcancer.org/symptoms/understand bc/ Radiol. 15, 689–696, http://dx.doi.org/10.3348/kjr.2014.15.6.689.
statistics [Accessed 25 Sep. 2018]. Michaelson, J.S., Shih, Y.T., Walter, L.C., Church, T.R., Flowers, C.R., Lamonte, S.J.,
Carneiro, G., Nascimento, J., Bradley, A.P., 2017. Automated analysis of Wolf, M.D., Desantis, C., Lortet-tieulent, J., Andrews, K., Manassaram-baptiste,
unregistered multi-view mammograms with deep learning. IEEE Trans. Med. D., Saslow, D., Smith, R.A., Otis, W., 2016. Guideline Update from the American
Imaging 36, 2355–2365, http://dx.doi.org/10.1109/TMI.2017.2751523. Cancer Society, 314., pp. 1599–1614, http://dx.doi.org/10.1001/jama.2015.
Cha, K.H., Hadjiiski, L., Samala, R.K., Chan, H.-P., Caoili, E.M., Cohan, R.H., 2016. 12783.Breast.
Urinary bladder segmentation in CT urography using deep-learning Moreira, I.C., Amaral, I., Domingues, I., Cardoso, A., Cardoso, M.J., Cardoso, J.S., 2012.
convolutional neural network and level sets. Med. Phys. 43, 1882–1896, http:// INbreast: toward a full-field digital mammographic database. Acad. Radiol. 19,
dx.doi.org/10.1118/1.4944498. 236–248, http://dx.doi.org/10.1016/j.acra.2011.09.014.
Cheung, Y.C., Lin, Y.C., Wan, Y.L., Yeow, K.M., Huang, P.C., Lo, Y.F., Tsai, H.P., Ueng, Muramatsu, C., Hara, T., Endo, T., Fujita, H., 2016. Breast mass classification on
S.H., Chang, C.J., 2014. Diagnostic performance of dual-energy mammograms using radial local ternary patterns. Comput. Biol. Med. 72,
contrast-enhanced subtracted mammography in dense breasts compared to 43–53, http://dx.doi.org/10.1016/j.compbiomed.2016.03.007.
mammography alone: interobserver blind-reading analysis. Eur. Radiol. 24, Nie, Dong, Cao, Xiaohuan, Gao, Yaozong, Wang, Li., SD, 2017. Estimating CT Image
2394–2403, http://dx.doi.org/10.1007/s00330-014-3271-1. from MRI Data Using 3D Fully. Convolutional Net. 1, 170–178, http://dx.doi.
Dhungel, N., Carneiro, G., Bradley, A.P., 2017. A deep learning approach for the org/10.1007/978-3-319-46976-8.
analysis of masses in mammograms with minimal user intervention. Med. Patel, B.K., Garza, S.A., Eversman, S., Lopez-Alvarez, Y., Kosiorek, H., Pockaj, B.A.,
Image Anal. 37, 114–128, http://dx.doi.org/10.1016/j.media.2017.01.009. 2017. Assessing tumor extent on contrast-enhanced spectral mammography
Diz, J., Marreiros, G., Freitas, A., 2016. Applying data mining techniques to improve versus full-field digital mammography and ultrasound. Clin. Imag. 46, 78–84,
breast Cancer diagnosis. J. Med. Syst., 40, http://dx.doi.org/10.1007/s10916- http://dx.doi.org/10.1016/j.clinimag.2017.07.001.
016-0561-y. Rokach, L., Oded, Z., 2008. Maimon. Data mining with decision trees: theory and
Domingues, I., Sales, E., Pereira, W.C.A., Tec, I., Faculdade, P., Porto, U., Engenharia, applications, 69. World scientific.
P.De, Coppe, B., Janeiro, R.De, 2012. Inbreast-Database Masses Rosenquist, C.J., Lindfors, K.K., 1998. Screening mammography beginning at age 40
Characterization., pp. 1–5. years: a reappraisal of cost-effectiveness. Cancer 82, 2235–2240, http://dx.doi.
Eigen, D., Rolfe, J., Fergus, R., LeCun, Y., 2013. Understanding Deep Architectures org/10.1002/(SICI)1097-0142(19980601)82:11<2235::AID-CNCR19>3.0.CO;2-
Using a Recursive Convolutional Network., pp. 1–9. V.
Elmore, J.G., Armstrong, K., Lehman, C.D., Fletcher, S.W., 2005. CLINICIAN’ S Rosset, A., Spadola, L., Ratib, O., 2004. OsiriX: an open-source software for
CORNER screening for breast Cancer. J. Am. Med. Assoc. 293, 1245–1256. navigating in multidimensional DICOM images. J. Digit. Imaging 17, 205–216,
Fallenberg, E.M., Dromain, C., Diekmann, F., Engelken, F., Krohn, M., Singh, J.M., http://dx.doi.org/10.1007/s10278-004-1014-6.
Ingold-Heppner, B., Winzer, K.J., Bick, U., Renz, D.M., 2014. Contrast-enhanced Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., Huang, Z., Karpathy,
spectral mammography versus MRI: initial results in the detection of breast A., Khosla, A., Bernstein, M., Berg, A.C., Fei-Fei, L., 2015. ImageNet large scale
cancer and assessment of tumour size. Eur. Radiol. 24, 256–264, http://dx.doi. visual recognition challenge. Int. J. Comput. Vis. 115, 211–252, http://dx.doi.
org/10.1007/s00330-013-3007-7. org/10.1007/s11263-015-0816-y.
Gao, F., Zhang, M., Wu, T., Bennett, K.M., 2016. 3D small structure detection in Samala, R.K., Chan, H.-P., Hadjiiski, L., Helvie, M.A., Wei, J., Cha, K., 2016. Mass
medical image using texture analysis. 2016 38th Annu. Conf. Proc. IEEE Eng. detection in digital breast tomosynthesis: deep convolutional neural network
Med. Biol. Soc., 6433–6436, http://dx.doi.org/10.1109/EMBC.2016.7592201. with transfer learning from mammography. Med. Phys. 43, 6654–6666, http://
Gillman, J., Toth, H.K., Moy, L., 2014. The role of dynamic contrast-enhanced dx.doi.org/10.1118/1.4967345.
screening breast MRI in populations at increased risk for breast cancer. Sasikala, S., Ezhilarasi, M., 2016. Combination of Mammographic Texture Feature
Women’s Heal. 10, 609–622, http://dx.doi.org/10.2217/WHE.14.61. Descriptors for Improved Breast Cancer Diagnosis. Asian J. Infor.Technol. 15
Han, X., 2017. MR-based synthetic CT generation using a deep convolutional neural (20), 4054–4062.
network method. Med. Phys. 44, 1408–1419, http://dx.doi.org/10.1002/mp. Simonyan, K., Zisserman, A., 2014. Very Deep Convolutional Networks for
12155. Large-Scale Image Recognition., pp. 1–14, http://dx.doi.org/10.1016/j.infsof.
He, K., Zhang, X., Ren, S., Sun, J., https://doi.org/10.1.1.725.4861 2014. Delving Deep 2008.09.005.
Into Rectifiers: Surpassing Human-level Performance on ImageNet Szegedy, C., Liu, W., Jia, Y., Sermanet, P., Reed, S., Anguelov, D., Erhan, D.,
Classification. Vanhoucke, V., Rabinovich, A., arXiv:1409.4842 2014. Going Deeper With
Huang, Jin, Ling, Charles X., 2005. Using AUC and accuracy in evaluating learning Convolutions., http://dx.doi.org/10.1109/CVPR.2015.7298594.
algorithms. IEEE Trans. Knowl. Data Eng. 17 (3), 299–310. Tajbakhsh, N., Shin, J.Y., Gurudu, S.R., Hurst, R.T., Kendall, C.B., Gotway, M.B., Liang,
Krizhevsky, A., Sutskever, I., Hinton, G.E., 2012. ImageNet classification with deep J., 2016. Convolutional neural networks for medical image analysis: full
convolutional neural networks. Adv. Neural Inf. Process. Syst., 1–9, http://dx. training or fine tuning? IEEE Trans. Med. Imaging 35, 1299–1312, http://dx.doi.
doi.org/10.1016/j.protcy.2014.09.007. org/10.1109/TMI.2016.2535302.
Lecun, Y., Bengio, Y., Hinton, G., 2015. Deep learning. Nature 521, 436–444, http:// Tan, Maxine, et al., 2011. A novel computer-aided lung nodule detection system for
dx.doi.org/10.1038/nature14539. CT images. Med. Phys. 38 (10), 5630–5645.
LeCun, Y., Bottou, L., Bengio, Y., Haffner, P., 1998. Gradient-based learning applied Warner, E., Plewes, D.B., Hill, K.A., Causer, P.A., Jong, R.A., Cutrara, M.R., Yaffe, M.J.,
to document recognition. Proc. IEEE 86, 2278–2323, http://dx.doi.org/10.1109/ Messner, S.J., Meschino, W.S., Piron, C.A., Narod, S.A., 2004. Surveillance of
5.726791. BRCA1 and BRCA2 mutation. JAMA 292, 1317–1325.
Lehrer, D., Jong, R.A., Pisano, E.D., Barr, R.G., Mahoney, M.C., Iii, W.P.E., Larsen, L.H., Yang, W., Chen, Y., Liu, Y., Zhong, L., Qin, G., Lu, Z., Feng, Q., Chen, W., 2017. Cascade
Morton, M.J., Mendelson, E.B., Farria, D.M., Cormack, J.B., Marques, H.S., Adams, of multi-scale convolutional neural networks for bone suppression of chest
A., https://doi.org/307(13):1394-1404 2012. Detection of Breast Cancer with radiographs in gradient domain. Med. Image Anal. 35, 421–433, http://dx.doi.
Addition of Annual Screening Ultrasound or a Single ScreeningMRI org/10.1016/j.media.2016.08.004.
toMammography inWomen With Elevated Breast Cancer Risk 307. Zeiler, M.D., Fergus, R., 2014. Visualizing and understanding convolutional
Lévy, D., Jain, A., 2016. Breast Mass Classification from Mammograms Using Deep networks arXiv:1311.2901v3 [cs.CV] 28 nov 2013. Comput. Vision–ECCV 8689,
Convolutional Neural Networks. 818–833, http://dx.doi.org/10.1007/978-3-319-10590-1 53.

You might also like