Clinical Interpretable Deep Learning Model For Glaucoma Diagnosis
Clinical Interpretable Deep Learning Model For Glaucoma Diagnosis
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.
1406 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 24, NO. 5, MAY 2020
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.
LIAO et al.: CLINICAL INTERPRETABLE DEEP LEARNING MODEL FOR GLAUCOMA DIAGNOSIS 1407
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.
1408 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 24, NO. 5, MAY 2020
Fig. 7. Optic Disc Activation Mapping: the weights are mapped back
to the previous convolutional layer to generate the Evidence Activation
Maps (EAMs) as the attention score for glaucoma classification. There
are n feature maps in all of the feature maps. Correspondingly there
are n weights learned from the previous process. Weighted summation
of weights and feature maps are used to generate EAM. The EAM
highlights the glaucoma-specific discriminative regions.
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.
LIAO et al.: CLINICAL INTERPRETABLE DEEP LEARNING MODEL FOR GLAUCOMA DIAGNOSIS 1409
III. EXPERIMENT
The effectiveness of the proposed EAMNet is validated on
two aspects: the accuracy of glaucoma diagnosis and precision
of evidence identification. We perform experiments on the chal-
lenging public datasets ORIGA [34]. The experimental results
verify the proposed EAMNet achieves state-of-the-art diagnosis
accuracy (0.88) and does an excellent performance on evidence
identification.
In our experiments, the localization of lesions and segmenta-
tion of the optic disc are employed as an instance of evidence
identification for our clinical interpretable EAMNet. The patho-
genesis of glaucoma, structural changes of optical nerve head,
are often observed on the optic disc [1]. It is believed that when
judging a fundus image, whether it is glaucoma, doctors focus
Fig. 8. ROC curve of our method and other methods. Our method, the
mostly on the optic disc and the lesions on it. Thus, when a CNN red one, performs better than others.
model provides diagnosis result, meanwhile giving the evidence
map where the optic disc is, we are convinced this model is
clinically interpretable. In this implementation, we make use of
superpixel to soften the gradient of local features and employ
ellipse fitting to obtain the segmentation of optic disc. To the best
of our knowledge, no previous work sets a criterion to measure
the interpretability of the model.
A. Criteria
In this paper, we utilise the area under the curve (AUC) of
the receiver operation characteristic curve (ROC) to evaluate
the performance of glaucoma diagnosis. The ROC is plotted as Fig. 9. The activated map represents optic disc and cup areas simul-
a curve which shows the tradeoff between sensitivity (TPR) and taneously with different activation amplitude. The first column is the raw
fundus image, and the second and third columns are the activation map
specificity (TNR), defined as: and segmented optic disc mask by our EAMNet.
TP TN
TPR = , TNR = . (3)
TP + FN FP + TN
conducted over three years from 2004 to 2007 by the Singapore
where T P and T N are the numbers of true positives and true Eye Research Institute and funded by the National Medical Re-
negatives, and F P along with F N are the number of false search Council. Singapore Malay Eye Study (SiMES) examined
positives and false negatives, respectively. 3,280 Malay adults aged 40 to 80, from which, 149 are glaucoma
We utilize the overlapping error E and balance accuracy A as patients. Retinal fundus images for both eyes were taken for
the evaluation metrics for optic disc segmentation. each subject in the study [34]. The 650 images with manual
Area(S ∩ G) 1 labelled optic disc mask are divided into 325 training images
E =1− , A = (T P R + T N R) (4) (including 73 glaucoma cases) and 325 testing images (including
Area(S ∪ G) 2
95 glaucomas).
with 1) Ablation Study: As shown in Figs. 8 and 9, the ablation
TP TN study demonstrates that our method can not only obtain accurate
TPR = , TNR = (5)
TP + FN FP + TN glaucoma diagnosis but also provides the more transparent inter-
pretation by highlighting the distinct regions recognised by the
where S and G denote the segmented mask and the manual
network. In Fig. 9, the ROC curve (in red) indicates that although
ground truth, respectively.
the detection of glaucoma based on colour fundus image is
a challenging task, our EAMNet obtains high sensitivity and
B. Dataset low specificity. Thanks to the accurate evidence and multi-scale
The origa dataset is used in the experiments to validate glau- feature aggregation, EAMNet obtains a state-of-the-art AUC
coma diagnosis, disc segmentation and lesion localization. The value with 0.88. It is much higher than the traditional image
ORIGA datasets are comprised of 168 glaucoma and 482 normal processing methods like Airpuff, Wavelet, Gabor, and GRI.
images from studies of a Malay population with ground truth cup It also performs better than the superpixel and CNN method
and disc labels along with clinical glaucoma diagnoses. It was (Chan et al. [4]). The further analytical result will be shown in
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.
1410 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 24, NO. 5, MAY 2020
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.
LIAO et al.: CLINICAL INTERPRETABLE DEEP LEARNING MODEL FOR GLAUCOMA DIAGNOSIS 1411
TABLE II accurate glaucoma diagnosis (0.88 AUC) and optic disc seg-
CLASSIFICATION ON THE ORIGA VALIDATION SET mentation (0.9 Adisc and 0.278 Edisc). Here, EAMNet obtains
precise boundaries of the optic disc and accurate glaucoma di-
agnosis simultaneously since the accurate segmentation of optic
disc originates from accurate glaucoma diagnosis. In addition,
the accurate segmentation (even evidence identification) pro-
motes and verify the accuracy of glaucoma diagnosis. Compared
with state-of-the-art methods, our EAMNet achieves accurate
glaucoma diagnosis, meanwhile obtains high performance on
evidence activation.
As shown in Table III, the results show that EAMNet deals
effectively with the challenging task of optic disc segmenta-
tion, even though the pixel-level is unavailable. Noting that our
overfit the model. It can be found that as the overfitted accuracy method is worse than other methods in the task of optic disc
raises the segmentation accuracy drops. It can be proven that segmentation. It is because that we did not use any pixel-level
although it looks like two independent tasks, the optic disc labels, and there is much less supervision information in our task
segmentation and glaucoma diagnosis in a unified framework are than fully-supervised method. We only use the fully-supervised
strongly related. We are convinced that the segmentation of optic method for comparison. And the comparison results are only
disc is guided by the procedure of glaucoma diagnosis, while for reference to prove that our semi-supervised method is as ef-
the accurate glaucoma diagnosis is also promoted by effective fective as other methods. Although using the image-level labels,
segmentation of optic disc as evidence map. EAMNet performs closely to fully-supervised OD segmentation
2) Comparison Results: In this section, we compare the methods. This phenomenon indicates that the main pathological
results of proposed EAMNet with different types of CNN ar- area of glaucoma is located in the optic disc, which matches do-
chitectures and show that our EAMNet obtains the state-of- main knowledge of glaucoma. And considering intuition clinical
the-art performance on glaucoma diagnosis. Same as above, evidence of glaucoma, like CDR, closely related to the optic disc,
to quantify the evidence activation, we compare the results of it is interpretable when CNN activation map covers it.
optic disc segmentation which is generated by evidence acti-
vation maps with a generic and straightforward segmentation
IV. CONCLUSION AND FUTURE WORK
method. The matched methods are as follow. Gabor [23] and
wavelet [24] method use manual features with Support Vector In this paper, we propose a novel clinical interpretable Con-
Machine (SVM) classifier to get the diagnostic result.GRI [25] vNet architecture named EAMNet not only for accurate glau-
is a probabilistic two-stage classification method to extract the coma diagnosis but also for the more transparent interpretation
Glaucoma Risk Index (GRI) that shows a good glaucoma detec- by highlighting the distinct regions recognized by the network.
tion performance. Superpixel [26] method proposes optic disc The EAMNet solves the lack of interpretability of CNN-based
and optic cup segmentation using superpixel classification for glaucoma diagnosis CAD system. Beside diagnosing glaucoma
glaucoma screening. Chen et al. [10] and Zhao et al. [9] propose with high precision, the proposed EAMNet also gives an in-
two CNN method both of them have good accuracy. Meanwhile, terpretation for diagnosis. It presents the ability of weakly-
U-Net [32] and M-Net + PT [36] are optic disc segmentation supervised optic disc segmentation. And it activates the extract
method also using CNN. glaucoma lesions like bleeding, notch, PPA and structural vari-
In the experiment, the manual labels are adopted as the ground ation of blood vessels. The proposed EAMNet employed the
truth. 10-fold cross-validation method is used in the experiment. ResNet and M-LAP. It consists of 3 GAPs connecting to 3 layers
We divided all samples into ten parts, each containing equal of the different resolution increasing the resolution of EAM sig-
proportions of glaucoma and normal individuals. Each time nine nificantly. The result shows that this method makes classification
samples were used as training samples, and the remaining one performance primarily preserved. And an additional function
was used as a test sample. Finally, each result was averaged to of optic disc segmentation is attached. We have demonstrated
obtain the final diagnosis result. As shown in Tables II and III, that our system produces high accuracy diagnosis and optic disc
experimental results show that the proposed EAMNet achieves segmentation results on ORIGA dataset.
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.
1412 IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, VOL. 24, NO. 5, MAY 2020
Based on this work, limitations and open questions are drawn. [17] K. Blekas, D. I. Fotiadis, and A. Likas, “Greedy mixture learning for
High-resolution feature maps are hard to be represented by GAP. multiple motif discovery in biological sequences,” Bioinformatics, vol. 19,
pp. 607–617, 2003.
Besides, the optic cup is also important and related to glaucoma [18] K. Kourou, C. Papaloukas, and D. I. Fotiadis, “Modeling biological data
diagnosis. Further studies need to be carried out to design a through dynamic bayesian networks for oral squamous cell carcinoma
more empirical model to deal with the clear cup segmentation classification,” in Proc. World Congr. Med. Phys. Biomed. Eng., 2018,
pp. 375–379.
by weakly-supervised evidence exploring. [19] Q. Zhang et al., “Interpreting CNN knowledge via an explanatory graph,”
in Proc. Thirty-Second Assoc. Advancement Artif. Intell. Conf. Artif. Intell.,
REFERENCES 2017, pp. 4454–4463.
[20] C. M. Olthoff et al., “Noncompliance with ocular hypotensive treatment
[1] E. Dervisevic, S. Pavljasevic A. Dervisevic, and S. Kasumovic, “Chal- in patients with glaucoma or ocular hypertension: An evidence-based
lenges in early glaucoma detection,” Med. Arch., vol. 70, pp. 203–207, review,” Ophthalmology, vol. 112, pp. 953–961, 2005.
2016. [21] H. I. Suk and D. Shen, “Deep learning in diagnosis of brain disorders,”
[2] H A. Quigley and A T. Broman, “The number of people with glaucoma Recent Progress in Brain and Cognitive Engineering, pp. 203–213, 2015.
worldwide in 2010 and 2020,” Brit. J. Ophthalmology, vol. 90, pp. 262– [22] R. Zhao et al., “Weakly-supervised simultaneous evidence identification
267, 2006. and segmentation for automated glaucoma diagnosis,” in Proc. Thirty-
[3] M. C. Leskea, A. Heijl, L. Hyman, B. Bengtsson, and E. Komaroff, Third Assoc. Advancement Artif. Intell. Conf. Artif. Intell., Jul. 2019,
“Factors for progression and glaucoma treatment: The early manifest vol. 33, no. 01, pp. 809–816.
glaucoma trial,” Current Opinion Ophthalmology, vol. 15, pp. 102–106, [23] U. R. Acharya et al., “Decision support system for the glaucoma using
2004. Gabor transformation,” Biomed. Signal Process. Control, vol. 15, pp. 18–
[4] W. Chan et al., “Analysis of retinal nerve fiber layer and optic nerve head 26, 2015.
in glaucoma with different reference plane offsets, using optical coherence [24] S. Dua, U. R. Acharya, P. Chowriappa, and S. V. Sree, “Wavelet-based
tomography,” Dig. World Core Med. J., 2006. energy features for glaucomatous image classification,” IEEE Trans. Inf.
[5] M. A. Fernandez-Granero et al., “Automatic CDR estimation for early Technol. Biomedicine, vol. 16, no. 1, pp. 80–87, Jan. 2012.
glaucoma diagnosis,” J. Healthcare Eng., vol. 2017, pp. 1–14, 2017. [25] R. Bock, J. R. Meier, and L. G. Nyl, “Glaucoma risk index: Automated
[6] M. H. Tan et al., “Automatic notch detection in retinal images,” in Proc. glaucoma detection from color fundus images,” Med. Image Anal., vol. 14,
IEEE Int. Symp. Biomed. Imag., 2013, pp. 1440–1443. pp. 471–481, 2010.
[7] S. Vernon, “How to screen for glaucoma,” Practitioner, vol. 239, pp. 257– [26] J. Cheng et al., “Superpixel classification based optic disc and optic cup
260, 1995. segmentation for glaucoma screening,” IEEE Trans. Med. Imag., vol. 32,
[8] J. B. Jonas et al., “Ranking of optic disc variables for detection of no. 6, pp. 1019–1032, Jun. 2013.
glaucomatous optic nerve damage,” Investigative Ophthalmology Vis. Sci., [27] Z. Wang et al., “Zoom-in-Net: Deep mining lesions for diabetic retinopathy
vol. 41, pp. 1764–1773, 2000. detection,” in Proc. Int. Conf. Med. Image Comput. Comput.-Assisted
[9] R. Zhao et al., “Automatic detection of glaucoma based on aggregated Intervention, 2017, pp. 267–275.
multi-channel features,” J. Comput.-Aided Des. Comput. Graph., vol. 29, [28] S. Barratt, “InterpNET: Neural introspection for interpretable deep learn-
pp. 998–1006, 2017. ing,” Preprint, 2017, arXiv:1710.09511.
[10] B. Zou et al., “Classified optic disc localization algorithm based on [29] K. He et al., “Deep residual learning for image recognition,” in Proc. IEEE
verification model,” Comput. Graph., vol. 70, pp. 281–287, 2018. Conf. Comput. Vis. Pattern Recognit., 2015, pp. 770–778.
[11] X. Chen et al., “Glaucoma detection based on deep convolutional neural [30] B. Zhou et al., “Learning deep features for discriminative localization,” in
network,” Eng. Medicine Biol. Soc., pp. 715–718, 2015. Proc. IEEE Conf. Comput. Vis. Pattern Recognit., 2016, pp. 2921–2929.
[12] P. Junior, A. Sarmento, “A HW/SW embedded system for accelerating [31] J. Ngiam et al., “Sparse filtering,” in Proc. Int. Conf. Neural Inf. Process.
diagnosis of glaucoma from eye fundus images,” in Proc. Int. Symp. Syst., 2011, pp. 1125–1133.
Rapid Syst. Prototyping: Shortening Path Specification Prototype, 2016, [32] M. Oquab et al., “Is object localization for free? - Weakly-supervised learn-
pp. 12–18. ing with convolutional neural networks,” in Proc. IEEE Conf. Comput. Vis.
[13] T. J. Jun et al., “2sRanking-CNN: A 2-stage ranking-CNN for diagnosis of Pattern Recognit., 2015, pp. 685–694.
glaucoma from fundus images using CAM-extracted ROI as intermediate [33] C. Szegedy et al., “Going deeper with convolutions,” in Proc. IEEE Conf.
input,” Preprint, 2018, arXiv:1805.05727. Comput. Vis. Pattern Recognit., 2015, pp. 1–9.
[14] S. Yousefi et al., “Glaucoma progression detection using structural reti- [34] Z. Zhang et al., “ORIGA-light : An online retinal fundus image database
nal nerve fiber layer measurements and functional visual field points,” for glaucoma analysis and research,” in Proc. IEEE Conf. Eng. Med. Biol.
IEEE Trans. Biomed. Eng., vol. 61, no. 4, pp. 1143–1154, Apr. Soc., 2009, vol. 2010, pp. 3065–3068.
2014. [35] O. Ronneberger, P. Fischer, and T. Brox, “U-Net: Convolutional networks
[15] Z. Gao et al., “Motion tracking of the carotid artery wall from ultrasound for biomedical image segmentation,” in Proc. Int. Conf. Med. Image
image sequences: A nonlinear state-space approach,” IEEE Trans. Med. Comput. Comput.-Assisted Intervention, 2015, pp. 234–241.
Imag., vol. 37, no. 1, pp. 273–283, Jan. 2018. [36] H. Fu, J. Cheng, Y. Xu, D. W. K. Wong, J. Liu, and X. Cao, “Joint optic
[16] Z. Gao, H. Xiong, and X. Liu, “Robust estimation of carotid artery wall disc and cup segmentation based on multi-label deep network and polar
motion using the elasticity-based state-space approach,” Med. Image Anal., transformation,” IEEE Trans. Med. Imag., vol. 37, no. 7, pp. 1597–1605,
vol. 37, pp. 1–21, 2017. Jul. 2018.
Authorized licensed use limited to: PSG COLLEGE OF TECHNOLOGY. Downloaded on December 03,2024 at 04:18:08 UTC from IEEE Xplore. Restrictions apply.