Combined_Optic_Disc_and_Optic_Cup_Segmentation_Network_Based_on_Adversarial_Learning (1)
Combined_Optic_Disc_and_Optic_Cup_Segmentation_Network_Based_on_Adversarial_Learning (1)
ABSTRACT Glaucoma is a group of diseases characterized by progressive optic nerve damage, ultimately
resulting in irreversible visual impairment. Early diagnosis through color fundus photography, including
measurement of the vertical cup-to-disk ratio (CDR), can help prevent vision loss. The normal range of
CDR values is usually 0.3-0.5, and if it exceeds 0.6, then there may be some problems. However, asym-
metrical thinning at the edges of the bottom-superior temporal-nasal region and large gaps in datasets pose
challenges for existing automatic segmentation methods. To address these challenges, this paper proposes a
joint segmentation method for the optic disc (OD) and optic cup (OC) based on an adversarial network,
incorporating new monitoring functions to guide the network optimization process. The effectiveness
and stability of this framework were evaluated using two public performance datasets of retinal fundus
images, namely Drishti-GS and REFUGE. On the Drishti-GS dataset, our method achieved a score of
0.850/0.964/0.086, while on the REFUGE dataset, it obtained a score of 0.887/0.975/0.061. These results
indicate the effectiveness of our approach.
INDEX TERMS Adversarial learning, deep learning, color fundus photography, glaucoma.
of the apparent cup. Subsequently, the extracted image is The optic disc, also known as the optic nerve papilla,
transformed using polar coordinates to increase the propor- is located at the posterior pole of the eyeball, approximately
tion of optic disc and optic cup regions in the entire image, 3 millimeters nasally and 1.5 millimeters in diameter. The
thereby enhancing the accuracy of subsequent segmentation. optic disc nerve plays an essential role in the fundus exam-
The segmentation process employs adversarial learning net- ination. Early methods of extracting optic disc boundaries
works, which are adjusted in terms of network structure and relied on the use of templates. Lowell et al. [29] utilized
loss function to improve overall performance. To evaluate image gradient changes to segment optic disc boundaries
the effectiveness of the proposed method, experiments were and incorporated the active contour method. Since both optic
conducted on two public datasets (Drishti-GS and REFUGE). discs and optic cups have ellipsoid shapes, methods based
The results demonstrate that the model achieves the desired on circular transformation technology have also been used
outcomes. [2], [3]. To enhance the robustness of the model, Fu et al
[9] incorporated local texture features in a multidimensional
TABLE 1. Advantages and disadvantages of the above methods. space. Pixel classification methods have achieved notable
results in current semantic segmentation tasks, particularly
in the field of medical image segmentation. By converting
the boundary segmentation task into a pixel classification
problem, researchers have found it to be more conducive
to solving the problem. Cheng et al. [30] used a superpixel
classifier to segment the optic disc and optic cup, and manu-
ally produced visual features to improve detection accuracy.
Abramoff et al. introduced parallax values extracted from
stereo image pairs to identify the optic disc and background
[5]. Although these methods have shown good results, they
all rely on manual annotation information, and are therefore
more susceptible to image quality and pathological changes.
OC Segmentation: An important indicator for diagnosis
glaucoma is the optic cup, situated in the center of the optic
This work’s primary contributions are as follows: disc within a brighter oval depression. The anterior movement
(1) We explored supervised adversarial networks to of the optic cup obstructs the optic disc, leading to glaucoma.
improve the adaptability of different segmentation networks Under normal circumstances, the optic cup is less than 1/3
to different datasets and to improve the generalization ability the size of the optic disc, but the proportion of the optic
of networks. cup is larger patients with glaucoma. Wong et al. proposed a
(2) A method of polar coordinate transformation is pro- level-set algorithm to automatically segment the boundary of
posed to transform the image so that the accuracy of the final OC [31]. Later, the information about blood vessel curvature
segmentation task can be improved. in retinal images has been shown to be beneficial for the seg-
(3) We evaluated our model on two public fundus image mentation of OC [32]. Due to the natural distortion of fundus
datasets, and achieved good results in OD and OC joint blood vessels near the optic disc (OC) boundary, the accu-
segmentation tasks. racy of OC segmentation based on the information of blood
The remainder of this paper is organized as follows. vessel distortion information is not satisfactory. In addition,
We review related techniques in Section II, and Section III Cheng et al. introduced the method of pixel classifier method
introduces the proposed method in detail. The evaluation and into the OC segmentation task [30]. More and more useful
results are presented in Section IV. Finally, we discuss the methods are being introduced to OC segmentation tasks [33],
results and draw conclusions in Section V and Section VI. [34]. All of the aforementioned methods depend on manually
labeled visual features, primarily focusing on the contrast
II. RELATED WORKS information between the edge of the optic nerve and the optic
Nowadays, many researchers are engaged in research on cup.
task segmentation of OD and OC, and many methods are Joint OD and OC Segmentation: Optic disc and optic cup
effective. These methods rely heavily on visual features of are closely related in physiological structure, and the optic
artificial markers for segmentation, such as image gradient cup is contained in the optic disc, which means that the pixels
information, features of stereoscopic image pairs, local tex- belonging to the optic cup also belong to the optic disc. The
ture features and superpixel based classifiers. The boundary joint segmentation of the optic disc can obtain high accuracy
between OC and OD is often difficult to distinguish, so OC in calculating the value of CDR [35]. Joshi et al. divided OD
segmentation is more difficult, and it relies more on man- and OC step by step [36]. Zheng et al. integrated the prior
ually annotated features. In recent years, it has been found graph cut framework into OD and OC segmentation [37]. The
that the joint segmentation of OD and OC can improve the above methods are based on the fact that any pixel in the
performance of segmentation networks [28]. fundus image only belongs to one part, such as background,
104900 VOLUME 12, 2024
Y. Liu et al.: Combined OD and OC Segmentation Network Based on Adversarial Learning
A. ROI EXTRACTION
Increasing the proportion of optic disc and optic cup in the
FIGURE 4. The proposed segmentation network architecture. It includes a
whole image can help improve the accuracy of segmentation. down-sampling part, an up-sampling part and a skip connection part.
For this purpose, we use the ROI extraction framework to And it includes the ASPP module.
Joint Morphology Loss: We used new functions to guide By constantly updating the parameters of the segmentation
the learning of the network. A dice coefficient loss LDL network and discriminator, better segmentation results can be
and a smoothness loss LSL are make up this compound loss obtained. We use formula 5 as the objective function of the
function. We can express it as: discriminator:
Lseg = λ1 LDice (pd , yd ) + λ2 LDice (pc , yc )
X
LD = − zlog(D(S(IcS ))) + (1 − z)log(1 − D(S(IcT )))
h i m,n
+ λ3 LSmooth (pd , yd ) + LSmooth (pc , yc ) (1) (5)
where pd , yd , represents OD prediction probability graph and where z = 1 if the patch prediction is from the training set,
binary true mask after polar coordinate transformation; pc and z = 0 if the patch prediction is from the testing set. In the
and yc represent the prediction probability graph of OC and process of segmentation, we use the supervision function
the binary ground true mask after polar transformation; by to ensure the reliability of the segmentation results, which
adjusting λ , the weight of each loss function can be changed. includes two parts and:
The dice coefficient loss [41] measures the overlap
between the prediction and ground truth, and is written as LS = Lseg (IcS ) + Ladv (IcT ) (6)
P X
2 i∈ pi · yi Ladv (IcT ) = − log(D(S(IcT ))) (7)
LDice (p, y) = 1 − P 2
P 2
(2) m,n
i∈ p + i∈ y
We used two public datasets, their training parts contain
where, represents all pixels in the image after polar coor-
the original images and the corresponding manual anno-
dinate transformation, p represents the predicted probability
tation images, in the optimization process of the network,
graph, and y represents the ground truth mask after polar
we can use the joint morphological perception to segment the
coordinate transformation. The smoothness loss is written as:
X X loss Lseg .
LSmooth (p, y) = Bi,j × yi × pi − pj (3)
i∈ j∈N4 (i)
( C. POLAR TRANSFORMATION FOR FUNDUS IMAGE
1 if yi = yj In order to improve the performance of our network, we use
Bi,j = (4)
0 otherwise the method of polar coordinate transformation. The image
The smoothness of the counter image can be improved by after transformation will make OD and OC have a certain
reducing the variation between adjacent pixels. Where N4 (i) constraint relationship in space, which accords with the actual
denote the four-connected neighbors of pixel i, p denote the situation of physiological structure. If we set the center of the
prediction and y denote ground truth. disk as O(u0 , v0 ), then any pixel P(u, v) in the original color
fundus image corresponds to pixel P′ (θ , r) in the transformed
image. Figure. 6 (C), if r is used to represent the radius to
the origin p, θ is used to represent the azimuth. We can make
polar coordinates correspond to Cartesian coordinates:
( ( p
u = r cos θ r = u2 + v2
⇔ (8)
v = r sin θ θ = tan−1 v/u
IV. RESULTS
Experiments and Results: To verify the validity of the pro-
posed method, we conducted experiments on two publicly
available datasets, Drishti-GS and REFUGE.
According to the experimental results, our approach has
demonstrated superior performance on the two publicly avail-
able datasets. As shown in Table 2, our method outperformed
other methods in the OD and OC joint segmentation task.
FIGURE 7. The visual examples of polar coordinate transformation on
Drishti-GS and REFUGE datasets. The top three lines are from the
Specifically, on the Drishti-GS dataset, our method achieved
Drishti-GS dataset, and the bottom three lines are from REFUGE datasets, a score of 0.850/0.964/0.086, while on the REFUGE dataset,
Where a) fundus image, b) ROI region c) the polar coordinate it obtained a score of 0.887/0.975/0.061. These results indi-
transformation corresponding to the ROI region, d) the polar
transformation corresponding to the label, d) the label for the ROI region. cate the effectiveness of our approach.
Figure 8 presents the segmentation results on the
Drishti-GS and REFUGE datasets. Based on these results,
The method is verified on an Intel i5-9400, 32G RAM, we can conclude that our method’s segmentation results
Nvidia 2080Ti GPU and Windows10 platform computer, (column C) are more accurate and closer to the gold stan-
which is implemented in Python language and the back-end is dard compared to U-Net (column D). This indicates that
Tensorflow. First, we crop the original image to 480×480 size our approach can provide more reliable technical support to
with the optic disc as the center, and pass this region to the clinicians.
network as the ROI region for the next operation. According In the experiment, we also calculated the CDR value
to the previous experience, too small datasets tend to cause to measure the role of the proposed method in glaucoma
overfitting of the network. Simple data enhancement, such as screening. The results of ORIGA and REFUGE dataset are
translation, rotation and other operations, can not improve the shown in columns 4 and 8 of Table 1. From the glaucoma
network overfitting problem. screening results, we have the following observations: Most
To address this issue, polar coordinate transformation is FCN-based segmentation networks rely heavily on pixel-
utilized to enhance the performance of the network. The level labels, resulting in unsatisfactory results. However, GNS
process is depicted in figure 7. During the training phase, the network has some advantages in network structure, which
Adam optimizer iteratively optimizes the network’s parame- can collect high latitude features and finally make the results
ters. The initial learning rate is set at 1e-3, which is adjusted better. As shown in figure. 9, the ROC values on Drishti-GS
in subsequent iterations. The discriminator network is also (AUC=0.8590) and REFUGE datasets (AUC=0.8788).
continuously optimized during training, with a simulation Ablation Experiments: Ablation experiments were con-
learning rate set at 2e-5 and 1e-5, respectively. The learning ducted on the Drishti-GS datasets. The result achieved by
rate is reduced by 10 times every 4 learning periods, for a total different components of the model is shown in Table 3. For
of 400 learning periods. After obtaining the predicted mask, the sake of presentation, we take the U-net framework as
morphological operation is performed as a post-processing the baseline. When GAN network is combined with baseline,
step to refine the segmentation mask. a better dice score can be obtained in optic disc segmentation,
Evaluation Metrics: We use the following indicators to with an increase of 4.65%, but the optic cup segmentation get
evaluate our segmentation results, Dice coefficient (Dice), a lower score.
Jaccard (Jac) and vertical cup-to-disc ratio (CDR). The cri-
teria are defined as:
V. DISCUSSION
2 × Ntp Glaucoma diagnosis primarily depends on retinal images,
Dice = (9) particularly the optic disc and cup region in the center of
2 × Ntp + Nfp + Nfn
the image. However, existing methods suffer from limita-
Ntp
Jaccard = (10) tions such as differences in labels, datasets, and physiological
Ntp + Nfp + Nfn structure between the optic disc and cup. To address these
VDcup issues, we propose a generative adversarial network. Our
δ = CDRp − CDRg , CDR = (11)
VDdisc approach offers several improvements over existing methods:
FIGURE 9. The ROC curves with AUC scores for glaucoma screening based
on the segmentation results on Drishti-GS and REFUGE datasets.
FIGURE 10. Example of polar coordinate transformation, when choosing different R-values for the same image. From left to right, R=400,
500, 600, 700, 800.
results in the references. In the next work we will explore the [4] A. Chakravarty and J. Sivaswamy, ‘‘Joint optic disc and cup
validation of model performance across different data sets. boundary extraction from monocular fundus images,’’ Comput.
Methods Programs Biomed., vol. 147, pp. 51–61, Aug. 2017, doi:
10.1016/j.cmpb.2017.06.004.
VI. CONCLUSION [5] M. D. Abramoff, W. L. M. Alward, E. C. Greenlee, L. Shuba, C. Y. Kim,
In this study, we utilized an enhanced generative adversarial J. H. Fingert, and Y. H. Kwon, ‘‘Automated segmentation of the optic disc
from stereo color photographs using physiologically plausible features,’’
network to segment the optic disc (OD) and optic cup (OC) Investigative Opthalmol. Vis. Sci., vol. 48, no. 4, p. 1665, Apr. 2007, doi:
in color fundus data. By transforming the joint segmentation 10.1167/iovs.06-1081.
task into a multi-label segmentation task, we successfully [6] M. Hayat, S. Aramvith, and T. Achakulvisut, ‘‘SEGSRNet for stereo-
endoscopic image super-resolution and surgical instrument segmentation,’’
achieved the separation of these two structures. Our pro- 2024, arXiv:2404.13330.
posed framework comprises two components: a segmentation [7] M. Hayat, S. Aramvith, and T. Achakulvisut, ‘‘Combined channel
network and a discriminator. The segmentation network is and spatial attention-based stereo endoscopic image super-resolution,’’
in Proc. IEEE Region 10 Conf. (TENCON), Chiang Mai, Thailand,
designed to learn the conditional distribution between the Oct. 2023, pp. 920–925, doi: 10.1109/TENCON58879.2023.
fundus image and its corresponding label, while the discrim- 10322331.
inator is responsible for distinguishing between the source of [8] A. Krizhevsky, I. Sutskever, and G. E. Hinton, ‘‘ImageNet classification
the image-label pair. To ensure that the OD and OC infor- with deep convolutional neural networks,’’ in Proc. Adv. Neural Inf. Pro-
cess. Syst. (NIPS), 2012, pp. 1097–1105, doi: 10.1145/3065386.
mation is well balanced in the image, we first employed an [9] H. Fu, J. Cheng, Y. Xu, D. W. K. Wong, J. Liu, and X. Cao,
EOI network to isolate the region of interest, followed by ‘‘Joint optic disc and cup segmentation based on multi-label deep
a polar coordinate transformation for optimal performance. network and polar transformation,’’ IEEE Trans. Med. Imag.,
vol. 37, no. 7, pp. 1597–1605, Jul. 2018, doi: 10.1109/TMI.2018.
Our experimental results on two separate datasets indicate 2791488.
the viability of our approach. This method has the potential [10] Y. Jiang, L. Duan, J. Cheng, Z. Gu, H. Xia, H. Fu, C. Li, and J. Liu, ‘‘Join-
to aid clinicians in diagnosis and we plan to evaluate its tRCNN: A region-based convolutional neural network for optic disc and
performance on additional public datasets in future work. cup segmentation,’’ IEEE Trans. Biomed. Eng., vol. 67, no. 2, pp. 335–343,
Feb. 2020, doi: 10.1109/TBME.2019.2913211.
Our approach also holds valuable insights for other related [11] Z. Gu, J. Cheng, H. Fu, K. Zhou, H. Hao, Y. Zhao, T. Zhang, S. Gao,
fields. and J. Liu, ‘‘CE-Net: Context encoder network for 2D medical image
segmentation,’’ IEEE Trans. Med. Imag., vol. 38, no. 10, pp. 2281–2292,
Oct. 2019, doi: 10.1109/TMI.2019.2903562.
REFERENCES [12] Z. Wang, N. Dong, S. D. Rosario, M. Xu, P. Xie, and E. P. Xing, ‘‘Ellipse
[1] Y.-C. Tham, X. Li, T. Y. Wong, H. A. Quigley, T. Aung, and C.-Y. Cheng, detection of optic disc-and-cup boundary in fundus images,’’ in Proc. IEEE
‘‘Global prevalence of glaucoma and projections of glaucoma bur- 16th Int. Symp. Biomed. Imag. (ISBI), Apr. 2019, pp. 601–604.
den through 2040,’’ Ophthalmology, vol. 121, no. 11, pp. 2081–2090, [13] K. K. Maninis, J. Pont-Tuset, P. Arbeláez, and L. Van Gool, ‘‘Deep retinal
Nov. 2014, doi: 10.1016/j.ophtha.2014.05.013. image understanding,’’ in Proc. Int. Conf. Med. Image Comput. Comput.-
[2] A. Aquino, M. E. Gegúndez-Arias, and D. Marín, ‘‘Detecting the Assist. Intervent. Cham, Switzerland: Springer, 2016, pp. 140–148, doi:
optic disc boundary in digital fundus images using morphological, edge 10.1007/978-3-319-46723-8_17.
detection, and feature extraction techniques,’’ IEEE Trans. Med. Imag., [14] J. Long, E. Shelhamer, and T. Darrell, ‘‘Fully convolutional
vol. 29, no. 11, pp. 1860–1869, Nov. 2010, doi: 10.1109/TMI.2010. networks for semantic segmentation,’’ in Proc. IEEE Conf. Comput.
2053042. Vis. Pattern Recognit. (CVPR), Jun. 2015, pp. 3431–3440, doi:
[3] S. Lu, ‘‘Accurate and efficient optic disc detection and segmentation by 10.1109/CVPR.2015.7298965.
a circular transformation,’’ IEEE Trans. Med. Imag., vol. 30, no. 12, [15] K. Simonyan and A. Zisserman, ‘‘Very deep convolutional networks for
pp. 2126–2133, Dec. 2011, doi: 10.1109/TMI.2011.2164261. large-scale image recognition,’’ 2014, arXiv:1409.1556.
[16] J. Zilly, J. M. Buhmann, and D. Mahapatra, ‘‘Glaucoma detection using [35] Y. Xu, J. Liu, S. Lin, D. Xu, C. Y. Cheung, T. Aung, and T. Y. Wong, ‘‘Effi-
entropy sampling and ensemble learning for automatic optic cup and cient optic cup detection from intra-image learning with retinal structure
disc segmentation,’’ Computerized Med. Imag. Graph., vol. 55, pp. 28–41, priors,’’ in Proc. MICCAI, vol. 15, 2012, pp. 58–65, doi: 10.1007/978-3-
Jan. 2017, doi: 10.1016/j.compmedimag.2016.07.012. 642-33415-3_8.
[17] F. Ding, G. Yang, J. Wu, D. Ding, J. Xv, G. Cheng, and X. Li, [36] G. D. Joshi, J. Sivaswamy, and S. R. Krishnadas, ‘‘Optic disk and cup seg-
‘‘High-order attention networks for medical image segmentation,’’ in mentation from monocular color retinal images for glaucoma assessment,’’
Proc. Int. Conf. Med. Image Comput. Comput.-Assist. Intervent. Cham, IEEE Trans. Med. Imag., vol. 30, no. 6, pp. 1192–1205, Jun. 2011, doi:
Switzerland: Springer, 2020, pp. 253–262, doi: 10.1007/978-3-030-59710- 10.1109/TMI.2011.2106509.
8_25. [37] Y. Zheng, D. Stambolian, J. O’Brien, and J. Gee, ‘‘Optic disc and cup
[18] S. Wang, L. Yu, X. Yang, C.-W. Fu, and P.-A. Heng, ‘‘Patch-based output segmentation from color fundus photograph using graph cut with pri-
space adversarial learning for joint optic disc and cup segmentation,’’ ors,’’ in Proc. MICCAI, 2013, pp. 75–82, doi: 10.1007/978-3-642-40763-
IEEE Trans. Med. Imag., vol. 38, no. 11, pp. 2485–2495, Nov. 2019, doi: 5_10.
10.1109/TMI.2019.2899910. [38] A. Sevastopolsky, ‘‘Optic disc and cup segmentation methods for glaucoma
[19] S. Hong, J. Oh, H. Lee, and B. Han, ‘‘Learning transferrable knowledge detection with modification of U-Net convolutional neural network,’’ Pat-
for semantic segmentation with deep convolutional neural network,’’ in tern Recognit. Image Anal., vol. 27, no. 3, pp. 618–624, Jul. 2017, doi:
Proc. IEEE Conf. Comput. Vis. Pattern Recognit. (CVPR), Jun. 2016, 10.1134/s1054661817030269.
pp. 3204–3212. [39] O. Ronneberger, P. Fischer, and T. Brox, ‘‘U-Net: Convolutional net-
[20] J. Hoffman, D. Wang, F. Yu, and T. Darrell, ‘‘FCNs in the wild: Pixel-level works for biomedical image segmentation,’’ in Proc. 18th Int. Conf.
adversarial and constraint-based adaptation,’’ 2016, arXiv:1612.02649. Med. Image Comput. Comput.-Assist. Intervent., vol. 9351. Cham,
[21] Y.-H. Chen, W.-Y. Chen, Y.-T. Chen, B.-C. Tsai, Y.-C. F. Wang, and Switzerland: Springer, 2015, pp. 234–241, doi: 10.1007/978-3-319-24574-
M. Sun, ‘‘No more discrimination: Cross city adaptation of road scene 4_28.
segmenters,’’ 2017, arXiv:1704.08509. [40] R. O. Duda and P. E. Hart, ‘‘Use of the Hough transformation to detect
lines and curves in pictures,’’ Commun. ACM, vol. 15, no. 1, pp. 11–15,
[22] M. Wang and W. Deng, ‘‘Deep visual domain adaptation:
Jan. 1972, doi: 10.1145/361237.361242.
A survey,’’ Neurocomputing, vol. 312, pp. 135–153, Oct. 2018, doi:
[41] J. Hu, L. Shen, and G. Sun, ‘‘Squeeze-and-excitation networks,’’ in
10.1016/j.neucom.2018.05.083.
Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018,
[23] S. Kadambi, Z. Wang, and E. Xing, ‘‘WGAN domain adaptation for the
pp. 7132–7141, doi: 10.1109/CVPR.2018.00745.
joint optic disc-and-cup segmentation in fundus images,’’ Int. J. Com-
[42] A. Sevastopolsky, S. Drapak, K. Kiselev, B. M. Snyder, J. D. Keenan, and
put. Assist. Radiol. Surg., vol. 15, no. 7, pp. 1205–1213, Jul. 2020, doi:
A. Georgievskaya, ‘‘Stack-U-Net: Refinement network for improved optic
10.1007/s11548-020-02144-9.
disc and cup image segmentation,’’ Proc. SPIE, vol. 10949, Mar. 2019,
[24] J. Sivaswamy, S. R. Krishnadas, G. D. Joshi, M. Jain, and A. U. S. Tabish, Art. no. 1094928, doi: 10.1117/12.2511572.
‘‘Drishti-GS: Retinal image dataset for optic nerve head(ONH) segmen- [43] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, and L.-C. Chen,
tation,’’ in Proc. IEEE 11th Int. Symp. Biomed. Imag. (ISBI), Apr. 2014, ‘‘MobileNetV2: Inverted residuals and linear bottlenecks,’’ in
pp. 53–56. Proc. IEEE/CVF Conf. Comput. Vis. Pattern Recognit., Jun. 2018,
[25] F. Fumero, S. Alayon, J. L. Sanchez, J. Sigut, and M. Gonzalez-Hernandez, pp. 4510–4520, doi: 10.1109/CVPR.2018.00474.
‘‘RIM-ONE: An open retinal image database for optic nerve evaluation,’’ [44] M. Tan, B. Chen, R. Pang, V. Vasudevan, M. Sandler, A. Howard,
in Proc. 24th Int. Symp. Comput.-Based Med. Syst. (CBMS), Jun. 2011, and Q. V. Le, ‘‘MnasNet: Platform-aware neural architecture search for
pp. 1–6. mobile,’’ 2018, arXiv:1807.11626.
[26] J. I. Orlando et al., ‘‘REFUGE challenge: A unified framework for [45] S. Ioffe and C. Szegedy, ‘‘Batch normalization: Accelerating deep network
evaluating automated methods for glaucoma assessment from fundus pho- training by reducing internal covariate shift,’’ 2015, arXiv:1502.03167.
tographs,’’ Med. Image Anal., vol. 59, Jan. 2020, Art. no. 101570, doi: [46] S. Yu, D. Xiao, S. Frost, and Y. Kanagasingam, ‘‘Robust optic disc
10.1016/j.media.2019.101570. and cup segmentation with deep learning for glaucoma detection,’’
[27] M. Arjovsky, S. Chintala, and L. Bottou, ‘‘Wasserstein GAN,’’ 2017, Computerized Med. Imag. Graph., vol. 74, pp. 61–71, Jun. 2019, doi:
arXiv:1701.07875. 10.1016/j.compmedimag.2019.02.005.
[28] P. V. C. Hough, ‘‘Method and means for recognizing complex patterns,’’ [47] X. Ren, S. Ahmad, L. Zhang, L. Xiang, D. Nie, F. Yang, Q. Wang,
U.S. Patent 3 069 654, Dec. 3, 1962. and D. Shen, ‘‘Task decomposition and synchronization for semantic
[29] J. Lowell, A. Hunter, D. Steel, A. Basu, R. Ryder, E. Fletcher, biomedical image segmentation,’’ IEEE Trans. Image Process., vol. 29,
and L. Kennedy, ‘‘Optic nerve head segmentation,’’ IEEE Trans. Med. pp. 7497–7510, 2020, doi: 10.1109/TIP.2020.3003735.
Imag., vol. 23, no. 2, pp. 256–264, Feb. 2004, doi: 10.1109/tmi.2003. [48] M. K. Kar and M. K. Nath, ‘‘Efficient segmentation of vessels and
823261. disc simultaneously using multi-channel generative adversarial network,’’
[30] J. Cheng, J. Liu, Y. Xu, F. Yin, D. W. K. Wong, N.-M. Tan, D. Tao, Social Netw. Comput. Sci., vol. 5, no. 3, p. 288, Feb. 2024, doi:
C.-Y. Cheng, T. Aung, and T. Y. Wong, ‘‘Superpixel classification based 10.1007/s42979-024-02610-0.
optic disc and optic cup segmentation for glaucoma screening,’’ IEEE
Trans. Med. Imag., vol. 32, no. 6, pp. 1019–1032, Jun. 2013, doi:
10.1109/TMI.2013.2247770.
[31] D. W. K. Wong, J. Liu, J. H. Lim, X. Jia, F. Yin, H. Li, and T. Y. Wong,
‘‘Level-set based automatic cup-to-disc ratio determination using reti-
nal fundus images in ARGALI,’’ in Proc. 30th Annu. Int. Conf. IEEE
Eng. Med. Biol. Soc., Aug. 2008, p. 10, doi: 10.1109/IEMBS.2008.
4649648.
[32] D. W. K. Wong, J. Liu, J. H. Lim, H. Li, and T. Y. Wong, ‘‘Automated
detection of kinks from blood vessels for optic cup segmentation in
retinal images,’’ Proc. SPIE, vol. 7260, no. 6, pp. 964–970, 2009, doi:
YONG LIU was born in Henan, China, in 1986.
10.1117/12.810784.
He received the bachelor’s degree in electrical
[33] Y. Xu, D. Xu, S. Lin, J. Liu, J. Cheng, C. Y. Cheung, T. Aung,
engineering and automation from Yanshan Uni-
and T. Y. Wong, ‘‘Sliding window and regression based cup detec-
tion in digital fundus images for glaucoma diagnosis,’’ in Proc. versity, in 2009, and the master’s degree in cir-
Int. Conf. Med. Image Comput. Comput.-Assist. Intervent. Berlin, cuit and system from Henan Normal University,
Germany: Springer, 2011, pp. 1–8, doi: 10.1007/978-3-642-23626- in 2013. He is currently pursuing the Ph.D. degree
6_1. in control science and engineering with Wuhan
[34] Y. Xu, L. Duan, S. Lin, X. Chen, D. Wong, T. Wong, and J. Liu, ‘‘Optic cup University of Science and Technology.
segmentation for glaucoma detection using low-rank superpixel represen- His research interests include artificial intelli-
tation,’’ in Proc. Int. Conf. Med. Image Comput. Comput.-Assist. Intervent., gence, medical image processing, and intelligent
2014, pp. 1–8, doi: 10.1007/978-3-319-10404-1_98. control.
JIN WU (Member, IEEE) was born in Anhui, XUEZHI ZHOU was born in Henan, China,
China, in 1967. She received the degree in elec- in 1993. He received the B.S. degree in electronic
tronic information engineering from Huazhong information engineering from Henan Normal Uni-
University of Science and Technology, in 1988, versity, Xinxiang, China, in 2014, and the Ph.D.
the master’s degree in detection technology and degree in information and communication engi-
automatic equipment from the University of Sci- neering from Xidian University, Xi’an, China,
ence and Technology Beijing, in 1997, and the in 2020. His research interests include radiomics,
Ph.D. degree in pattern recognition and intelli- medical image analysis, and bioinformatics.
gent systems Huazhong University of Science and
Technology, in 2006.
She became the Director of the Department of Electronic and Information
Engineering, Wuhan University of Science and Technology, in 1998, and
the Associate Dean of the School of Information, Wuhan University of
Science and Technology, in 2007. Her research interests include image pro-
cessing, pattern recognition and intelligent systems, signal and information
processing, and multimedia communication.
Dr. Wu is a member of the Professional Committee of Image and Video
Processing and Communication of Chinese Society of Image and Graphics,
and a Senior Member of Chinese Institute of Electronics and Chinese Optical
Society.