A hybrid multi model artificial intelligence approach for glaucoma screening using fundus images

A hybrid multi model artificial intelligence approach for glaucoma screening using fundus images

Introduction

Glaucoma is an irreversible and progressive optic nerve disease, resulting in permanent loss of vision1. As life expectancy continues to increase, the number of glaucoma cases is on the rise with its prevalence surpassing that of conditions like diabetic retinopathy (DR) and age-related macular degeneration. This trend poses a significant public health challenge. Despite advancements in eye care, disparities in the availability of ophthalmologists, especially in resource-limited settings, highlight the need for scalable solutions2. Artificial intelligence (AI) and deep learning (DL) hold great promise in bridging this gap, particularly in automating the screening process for eye diseases.

AI technologies have already demonstrated significant success in automatic disease detection. Notably, the FDA-approved IDx-DR [Digital Diagnostics, Coralville, IA, USA, initially started as a project at the University of Iowa, under the name of the Iowa Detection Program (IDP)], and EyeArt AI systems have achieved significant milestones in diagnosing DR without requiring human interpretation3,4. These systems showcase AI’s transformative potential in eye care. However, glaucoma presents unique challenges that complicate the development of AI-based screening solutions5. Unlike DR, which relies primarily on fundus imaging, glaucoma diagnosis often requires a combination of information from fundus images, optical coherence tomography (OCT), intraocular pressure (IOP) measurements, and visual field (VF) testing6,7,8.

Glaucoma screening relies heavily on fundus imaging due to its cost-effectiveness, portability, and practicality in detecting glaucomatous features9. The screening process is conducted by analyzing the optic disc and surrounding areas [optic disc size and shape, cup to disc ratio (CDR), neuro-retinal rim, peripapillary atrophy, nasalization of blood vessels, presence of retinal nerve fiber layer defect (RNFLD), and optic disc hemorrhage (DH)]. Generally, a specialist can perform a qualitative assessment to detect glaucoma in less than 10 s. However, the initial stage known as pre-perimetric (PPG) is asymptomatic (minor/fewer changes in the retina). The optic nerve shows glaucomatous damage, but VF defects are not yet detectable by standard perimetric tests10,11,12,13,14. Therefore, sensitivity and specificity of VF testing in screening of early glaucoma is considerably low15. The PPG hinges on identifying minor structural changes in the retina8, which can challenge even experienced ophthalmologists. Factors such as DH and the extent of RNFL thinning are considered, but there is no consensus on a predictive model that accurately identifies progression risk for all patients, making PPG debatable in terms of diagnosis, progression, and treatment strategies. While technologies like OCT excel in detecting RNFLD, the identification of DH often requires fundus photography16,17,18,19,20. Nevertheless, detecting these early features during screening is crucial for identifying suspected glaucoma cases. Early detection enables timely intervention and close monitoring, thereby reducing the risk of blindness.

Real-world deployment of AI in glaucoma screening faces several hurdles, including variations in imaging modalities, differences in image quality, and population diversity5,21. While convolutional neural networks (CNNs)/DL models have shown excellent performance in binary classification (glaucoma/normal), their sensitivity to thresholds and overlapping image features limits their reliability, particularly for early-stage detection22,23,24,25,26,27,28,29,30,31. There are significant variations in reported results. For example, the area under the curve (AUC) varies from 71.8 to 99.7%, depending on the model used and the dataset’s ground truth definition. Typically, a binary classification model outputs a probability score between 0 and 1. The decision to classify a fundus image as either normal or glaucoma is based on the chosen threshold. When the features of glaucoma and normal images are well-separated, the model is likely to produce output probabilities that are confidently close to 0 or 1, leading to stable and good performance. However, the model’s performance becomes sensitive when there is significant overlap in image features between the two categories. As a result, the detection of moderate to advanced stages of glaucoma tends to be relatively accurate due to pronounced structural differences in fundus images. In contrast, performance becomes highly dependent on the chosen threshold, model architecture, and dataset in the early stages due to the subtle retinal changes10,11,12,13,14. Addressing these challenges requires advanced models capable of detecting minor retinal changes, which can be achieved by incorporating task-specific models.

Another challenge is the shortage of large, diverse clinical datasets for training AI models9. Variations in the characteristics of fundus cameras and image quality can significantly impact model performance. While transfer learning architectures like VGGNet, AlexNet, and GoogLeNet have been employed for image classification, these models are often optimized for non-clinical datasets (e.g., flowers or animals), limiting their effectiveness in glaucoma detection. Additionally, the high computational demands associated with deploying DL models pose obstacles to seamless integration with telehealth platforms. The lack of transparency in current AI models further complicates their adoption, as clinicians often require interpretable parameters to support decision-making. While a black-box approach can be highly effective for screening, an ideal scenario might involve a hybrid approach, where the algorithm not only provides predictions but also outputs critical parameters that assist in disease identification. This approach balances the strengths of AI-based screening with the need for transparency and trust in clinical practice.

Despite these challenges, macula-centered fundus images offer a promising pathway for glaucoma screening and the detection of other eye diseases beyond glaucoma15,32. These images provide a broader retinal view, enabling analysis of both the optic disc and macula. This perspective also facilitates the assessment of the RNFLD trajectory and provides the optic disc size index, which is essential for accurately interpreting the CDR. Although binary classification models trained on diverse fundus images can recognize disease-related features, they may overlook early indicators of glaucoma, such as RNFLD and DH, due to their emphasis on more prominent features like optic disc cupping. The automatic detection of RNFLD and DH17,33, by DL models remains challenging and is an underrepresented area in the existing literature.

Present studies have largely focused on estimating the optic disc parameters using segmentation models or predicting glaucoma through binary classification models. However, an integrated framework combining multiple diagnostic features is still lacking. To address this gap, we propose an AI-based glaucoma screening (AI-GS) network that acts as a virtual glaucoma specialist. Leveraging macula-centered fundus images, the AI-GS network predicts glaucoma while evaluating key optic disc features. By detecting multiple risk factors and integrating outputs from various models, this network enhances sensitivity to minor structural changes associated with glaucoma, significantly improving early detection during screening in comparison to binary classifiers. Moreover, the AI-GS network is designed to integrate seamlessly with telehealth platforms, advancing accessible and accurate glaucoma screening for timely intervention and prevention of vision loss.

Results

Architecture of AI-based glaucoma screening network

Figure 1 shows the block diagram of the AI-GS network. It consists of multiple DL models (multitask learning, segmentation, and classification models). The input to the models is a square-shaped fundus image (512 × 512 pixels), processed with contrast limited adaptive histogram equalization (CLAHE) method. The input images for DH segmentation and classification models are cropped from the original fundus image (512 × 512 pixels) around the optic disc as a region of interest (ROI). The output includes an image report along with a CSV file. The image report depicts boundaries of cup and disc, position of the macula, and location of DH (if detected). It also includes the numerical values of disc assessment, detected structural changes, glaucoma decision based on direct DL output and cupping analysis along with the final recommendations. The detailed disc assessment parameters include disc size index, disc size (large, small), center of the disc, cup and fovea, disc fovea angle, CDR (major axis, vertical, horizontal, area, and perimeter), area of the neuroretinal rim, circularity index of the cup and disc, and the decision probabilities of different models. The AI-GS network includes six lightweight DL models for tasks such as segmentation, image classification, and decision-making based on numerical features or probabilities. The combined memory size of all the models is ~112 MB. The layered architecture of the DL models used in the AI-GS network is inspired from our previously reported lightweight DL model (LWBNA-unet)34. This model was specifically designed for biomedical image segmentation and has several advantages over the traditional Unet model, which is popular for this task. Compared to Unet, LWBNA-unet is significantly lighter, with its weight reduced by a factor of 10. It achieves better segmentation accuracy and demonstrates improved reproducibility during model training. This model was specifically used for segmenting DH in fundus photographs. Furthermore, the LWBNA-unet model’s versatility allows for easy modification to create a DL model for binary classification and multi-task learning.

Fig. 1: Architecture of the AI-based glaucoma screening (AI-GS) network.
A hybrid multi model artificial intelligence approach for glaucoma screening using fundus images

The network integrates multiple DL models to extract glaucomatous features from fundus images, such as disc hemorrhage (DH) and retinal nerve fiber layer defects (RNFLD), and measure key parameters like the cup-to-disc ratio (CDR) and disc size index. These parameters are combined through a fully connected network (FCN) to estimate glaucoma probability, which is further refined using image-wide predictions—incorporating direct DL outputs, RNFLD, and DH—to generate the final screening decision, along with corresponding numerical values.

Full size image

Binary classification model based on LWBNA-unet

The architecture of the binary classifier used for identifying fundus images with the RNFLD and DH is shown in Fig. 2a. The encoder part and the Bottleneck Narrowing with Attention (BNA) module, i.e., the midblock, of the LWBNA-unet, were used34. The encoder takes a three-channel RGB image as input. In the first convolutional layer, the input image is expanded into a 16-channel feature map using 16 different filters. These filters detect low-level image features such as edges, textures, and patterns. In successive layers of the encoder, the number of channels (or filters) doubles, while the image size is halved. This enables the model to learn more complex and abstract features. At the end of the encoder, the initial matrix, which was 512 × 512 x 16 (width x height x channels) in the first layer, transforms into a feature matrix of size 16 × 16 x 256. The mid-block, or BNA module, further refines the extracted features and reduces the dimensions to the smallest size (16 × 16 x 8). The BNA module applies channel narrowing with attention, a technique that selectively focuses on the most relevant channels while discarding less informative ones. This process condenses the high-level image features into a more compact representation. The resulting multi-dimensional array (16 x 16 x 8) is then converted into a single-dimensional vector by flattening. Flattening stacks the values in the array row-by-row, creating a long continuous vector that retains the condensed high-level features in a linear format. Before reaching the binary output layer (the final layer that produces the classification prediction), the flattened vector is connected to fully connected (FC) and dropout layers twice. The FC layers help in learning complex relationships between the features, and dropout layers reduce overfitting by randomly disabling some neurons during training. The total number of model parameters is 1,718,770, which represents the weights and biases learned during the training process. Despite the large input image size, the model is relatively small at approximately 20.5 MB, achieved through channel narrowing with attention. This technique reduces the number of parameters while maintaining prediction accuracy, making the model more efficient and lighter.

Fig. 2: Architectures of lightweight models used in the AI-GS network.
figure 2

a Binary classifier based on LWBNA-unet model34, featuring the bottleneck narrowing with attention (BNA) module, which reduces the model’s parameter count without compromising prediction accuracy. b Layer details of the multi-task deep learning model based on LWBNA-unet (MTL_LWBNA-unet). c Feedforward fully connected network (FFCN), which combines numerical values and probabilities from the classification models to produce a binary output.

Full size image

Multi-task learning model based on LWBNA-unet

Segmentation and classification tasks are crucial in medical image analysis, and certain scenarios require the achievement of both objectives. For instance, DL can directly classify fundus images as glaucoma or normal. Traditionally, the CDR has been employed for this purpose, necessitating the segmentation of cup and disc regions in the fundus image. The LWBNA-unet offers an efficient and flexible solution. It can be tailored for multi-tasking, performing both segmentation and classification within a single model. Multi-tasking enhances computational efficiency, reduces model size, and enables end-to-end learning. Figure 2b shows the details of the layers in the LWBNA-unet-based multi-task learning model (MTL_LWBNA-unet). The segmentation output is a three-channel image corresponding to the disc (red), cup (green), and fovea (white). A new branch from the end of the bottleneck between the encoder and decoder paths of LWBNA-unet is created to output classification of image into glaucoma/normal categories. Like the classification model (Fig. 2a), the multi-dimensional array size is smallest (32 × 32 x 16) at this branching point. It is then converted to a single dimension vector by flattening and followed by FC and dropout layers twice before the binary classification output layer (Fig. 2b).

Model for numerical parameter-based classification

The AI-GS network also incorporates a DL model to make decisions using numerical values and probabilities derived from classification and segmentation models. For example, disc assessment parameters—such as vertical and horizontal CDR, disc size index, and cup and disc circularity index—determined from the segmented images of the cup, disc, and fovea are used to classify the fundus image as glaucoma or normal, i.e., a cupping-based decision. For this purpose, we used a simple feedforward FFCN (Fig. 2c). The model consists of five dense layers, each followed by a dropout layer for regularization. The number of neurons in each layer increases from 16 to 32, 64, 128, and then decreases to 64, 32, and 16. The final dense layer, which contains binary classification category neurons, uses the softmax activation function for multi-class classification.

Segmentation of cup, disc and fovea by LWBNA-unet

Figure 3a, b shows an example of a fundus image and its mask created using the Eq. (1) for the REFUGE Dataset (as described in the “Methods” section). The disc is marked with red, the cup in yellow, and the fovea in white. The average D obtained on the validation dataset for disc and cup after six rounds of training with the LWBNA-unet is 0.9582 ± 0.0204 and 0.8896 ± 0.0573, respectively. The best values of D for disc and cup are 0.9601 ± 0.0202 and 0.8895 ± 0.0609, respectively. The performance of the LWBNA-unet is very close to the best results reported for this dataset35. As described in the “Datasets” section, an extended dataset was created using Topcon’s fundus images using the trained LWBNA-unet model. Remarkably, the D for the cup, disc, and fovea were consistent with those obtained without incorporating the Topcon fundus images, differing only at the third and fourth decimal places. These outcomes indicate the robustness of the model for cup, disc, and fovea segmentation, making it suitable for generating masks for diverse fundus images.

Fig. 3: Example of input images and glaucoma screening results from binary classification models.
figure 3

a Macula centered fundus image. b Corresponding mask showing the optic disc (red + yellow), cup (yellow), and fovea (white) used for training the of MTL_LWBNA-unet model. c ROC curves comparing the best-performing MTL_LWBNA-unet and the cupping parameter based FFCN model. The MTL_LWBNA-unet model demonstrates superior glaucoma prediction performance compared to the cupping parameter-based approach.

Full size image

Glaucoma prediction and estimation of cupping parameters by MTL_LWBNA-unet

Training of the MTL_LWBNA-unet model utilized 2,953 fundus images of glaucoma and 5,195 normal images, gathered from various sources (as described in the “Dataset” section)36,37,38. The masks for the cup, disc, and fovea (Fig. 3a, b), corresponding to each fundus image, were generated by the trained LWBNA-unet model as described earlier. Five-fold cross-validation of the MTL_LWBNA-unet model was conducted. After training five times, ROC curves were plotted for each validation dataset. The average values along with the standard deviation (std) for the five-fold cross-validation testing at a threshold of 0.5 are as follows: prediction accuracy = 96.2812 ± 0.4451%, AUC = 0.9904 ± 0.0014, specificity = 0.9692 ± 0.0074 and sensitivity = 0.9516 ± 0.0151. The low std values suggest that the model’s performance is consistent, and likely to generalize well to unseen images. Figure 3c shows the ROC curve of the best-performing model among the five trained models. At an optimum threshold of 0.5245, sensitivity and specificity for the detection of glaucoma on the validation dataset by DL is ~ 96.19% and ~ 97.63%, respectively. Details of the results, along with the 95% CI, are given in Table 1. These results highlight the model’s effectiveness in detection of glaucoma, demonstrating its potential for clinical application.

Table 1 Glaucoma detection results for the validation, testing, and real-world screening datasets, using different DL models for classification and segmentation based on LWBNA
Full size table

Traditionally, glaucoma diagnosis relies on analyzing optic disc and cup features in fundus images. The present MTL_LWBNA-unet model segments the optic disc, cup, and fovea boundaries, enabling the estimation of size and shape for various fundus image features. We utilized several parameters for glaucoma detection, including the disc size index (disc-to-fovea distance/maximum disc diameter), major axis CDR, vertical CDR (VCDR), horizontal CDR (HCDR), area CDR (ACDR), perimeter CDR (PCDR), neuro-retinal rim area (NRRA)/disc area (DA) ratio, and circularity indices (CI) for both the disc and cup. These parameters were used to train a simple FFCN model, as shown in Fig. 2c. The ROC curve of the best-trained FFCN model from five-fold cross-validation is shown in Fig. 3c. The results obtained at the optimal threshold of 0.4638 are presented in Table 1. As anticipated, glaucoma detection using cupping parameters (sensitivity = 85.99% and specificity = 93.39%) is less effective than the DL-based approach.

To assess the precision of the MTL_LWBNA-unet model in segmentation and CDR estimation, we used a distinct dataset known as “Chaksu”39. Out of 500 images, we successfully calculated the VCDR, HCDR, and ACDR for 481 images (85 from Bosch, and 396 from Remidio cameras). We compared these results with data annotated by five experts in the Chaksu dataset. The dataset provides mean, median, and the majority values for VCDR, HCDR, and ACDR as assessed by different experts. To compare the accuracy of automatic CDR estimation by MTL_LWBNA-unet with manual estimation by experts, we conducted Bland-Altman analysis—an approach that assesses agreement between two quantitative measurements. Figure 4 illustrates Bland-Altman plots comparing VCDR, HCDR, and ACDR estimated from the present model and majority values from Chaksu dataset experts. In each plot, the centerline represents bias, while the upper and lower lines depict limits of agreement (mean difference ± 1.96 times the std of differences). This range reflects the expected variability between the two methods (i.e., ~95% CI). The bias and std of differences in VCDR, HCDR and ACDR are 0.0256 ± 0.0552, −0.0020 ± 0.0806, and 0.0057 ± 0.0701, respectively. These values, approaching zero, indicate close agreement between MTL_LWBNA-unet estimates and manually estimated values from experts. To understand further the variability in measurements, we plotted the Bland-Altman plots of five experts with majority values in the Chaksu dataset, and the results are given in Table 2. Comparison of DL-based parameters with experts suggests that the current model outperforms some experts, with a higher percentage of data points falling within the 95% CI.

Fig. 4: Comparison of optic disc cupping parameters estimated by the MTL_LWBNA-unet model with those manually estimated by experts.
figure 4

Bland-Altman plots comparing the vertical cup-to-disc ratio (VCDR), horizontal cup-to-disc ratio (HCDR), and area cup-to-disc ratio (ACDR) automatically estimated by the deep learning model with expert-derived values from the Chaksu dataset. In each plot, the centerline represents the bias, while the upper and lower lines indicate the limits of agreement (mean difference ± 1.96 times the standard deviation of differences). Values near zero indicate close agreement between the MTL_LWBNA-unet model’s estimates and the experts’ manual estimates.

Full size image
Table 2 Results of Bland-Altman plots of five experts with majority values in the Chaksu dataset
Full size table

To test the classification accuracy of glaucoma and non-glaucoma cases, we created a balanced testing dataset of 8370 images. None of the fundus images in the testing dataset were part of MTL_LWBNA-unet training or validation datasets. Figure 3c shows the ROC curves illustrating glaucoma classification via direct model prediction and based on parameters estimated for cupping. The testing results are summarized in Table 1, highlighting that direct predictions by the MTL_LWBNA-unet significantly outperformed parameter-based predictions. Specifically, parameter-based predictions achieved a sensitivity of 70.8% at 95% specificity, while the DL approach attained 94.0% sensitivity, significantly exceeding the 85% sensitivity threshold recommended by Prevent Blindness America for glaucoma screening tools40. As noted in the introduction, a primary goal of this study is to detect PPG/suspected glaucoma cases. To achieve this, we segregated PPG, NTG, and POAG cases from the testing dataset to create specialized subsets. Glaucoma detection results are presented in Table 1. Predictions for NTG and POAG groups were comparable, while performance for the PPG group was notably lower. Performance was poorest in real-world scenarios like the Miyagi Screening, with sensitivities at 95% specificity as follows: 48.9% (Miyagi Screening), 83% (PPG), 94.2% (POAG), and 95.3% (NTG). A similar trend was observed in parameter-based predictions (Table 1). The markedly lower cupping-based sensitivity for PPG ( ~ 47% at 95% specificity) reflects minimal optic disc changes in early glaucoma stages. This highlights the importance of detecting subtle structural changes for accurate PPG detection. The model’s superior direct prediction accuracy likely stems from its ability to detect glaucomatous features beyond optic disc changes. Interestingly, the cupping-based sensitivity for Miyagi Screening at 95% specificity (~60%) exceeded that of the DL model. This suggests that ophthalmologists’ screening decisions may heavily rely on optic nerve head cupping along with other structural changes. Integrating specialized DL models that detect structural changes such as DH and RNFLD into a network could potentially enhance the performance of screening PPG and glaucoma in real-world scenarios.

Detection of disc hemorrhage

Detecting DH is crucial for accurate glaucoma diagnosis. DH appears as circular or flame-shaped patches of bleeding around the optic disc. Patients with DHs face a higher risk of glaucoma progression, requiring closer monitoring and timely intervention. DHs have been identified as an independent risk factor for the conversion of ocular hypertension to POAG or the progression of open-angle glaucoma17,18,41. Automatically detecting DHs from fundus photographs remains challenging, as they have been resistant to DL-based detection. Notably, publicly available datasets for DH remain scarce. To address this, we created a dataset for training LWBNA-based binary classification model. Representative image examples are shown in Fig. 5a–c. The model input is a 512 × 512 pixels image cropped around the optic disc. We used the original image (Fig. 5a) and its intensity-equalized version, processed using the CLAHE algorithm (Fig. 5b). Binary classification model (Fig. 2a) was trained ten times, and the best-performing model was selected based on validation accuracy. Figure 5d shows ROC curves for both the validation and testing datasets, with a summary of results in Table 1. On the validation dataset, the top-performing binary DH classification model achieved an AUC of 0.8394, an accuracy of 78.56%, and a Youden index of 0.5647 at a threshold of 0.5093. On the more balanced testing dataset, consisting of single image per eye, the model achieved an AUC of 0.8723, an accuracy of 75.49% and a Youden index of 0.5879 at a threshold of 0.4373. The predictive accuracy of the DH classification model is notably lower than that of the glaucoma classification model. This discrepancy may be explained by the small size of DHs and their similarity in color to other retinal features, such as vessels and pigmentation.

Fig. 5: Automatic detection of disc hemorrhage (DH) in a fundus image.
figure 5

a Typical example of a DH-negative image cropped around the optic disc. b Intensity-equalized image using the CLAHE algorithm. c Appearance of DH at the inferior-temporal boundary of the optic disc, highlighting DH as a time-dependent phenomenon. d ROC curves for the validation and testing datasets. The results indicate that DH detection accuracy in binary classification improves significantly when the probable DH area is provided through segmentation, i.e., using a DH mask.

Full size image

To enhance the classification accuracy of DH, we can provide the algorithm with spatial information about possible DH location. This approach mimics an ophthalmologist’s process of searching for DH in fundus images. For instance, initial observations may focus on retinal pigmentation or identifying true DH + , especially when they are large. Additionally, localized vessel contrast may be considered. However, a careful examination of these locations can help in ruling out false DH+ cases. To provide spatial information to the algorithm, we can either create masks outlining possible DH+ areas through segmentation. Alternatively, we can use bounding boxes, similar to object localization in object detection models. Although segmentation requires more effort to generate masks, it provides a detailed characterization of DH, including size, precise location, shape, and other attributes. Therefore, we created a dataset of 1704 DH+ fundus images and their corresponding masks. The LWBNA-unet segmentation model was trained using dice loss, similar to the approach used for segmenting FAZ in OCTA images34. A typical example of the image and its corresponding mask input to the model is shown in the inset of Fig. 6. Despite the often-unclear boundaries of DH, the model’s training and validation curves are consistent and repeatable (Fig. 6). We achieved a significantly high D of 0.8291 ± 0.1576 for the validation dataset (296 images). In many cases, the DH segmentation produced by our model surpasses the precision of the ground truth masks used for training. This improvement is illustrated in Supplementary Figs. 1 and 2.

Fig. 6: Segmentation of the disc hemorrhage region in a fundus image.
figure 6

Training and validation curves of the LWBNA-unet model trained for DH segmentation. The inset shows a typical example of an input image and its corresponding DH mask. A high Dice coefficient indicates successful delineation of the DH area by the segmentation model.

Full size image

Figure 7 shows examples of DH boundaries detected and marked on the fundus ROI images for both DH+ and DH- cases. Both the location and area of the hemorrhage can be estimated. For DH+ cases, the marked hemorrhage area is highly precise, even when it is much smaller than the overall image area (Fig. 7a). The model can recognize both single and multiple DH+ events in images captured from various cameras (Fig. 7b, c). In complex situations, such as in Fig. 7d–f, the model may detect multiple DH+ events that appear true at first glance. For example, in Fig. 7d, the DH marked in the inferior-temporal (IT) region just outside the disc is real, whereas the one inside the disc is false and represents winding or jumbled blood vessels. This image is from the AIROGS dataset (TRAIN090753) and is categorized as referral glaucoma (RG). The observed DH is located at the edge of the RNFLD, a common site of hemorrhage in glaucoma. False DH+ situations may arise when the algorithm analyzes DH− images, as shown in Fig. 7e, f. The model may output either no DH regions or multiple small spots (a few pixels) resembling DH+ in color. In most cases, these regions are easily eliminated. However, in certain situations, such as blood vessel winding (Fig. 7d, f) or local pigmentation (Fig. 7e), which resemble DH+ in color, size, location, and shape, distinguishing them can be challenging. This reflects the real-world challenges faced by ophthalmologists, where experience plays a crucial role. Thus, segmenting potential DH+ areas alone is insufficient for DH classification.

Fig. 7: Examples of disc hemorrhage (DH) boundaries detected by the LWBNA-unet model.
figure 7

a DH at the superior-temporal (ST) region. b DH at the inferior-temporal (IT) region. c DH at both ST and IT regions. d DH at IT, with the marked area inside the disc not representing DH but a winding vessel. e Multiple markings, where the IT at the disc boundary likely represents real DH, but the ST marking is pigmentation. f A DH-negative case where a curved vessel resembles a hemorrhage.

Full size image

Therefore, the dataset used for training the DH classification model was reprocessed using the segmentation model to identify potential DH+ areas for retraining. The input to the model now consists of a six-channel image: three RGB channels for the original ROI image and three channels for the segmentation mask. The DH classification accuracy improved significantly, increasing from approximately 76% to 91%. The best-performing model (from ten training runs) achieved a specificity of ~94% and a sensitivity of 87% on the testing dataset, with an AUC of 0.9678 and a Youden index of 0.8108 at a threshold of 0.4004. Detailed results are summarized in Table 1. This improvement is likely due to the spatial information about DH provided to the classification model through the segmentation mask.

Detection of retinal nerve fiber layer defects

RNFLD may be the earliest sign of glaucoma, often appearing long before visual defects become noticeable13,14,42,43. Detecting RNFLD in a fundus image can be challenging due to the insignificant intensity differences between the affected and normal regions. RNFLD typically appears as an arc extending from the optic disc toward the fovea. Few reports exist on the detection of RNFLD44,45,46, and there is currently no publicly available dataset. Creating such a dataset is challenging and prone to inconsistent labeling, leading to errors in model generalization and discrepancies in results across the training, validation, and testing datasets. Therefore, we created a dataset of 3721 fundus images and conducted a five-fold cross-validation training of the LWBNA-based model shown in Fig. 2a. Typical examples of input images (RNFLD− and RNFLD+) are shown in Fig. 8a, b. For demonstration purposes, the presence of RNFLD is marked by arrows in Fig. 8b, highlighting an arcade-like darker contrast starting from the edge of the optic disc.

Fig. 8: Detection of retinal nerve fiber layer defects (RNFLD) in macula-centered fundus images.
figure 8

a Example of an RNFLD-negative image. b Example of an RNFLD-positive image. c ROC curves for five models trained using a five-fold cross-validation scheme. The details of the curves are shown in the inset. d Training and validation curves for the best-performing model. The best-performing model achieved high prediction accuracy, with an AUC of 0.9933, a Youden index of 0.9182, and an accuracy of 95.97%.

Full size image

Figure 8c shows the ROC curves for five models trained with a five-fold cross-validation scheme. The average prediction accuracy, AUC, and Youden index are 91.71 ± 3.21%, 0.9731 ± 0.159, and 0.7990 ± 0.0898, respectively. Although some variations exist between folds, they are not statistically significant. The best-performing model achieved a high prediction accuracy, AUC, and Youden index of 95.97, 0.9933, and 0.9182, respectively (Table 1). The training and validation curves for this model are displayed in Fig. 8d. The training loss consistently decreases, while the validation loss stabilizes and then begins to increase with increasing epochs. Differences in training and validation accuracy across epochs may be attributed to variations in labeling. Overall, the model’s training behavior on the created dataset appears consistent, and the detection results for RNFLD are promising.

Glaucoma screening by multi-model AI-GS network

The AI-GS network combines the results of the aforementioned models using an FFCN to make a final screening decision. The inputs to the FFCN model consist of predictions (0/1) from the MTL_LWBNA-unet, disc assessments for cupping based on the FFCN model, DH predictions, and RNFLD predictions. The dataset for training the final FFCN model was created by combining all fundus images used for training and validation of the MTL, DH, and RNFLD detection models. A fundus image is labeled as positive (a case for referral to an ophthalmologist) if any prediction from the four models—direct DL, cupping-based FFCN, DH, or RNFLD—is positive. With this labeling scheme, the dataset contains 5062 fundus images labeled as negative (i.e., no glaucoma referral) and 4867 as positive. Compared to the direct prediction of glaucoma by the MTL_LWBNA-unet, the AI-GS network shows a slight improvement in five-fold cross-validation accuracy (96.28 ± 0.45% vs. 97.47 ± 0.04%). Table 1 summarizes the performance of the best-performing AI-GS network. Considering the 95% CI, the results for the AI-GS network and direct predictions by the MTL_LWBNA-unet do not differ significantly across the overall testing dataset when using the optimal threshold determined by the Youden index. However, in the real-world screening scenario using the “Miyagi Screening“ dataset, the AUC increased from 87.8% to 92.5%, and sensitivity improved from 56.5% to 80.5%, with only minor changes in specificity (~91%).

To further understand the differences in direct glaucoma prediction between the single DL model (MTL_LWBNA-unet) and the AI-GS network, we performed Delong’s test47. This test assesses whether the differences in the AUCs between two ROC curves are statistically significant. The extremely low p values—1.7 ×10−9 on the Miyagi screening dataset and even smaller on the testing datasets (PPG, NTG, and POAG)—indicate highly significant differences. As presented in Table 1, the threshold values for each model were determined based on the Youden index, often considered the optimal cutoff for decision-making in diagnostic tests. To better understand the strengths and weaknesses of the current models, we analyzed and visualized sensitivity, specificity, and accuracy with respect to threshold levels. These analyses are shown in Fig. 9 for the entire testing dataset, Miyagi Screening, and separately for PPG, while corresponding curves for NTG, POAG, and the cupping-based dataset are included in Supplementary Fig. 3. It is noteworthy that these parameters vary with the threshold in the case of a single DL model-based binary classification (MTL_LWBNA-unet) but remain almost invariant over a wider range of thresholds for the AI-GS network. These results clearly suggest that the AI-GS network’s performance is robust and better suited for large-scale screening programs where consistent and reliable performance is crucial, irrespective of the disease stage or severity.

Fig. 9: Dependence of sensitivity, specificity, and prediction accuracy on the binary classification threshold for a single model (MTL_LWBNA-unet) and the multiple-model AI-GS network.
figure 9

a Results from the MTL_LWBNA-unet model for the testing dataset. b Results from the AI-GS network for the testing dataset. c MTL_LWBNA-unet results for the PPG dataset. d AI-GS network results for the PPG dataset. e MTL_LWBNA-unet results for the Miyagi screening dataset. f AI-GS network results for the Miyagi screening dataset. The results demonstrate that the AI-GS network maintains consistent performance across a broader range of thresholds compared to the MTL_LWBNA-unet model.

Full size image

Discussion

Based on the obtained results, it is evident that the AI-GS network can detect glaucoma from macula-centered fundus images, evaluate disc features, and identify early structural changes. However, a definitive diagnosis of glaucoma cannot be made without VF testing. Since PPG is a stage where structural changes are present but no VF abnormalities are observed, this model facilitates the detection of PPG. This is a significant achievement, as early detection of PPG is crucial in the era of 100-year lifespans, contributing to timely diagnosis and intervention in glaucoma management. Additionally, the efficient design, based on the LWBNA-unet model34 minimizes computational overhead to nearly one-third of the size of the widely used Unet48 model employed in image segmentation tasks. This represents a major step toward integrating glaucoma screening into telehealth, making faster and more accurate eye care accessible. While many models for glaucoma detection from fundus images have been reported5,25,26,28,29,49,50,51,52, they often provide limited quantitative data. The AI-GS network not only offers screening but also delivers detailed metrics on the optic cup, disc, and fovea, supporting ophthalmologists in diagnosis. This quantitative information can help in clarifying the relationship between structural changes and disease progression, which may differ across populations. This study emphasizes the importance of assessing the neuroretinal rim (NRR) and optic cup relative to disc size in glaucoma detection.

In the current study, the AI-GS network aims to provide numerical data (disc and fovea assessments) and structural information (DH, RNFLD, and cupping), similar to that used in the labeling and diagnosis of glaucoma36,53. The MTL_LWBNA-unet model has demonstrated high accuracy in analyzing optic disc parameters from images captured by various cameras, with Ddisc = 0.9601 ± 0.0202 and Dcup = 0.8895 ± 0.0609 (Fig. 4). Thus, the AI-GS network effectively facilitates the analysis of correlations among various optic disc features and their vulnerability to glaucoma. To demonstrate this, the training, validation, and testing datasets of AIROGS (NRG: 5540, and RG: 1436) and THU (Normal: 2726, NTG: 2511, and POAG: 1357) were used. Figure 10a displays the correlation heatmap between various features of the fundus images, including p values to denote their statistical significance. The optic disc in a normal eye is slightly oval, and it is reported that axial length is significantly associated with disc ovality, thus correlating with myopia54. Based on the optic disc shape, we defined the “Myopic Factor,” or the disc ovality index, as the ratio of the disc’s major axis to its minor axis, which increases with the severity of myopia. Our results indicate that myopia negatively correlates with the disc circularity index (Disc CI), as expected (Fig. 10a). As anticipated, myopia does not affect the disc-fovea angle55. The superior and inferior neuroretinal rim areas (Ratio S area to Disc, and Ratio I Area to Disc) exhibit a positive correlation with myopia, while the temporal and nasal areas are negatively correlated. The strength of the negative correlation is weaker in the temporal area compared to the nasal area. These correlation patterns significantly weaken in cases of glaucoma (Fig. 10b), potentially due to the preferential (non-uniform) loss of nerve fibers affecting the NRR area. Like the Myopic Factor, correlation among other retinal features can be understood from the heatmaps in Fig. 10. Bootstrap analysis of the difference in correlation coefficients between normal and glaucoma groups provides insights into how variable relationships differ across these groups. A difference close to zero implies that glaucoma does not significantly alter the relationship between certain parameters. Conversely, a non-zero difference could reveal insights into the mechanisms of glaucoma-induced changes. Figure 10c shows the heatmap of the difference in correlation coefficients among normal and glaucoma groups in the AIROGS dataset. Interestingly, the correlation difference between the Myopic factor and disc circularity index is not significant, indicating that glaucoma does not alter this relationship. However, the relationship between the cup circularity index and the Myopic factor has changed markedly, possibly due to irregular loss of retinal nerve fibers. Glaucoma appears to modify the relationship of nasal NRR with other disc features (Fig. 10c). A parallel analysis using the THU dataset (shown in Supplementary Figs. 4–8) reveals that the correlation heatmaps for normal and glaucoma groups (NTG and POAG) are similar to the AIROGS dataset, but there are noteworthy variations in the heatmap of correlation coefficient differences among NTG and POAG groups. The difference heatmap of THU-POAG is like that of AIROGS, suggesting prevalence of POAG glaucoma type. It is reported that among overall glaucoma cases worldwide, approximately 74% are of the POAG type56. This may explain the similarity between the glaucoma and POAG groups in these datasets.

Fig. 10: Correlation coefficient heatmap among different optic disc features in AIROGS dataset.
figure 10

a Normal. b Glaucoma. c Difference between normal and glaucoma, analyzed using the bootstrap method. These heatmaps, derived from the AI-GS network, provide insights into the relationships between various structural features in fundus images and their associations with disease and population characteristics. Star marks indicate statistical significance (i.e., p values).

Full size image

Further insights into the population-based differences and risk factors can be gained by analyzing the histograms of various retinal features. For example, the mean Myopic factor for the AIROGS dataset in normal (NRG = 1.11) and glaucoma (RG = 1.10) groups is similar and close to the reported mean (1.09)57. The THU dataset has the same mean value for normal (1.10), but it is slightly higher in POAG ( = 1.12) and NTG ( = 1.14) groups. Histograms, along with their cumulative probabilities (Fig. 11a) reveal that the prevalence of myopia is relatively high in the THU dataset (i.e. Japanese population), and it increases further in glaucoma. The results of Fig. 11a suggest that high myopia could be a risk factor for developing glaucoma in the Japanese population, especially the NTG type. These findings are consistent with the reported literature58. It was hypothesized that, parallel to the myopic stretching and thinning of sclera, the optic disc is also stretched, leading to an abnormally large and irregularly configured optic nerve head. This might be one of several reasons for the increased risk of glaucoma59. An increase in the negative correlation strength of the myopic factor with the disc circularity index for glaucoma in the THU dataset appears consistent. Another noticeable difference between these two datasets is in the disc size index (the ratio of disc-fovea distance to maximum disc diameter) among normal and glaucoma groups. Based on Fig. 11b, the prevalence of relatively larger disc size is high for glaucoma in the AIROGS dataset, whereas the opposite is true for the THU dataset. Both larger and smaller optic discs have been reported as potential risk factors for glaucoma. The higher prevalence of glaucoma in certain African populations, which have larger optic discs, led to the hypothesis that larger optic discs are prone to glaucoma60. It is reported that the pressure differential across the lamina cribrosa could produce increased deformation and displacement of central tissues in larger discs, leading to greater glaucoma susceptibility. In contrast, a larger disc has a larger lamina cribrosa area and more pores than a small disc, providing more space for nerve fibers and possibly reducing the likelihood of focal compression to axons, resulting in a lower risk of glaucoma compared to smaller optic discs. Eyes with smaller optic discs also have a smaller number of nerve fibers61. Although studies suggest that disc size alone might not be the risk factor62, the AI-GS network can provide further insights through rapid analysis of large-sized datasets from different population groups.

Fig. 11: Cumulative probabilities obtained from histograms.
figure 11

a Myopic factor. b Disc size index for the AIROGS/Rotterdam dataset, including normal (RTD_N) and glaucoma (RTD_G) cases, as well as Tohoku University normal (THU_N), NTG (THU_NTG), and POAG (THU_POAG) cases. The results suggest that myopia and smaller optic disc size are risk factors for glaucoma in the Japanese population.

Full size image

Figures 10 and 11 highlight differences in optic disc features and their correlations in normal, POAG, and NTG groups. These differences can be utilized for diagnostic purposes. Some of the clinically investigated parameters include vertical and horizontal CDR, the NRR area as a whole, and measurements in various disc sectors (such as inferior, superior, temporal, and nasal, or combinations thereof)63. The inferior (I) and superior (S) regions of the optic disc are often the first and most severely affected in glaucoma. These regions are used by ophthalmologists to differentiate glaucomatous damage from other types of optic neuropathies. Monitoring changes in the NRR area, especially in the inferior and superior temporal (IT and ST) regions, can be used to assess the risk of developing glaucoma. We analyzed the ratio of the IT- and ST-NRR area to the disc area for both the AIROGS and THU datasets. Histograms with cumulative probabilities are presented in Fig. 12. The distribution of the area for normal and glaucoma groups follows an almost normal distribution in both IT and ST regions for the AIROGS dataset. As expected, the peak area ratio of the distribution for normal and glaucoma groups is larger in the IT region compared to the ST. A narrower distribution, indicated by full width at half maximum (FWHM), in the IT region compared to the ST suggests that glaucoma produces rapid changes in the IT region. This finding is consistent with the reported literature64,65. The THU dataset, which is divided into POAG and NTG types, shows some interesting differences. The distribution in the IT region is close to normal for both NTG and POAG, but it deviates in the ST. This deviation is weaker for POAG and is similar to the AIROGS dataset. The changes in the area ratio are slower (i.e., indicating a wider distribution) for NTG compared to POAG (Fig. 12d). In POAG, elevated IOP is a significant risk factor, and the damage typically follows the characteristic pattern of glaucomatous optic neuropathy. Both the IT and ST regions of the NRR are affected, with the inferior region often showing damage first or more prominently. The pathophysiology of NTG is still not fully understood, but it is believed to involve factors beyond just IOP, including vascular and biomechanical aspects66. Vascular theories suggest that insufficient blood supply and vascular dysregulation play a role in NTG. These factors could contribute to the specific patterns of optic nerve damage observed in NTG, which may differ from those in other types of glaucoma. Nevertheless, significantly larger differences observed in the optic disc parameters for normal and glaucoma groups can be utilized for their separation. The ROC curves were plotted, and the AUCs for different parameters in the AIROGS and THU datasets are shown in Fig. 13. In general, all the optic disc parameters have high AUC (0.82–0.98). Vertical CDR, area CDR, and NRR normalized with the disc are the best for separating a normal eye from glaucoma. These results are consistent with the reported literature63.

Fig. 12: Normalized histograms and cumulative probabilities ratio.
figure 12

a Inferior-temporal neuroretinal rim (IT-NRR) area to disc area for AIROG. b Superior-temporal neuroretinal rim (ST-NRR) area to disc area for AIROG dataset. c IT-NRR to disc area for THU datasets. d ST-NRR to disc area for THU datasets. The results suggest that changes in the IT area of the optic disc are rapid and similar in both NTG and POAG cases, while changes in the ST region are slower and differ between NTG and POAG groups.

Full size image
Fig. 13: Glaucoma diagnostic ability of different optic disc parameters.
figure 13

a In the AIROGS dataset. b Tohoku University POAG dataset. c Tohoku University NTG dataset. As expected, the vertical cup-to-disc ratio (VCDR) demonstrates the highest diagnostic ability. The area under the curve (AUC) is higher for POAG compared to NT.

Full size image

The AI-GS network can diagnose glaucoma either through direct DL or by integrating early structural changes with DL predictions. Specifically, for diagnoses excluding PPG, the MTL_LWBNA-unet model offers superior sensitivity and specificity based on an optimal threshold detailed in Table 1. This threshold, however, varies across datasets such as PPG, Miyagi screening, POAG, and NTG, highlighting the necessity for dataset-specific adjustments (refer to Fig. 9 and Supplementary Fig. 3). The variation in fundus image quality and disease severity may influence the classification threshold, affecting the DL’s ability to detect disease-specific features. A potential solution involves developing a method to automatically adjust the threshold based on image quality and disease stages or creating a model whose predictions are robust across a broad threshold range. In this study, the AI-GS network achieved this by leveraging multiple models focusing on different aspects of glaucoma, maintaining stable prediction results across a threshold range of 0.1 to 0.65 for the analyzed datasets. This approach enhanced the network’s robustness against variations in image quality and disease stage.

In situations such as mass screenings or when resources are limited, ensuring a cost-effective testing protocol is important. Selecting a threshold that either maximizes sensitivity or balances sensitivity and specificity in a cost-effective manner is crucial. Prevent Blindness America40 recommends a sensitivity of at least 85% at a specificity of 95%. On the entire testing dataset, the AI-GS network showed ~94% sensitivity at 95% specificity and maintained similar specificity and sensitivity (~93–94%) across a wide range of thresholds. However, sensitivity drops to ~83–81% (specificity ~94–91%) for the PPG and Miyagi screening datasets. The AUC for detecting PPG using techniques like Pulsar, Flicker, OCT-disc and OCT-macular ranges from 0.663 to 0.842, whereas for early glaucoma, it ranges from ~ 0.851–0.90767. The AUC from the AI-GS network for PPG is ~ 0.9123, which is significantly high. Differences in labeling for PPG among the ophthalmologists may result in lower AUC/sensitivity because distinguishing between GS/PPG and normal is challenging, and no common consensus exists68. For instance, Deshpande et al.13 defined PPG as normal VF and one or more localized RNFLDs (on fundus image) associated with glaucomatous disc and IOP greater than 21 mmHg. Factors associated with glaucoma, such as DH, and a higher VCDR might be considered. Since glaucomatous morphological changes and VF disturbances are irreversible, examination should be performed regularly to identify disease progression as early as possible during the PPG stage (before the disease progresses to VF impairment)11. In such scenarios, where stringent detection is crucial and missing a case can have serious consequences, the priority is typically to maximize sensitivity.

The AI-GS network can be adjusted to meet different requirements as discussed above. For example, cases predicted as normal by AI-GS but with significantly high prediction probabilities for structural changes from sub-models (such as cupping, RNFLD, and DH) can be further classified as glaucoma suspects (GS). By changing the prediction from normal to GS when any of the following conditions are met—cupping probability >0.65, RNFLD probability >0.8, and DH probability >0.75—the sensitivity for detecting PPG can be increased further from 83.1% to 87.6%. The improvement is even more pronounced for the real-world screening dataset, where sensitivity increases from 80.5% to 93.0%. However, this comes at the cost of decreased specificity (Table 1). The glaucoma group of Miyagi screening includes subjects evaluated for 3, 2, or 1 year. The sensitivity of this adjusted AI-GS network for different groups was as follows: 0.9574 for Group 1 (consistent over 3 years), 0.8654 for Group 2 (2 years), and 0.75 for Group 3 (1 year). On the other hand, the sensitivity for these groups is 0.8571, 0.6442, and 0.625 for the AI-GS network, and 0.6165, 0.4327, and 0.2917 for the single DL binary classification model (MTL_LWBNA-unet), respectively. The cupping model, in particular, demonstrated the most consistent results, closely matching the assessments of the screening ophthalmologists. Sensitivity variations were 0.9449, 0.8173, and 0.8333 for Groups 1, 2, and 3, respectively, suggesting that screening ophthalmologists prefer to make decisions based on cupping.

To further assess the robustness of the AI-GS network compared to a binary classification model for real-world glaucoma screening, we reanalyzed 230 fundus images flagged as “Glaucoma” by the screening ophthalmologist but misclassified as “Normal” by the single DL binary classification model in the Miyagi screening dataset (Table 1). Three glaucoma experts (NT, TN, MS) independently reviewed the images without prior knowledge of AI predictions or screening outcomes (details in Supplementary Table 1). Results revealed significant variability in screening outcomes: the screening ophthalmologist (not necessarily a glaucoma specialist), prioritizing sensitivity to minimize missed positive cases, flagged all 230 images as glaucoma. In comparison, glaucoma experts flagged 206, 178, and 172 cases as glaucoma-positive, reflecting differences in criteria, clinical experience, and thresholds. The single binary classification model (MTL_LWBNA-unet) performed poorly, classifying all cases as “Normal,” highlighting its lack of sensitivity for real-world screening. In contrast, the AI-GS network correctly identified 127 glaucoma cases, and the Adjusted AI-GS model further improved to 193, closely matching glaucoma expert’s opinions. Supplementary Fig. 9 illustrates the AI-GS network’s superior sensitivity, accurately detecting critical features such as cupping, RNFLD, and DH, reclassifying images overlooked by the binary model, and identifying subtle pathologies missed even by the experts. Notably, the Adjusted AI-GS achieved higher agreement with all three experts, effectively complementing their evaluations and addressing diagnostic variability. Detection of these pathological features can help identify borderline cases often missed by screening ophthalmologists due to time constraints. Ophthalmologists may reconsider cases flagged by the AI-GS network. For instance, Supplementary Fig. 9d shows a case marked as normal by both the screening ophthalmologist and the binary classification model, but flagged as a referral by AI-GS, aligning with the glaucoma expert’s decision. These findings establish the AI-GS network as a sensitive and reliable tool for improving detection in challenging cases, addressing critical gaps in real-world glaucoma screening. We believe that integrating the AI-GS network into screening processes can enable general practitioners (GPs) to detect glaucoma at a level comparable to glaucoma experts, and in a significantly shorter time (less than a second).

The enhanced sensitivity for detecting initial structural changes may lead to the detection of DH, RNFLD, and cupping, which might not be related to glaucoma. Retinal diseases such as DR and branch retinal vein occlusion (BRVO) might exhibit hemorrhages and RNFLD. We have analyzed some of the fundus images from the NRG group in the AIROGS dataset, which were predicted to be GS. Supplementary Fig. 10 shows a fundus image where RNFLD in both the ST and IT regions is visible, but cupping and the DL-based glaucoma predictions are negative. In addition to RNFLD, the fundus image exhibits exudates and suggests a case of DR. In another example in Supplementary Fig. 11, DH in the ST-region without glaucomatous features was detected by AI-GS, classifying it as GS. However, some exudates and hemorrhages exist in an arcade position near the fovea, which might suggest a case of BRVO. Supplementary Fig. 12 does not exhibit any obvious features of retinal diseases, except for the visual appearance of choroidal patterns related to myopia. The AI-GS network detected DH in IT, which could be a risk factor for glaucoma (i.e., GS). The AI-GS network detected all features related to glaucoma (DH+, RNFLD+, glaucoma+, and cupping+) in Supplementary Fig. 13. It appears to be a case of glaucoma, but without a VF test, confirmation is difficult.

The present study may have some limitations in terms of the datasets and the results inferred from the analysis of glaucoma. About 71% of fundus images in the normal and 25% in the glaucoma category are from public datasets. Normal group images are non-glaucomatous but may include other retinal diseases or PPG cases (as noted, the AIROGS dataset36,69 was initially aimed for DR screening), which may affect retinal features and prediction accuracies. For the glaucoma group, detailed diagnostic procedures (such as perimetry, IOP) are unavailable. Although the diagnostic procedures of the THU-dataset are well supported by various measurements and ophthalmologists70 it is a hospital-based dataset where some bias in terms of disease severity may exist. The present study used macula-centered images to enhance the model’s ability to detect RNFLD and estimate the optic disc size index. Depending on the quality and resolution of the fundus images, there may be differences in optic disc parameters when estimated from disc-centered images. However, we believe macula-centered images are better suited for screening purposes, as they allow for the detection of other eye diseases beyond glaucoma. To fully assess the generalizability of the AI-GS network, testing it on additional screening datasets is necessary. Unfortunately, publicly available glaucoma datasets that include both macula-centered fundus images and comprehensive diagnostic information (such as IOP, VF, OCT, etc.) are currently limited. Despite these limitations, the AI-GS network demonstrated high accuracy in detecting glaucoma from the macula-centered images in the datasets used for this study. We believe that future data collection efforts from diverse population groups will help in addressing these challenges. Nevertheless, the AI-GS network’s architecture remains robust.

In conclusion, we have introduced an innovative AI-based Glaucoma Screening (AI-GS) network that leverages multiple DL models to accurately mimic the diagnostic approach of an ophthalmologist. Unlike traditional methods, the AI-GS network performs both segmentation and classification tasks, utilizing a hybrid approach that combines black-box predictions with conventional glaucomatous optic disc feature analysis. This dual methodology allows for the early detection of glaucoma indicators such as DH and RNFLD, crucial for preventing irreversible vision loss. On a testing dataset of over 8000 images, the AI-GS network achieves a detection accuracy of ~94%. A comparative analysis in real-world screening scenarios revealed a stark difference in sensitivity between direct DL-based approaches and the AI-GS network. This exceptional performance is attributed to the incorporation of multiple DL models trained on extensive datasets, featuring images captured by various cameras across different locations and ethnicities. The lightweight nature of AI-GS network (total memory size approximately 110 MB) enables its integration into everyday devices, such as smartphones and tablets. This capability is particularly transformative for remote or underserved communities, where access to healthcare resources and specialized ophthalmologic services is limited. By making accurate glaucoma screening accessible to individuals at their convenience, the AI-GS network stands to make a profound impact on global health, offering a scalable solution to prevent the progression of glaucoma and the consequent loss of vision. Future research could explore the integration of additional models trained on other eye diseases to broaden its screening capabilities.

Methods

Subjects and datasets

The present study adheres to the tenets of the Declaration of Helsinki, and the protocols were approved by the Clinical Research Ethics Committee of the Tohoku University, Graduate School of Medicine (No 2023-1-896). The ethics committee approved the study procedure, and all the methods were carried out in accordance with the relevant guidelines and regulations. Informed consent was obtained from all participants and/or their legal guardians.

Training of different DL models in the AI-GS network requires datasets and creating them is time-consuming. For example, the MTL_LWBNA-unet model requires labeling fundus images as glaucoma or normal, along with masks for the cup, disc, and fovea (Fig. 3a, b). It is hard to find such a large-sized dataset needed to train a robust DL model9. For the detection of RNFLD, fundus images of glaucoma patients are required. The detection of RNFLD is highly dependent not only on the pathological presence of the condition but also on several technical and patient-related factors. These include the type of fundus camera used, which affects image resolution and contrast, and age-related ocular conditions such as cataracts and corneal opacities. Furthermore, myopia-related changes, such as the tigroid/stripe pattern, can obscure RNFLD in fundus images. The situation is more complex for DH, as it is a time-dependent phenomenon (appearing/disappearing over time) and is not observed frequently. Separation of fundus images into normal/glaucoma or RNFLD-positive/-negative categories is relatively easier compared to the preparation of masks, which requires precise drawing of cup and disc boundaries. Additionally, mixing images obtained from different cameras and facilities can be beneficial for making a DL model more resilient to changes in domain or modality. In the present work, we searched various publicly available datasets36,37,39, as explained in Table 3, and mixed them with the fundus images of glaucoma/normal patients diagnosed at Tohoku University Hospital70. For the present study, we considered fundus images centered on the fovea/macula. For the subdivision of datasets into PPG, POAG, and NTG categories, non-glaucoma cases were randomly selected to create balanced datasets. For instance, the testing dataset includes 1373 fundus images of the POAG type. An equal number of non-glaucoma images were also randomly chosen from a pool of 4185 images, as detailed in Table 3.

Table 3 Details of fundus image datasets used for training, validation, and testing of DL models
Full size table

EyePACS-AIROGS (Rotterdam) dataset

The Rotterdam EyePACS Artificial Intelligence for Robust Glaucoma Screening (AIROGS) dataset36 is a comprehensive collection of color fundus images from various subjects and sites, designed for glaucoma detection and classification. It includes 113,893 images obtained from 60,357 subjects across approximately 500 different sites, featuring individuals of diverse ethnic backgrounds. Images are labeled as either “referable glaucoma (RG)” or “non-referable glaucoma (NRG)” based on adherence to the guidelines established by the International Council of Ophthalmology. From this dataset, we used 1544 images of RG and 2284 of NRG in our training dataset. In the testing dataset, 3284 images of NRG are also included. The selection was based on the fundus image file size (>1.2 MB) and the presence of both optic disc and fovea.

Retinal Fundus Glaucoma Challenge (REFUGE) dataset

The REFUGE dataset37 was introduced as a satellite event at the 2018 MICCAI conference. This dataset comprises 1200 fundus images, each accompanied by masks specifying the positions (x, y pixel coordinates) of the optic cup, disc, and fovea. The images were acquired using two distinct camera models: Canon CR2 (400 images for testing and 400 for validation) and Zeiss Visucam 500 (400 images for training). Among these 1200 images, 120 are labeled as glaucoma. We used this dataset for segmenting the cup, disc, and fovea boundaries as well as for training the MTL_LWBNA-unet for glaucoma classification.

Creation of cup, disc, and fovea masks

To create a dataset for training the MTL_LWBNA-unet model, we initially trained a segmentation model, the “LWBNA-unet”34, using the publicly available REFUGE dataset (Table 3), which provides masks and positional information for the optic cup, optic disc, and fovea. To enhance the model’s robustness to domain shifts, we combined the original training and testing datasets into a new training set, while retaining the original validation dataset for evaluating the LWBNA-unet model. For detecting the fovea center, we experimented with both binary circle masks and regression-based outputs. While the model produced good segmentation results for the optic cup and disc, it frequently failed to accurately detect the fovea center. This issue likely arises because multiple regions in the fundus image exhibit pixel intensities similar to those of the fovea, making precise localization challenging.

To address this issue, we introduced spatial context to simulate the variations in pixel intensity around the fovea71. A 2D Gaussian distribution was used to generate a heat map of the fovea within the binary mask of the optic cup and disc. The 2D Gaussian kernel is centered on the fovea, with pixel intensity decreasing as the distance from the center coordinates increases. The spread and concentration of the heat map can be controlled by adjusting the standard deviation (σ). The equation below is used to generate the heat map of the fovea71.

$$H(x,y)=exp left(-frac{{left(x-alpha right)}^{2}+{left(y-beta right)}^{2}}{{2sigma }^{2}}right)$$
(1)

Where (α, β) is the center of the fovea, and (x, y) denotes the coordinates of an image vector whose dimensions are 20% of the square-shaped fundus image.

The fovea masks generated using Eq. (1), along with the disc and cup masks, were used to train the LWBNA-unet model six times. We then applied the LWBNA-unet model, trained on the REFUGE dataset, to generate disc, cup, and fovea masks for fundus images captured by a Topcon fundus camera equipped with DRI OCT. This allowed us to create an extended dataset by combining REFUGE dataset images with images from the Topcon camera. While the number of validation images remained the same (400), the training dataset increased to 2058 images. Data augmentation techniques, such as brightness adjustments, minor cropping, and rotation, further expanded the training set to 6174 images, while the validation set grew to 1173 images. The LWBNA-unet model was trained six more times.

Chaksu dataset

It is a publicly available dataset39, encompassing a total of 1345 fundus images captured by three different commercial cameras (Remidio fundus on Phone, Forus 3Nethra Classic, and Bosch handheld cameras). The dataset has been divided into distinct training and testing subsets. In the context of this study, particular attention was directed towards the images from the training subset, specifically those acquired using the Remidio Fundus on Phone and Bosch handheld cameras. Images from the Forus camera were excluded due to cropping issues. We used this dataset to test the segmentation accuracy of the optic cup and disc. Notably, while the MTL_LWBNA-unet model was trained on fundus images centered around the fovea, the images in the Chaksu dataset were centered around the optic disc. The typical method of estimating CDR involves detecting both the optic disc and cup; however, the AI-GS network encounters errors if it fails to detect the fovea’s position, a crucial element for estimating the optic disc size index. As a result, we selected 500 images from the initial 955, basing our choice on visual image quality and the algorithm’s ability to detect the cup, disc, and fovea positions without error. It is worth mentioning that if required, the algorithm can be tailored to obtain parameters related to the optic disc only (without fovea), facilitating the processing of disc-centered fundus images.

Tohoku University Hospital (THU) dataset

Data from patients who visited Tohoku University Hospital in the last 10 years were considered for the present study. Glaucoma/Normal diagnoses relied on multiple tests (OCT, fundus photography, VF, IOP) performed by the specialists and their clinical characteristics have been reported previously70,72,73. For PPG cases, the optic nerve showed glaucomatous damage, but VF defects were not yet detectable by standard perimetric tests. To eliminate medication effects, we used fundus images from patients’ initial visits. We included 1832 fundus images of normal eyes and 1192 images of glaucoma in the training dataset. The testing dataset includes 901 images of normal and 4185 images of glaucoma. Out of 4185 glaucoma images, 2522, 1373, and 290 were of NTG, POAG, and PPG types, respectively.

ODIR (Ocular Disease Intelligent Recognition) dataset

The ODIR dataset38 is a structured ophthalmic database designed to facilitate research in ophthalmology and fundus image analysis. This dataset encompasses data from 5000 patients, including age, color fundus photographs from both left and right eyes, and diagnostic keywords provided by medical professionals. The dataset can be employed for multi-label, multi-disease classification of fundus images, encompassing conditions like diabetic retinopathy, glaucoma, cataract, age-related macular degeneration, hypertension, pathological myopia, and other abnormalities. In the current study, we only included the fundus images labeled as glaucoma in the training dataset. This selection was made to avoid a significant class imbalance (Glaucoma/non-glaucoma).

Real-world screening data from Miyagi Glaucoma Screening

For testing the performance of the AI-GS network on real-world glaucoma screening data, we used images from the Miyagi Glaucoma Screening checkup. We received images captured by a Canon fundus camera over three consecutive years (2020–2022). All participants in this dataset are Japanese. There were ~11,000 right eyes marked as normal for three consecutive years of checkup. We analyzed 10,767 fundus images of normal right eyes from the testing year of 2020. The demographic breakdown in the normal group shows a male-to-female ratio of 1.25. The mean and std of age are 54.08 ± 15.06 years, and the BMI is 23.74 ± 3.79. Other health metrics include systolic blood pressure at 129.77 ± 19.81 mmHg, diastolic blood pressure at 78.48 ± 12.40 mmHg, HDL levels at 63.90 ± 16.50 mg/dL, LDL levels at 122.68 ± 31.33 mg/dL, and HbA1c at 5.72 ± 0.57. For glaucoma detection, we considered 529 fundus images from 209 subjects over 3 years. Among these, 133 subjects were present during all 3 years, 52 for 2 years, and 24 for only 1 year. The male-to-female ratio in this group is 1.09, with an average age of 61.87 ± 10.36 years, BMI of 24.15 ± 4.22, and similar ranges for other health parameters [systolic/diastolic blood pressure (mmHg) = 132.77 ± 20.07/79.87 ± 12.16, HDL/LDL = 64.40 ± 15.97/122.88 ± 29.16, and HbA1c = 5.89 ± 0.52]. Labeling for glaucoma and normal fundus images was performed by screening ophthalmologists.

Dataset for DH detection and segmentation

Publicly available datasets for DH remain scarce. Considering this, we constructed a training dataset comprising fundus images with both positive (DH+ = 918) and negative DH instances (DH− = 1127). For validation, the dataset contains 252 images of DH− and 205 for DH+. Both the testing and validation datasets have multiple images of the same eye (both eyes), but none of the images of a single patient are present in both the datasets. Given that the region of interest (ROI) pertains to the area around the optic disc, the original images necessitate cropping. To achieve this, the contour of the optic disc, detected by MTL_LWBNA-unet, was fitted to a circle, allowing determination of its radius and center (i.e., disc center). Subsequently, the fundus images were cropped into square shapes, with dimensions approximately two times the disc diameter. These ROI-selected images were resized to 512 × 512 pixels and subjected to the CLAHE algorithm to adjust image intensity and contrast. In the training of the DL model for detecting DH, both the original ROI-selected images and the images processed with CLAHE were employed to account for variations in image quality. This augmentation strategy effectively doubles the volume of images utilized for training (4090) and validating (914) the model. We also created a dataset for model testing, consisting of fundus images categorized into two groups: DH− (179 images) and DH+ (180 images). Importantly, this dataset contains only single images of individual eyes.

For the segmentation of DH+ areas in fundus images, we selected 1704 images with DH+. As mentioned above, these were cropped to select images with the ROI. Binary masks for DH were created manually. The images were split into training (1408 images) and validation datasets (296 images). The dataset contains multiple images of the same eye recorded at different time intervals, but the same eye image is not present in both the training and validation datasets. By using image data augmentation (random rotation, cropping, stretching, changing brightness/contrast, etc.) the number of images was increased to 8307 (training: 6837, and validation: 1470).

Dataset for RNFLD detection

There is currently no publicly available dataset for RNFLD detection. Creating such a dataset is challenging, and it is likely to have inconsistent labeling due to unclear presence in fundus images. We created a dataset of 3721 fundus images, comprising 2438 RNFLD-negative and 1283 RNFLD-positive cases. The dataset was divided into five sets for five-fold cross-validation training, ensuring a consistent ratio of RNFLD-positive and RNFLD-negative images (20%). To enhance the robustness of the trained model, we employed data augmentation and increased the number of images to 14,884.

Implementation of DL model

A workstation (DELL Precision 7820 Tower) equipped with an NVIDIA RTX A6000, 48 GB GPU, and 192 GB RAM was used for training and validation of DL models shown in Fig. 2. The open-source programming language, Python 3 backed by Keras API integrated in Tensorflow 2, was used. Details of the parameters used for training the LWBNA-unet model for segmentation have been reported in our previous paper34. For the training of classification models, we used categorical cross-entropy as the loss function.

Evaluation metrics

The Dice coefficient (D) is utilized specifically to evaluate the performance of an image segmentation model34. It quantifies the similarity between the predicted segmentation and the ground truth, with values ranging from 0 (indicating no overlap) to 1 (indicating perfect overlap). For evaluating the performance of classification models, several different metrics are employed. Prediction accuracy is the ratio of correct predictions (glaucoma and normal) to the total number of cases, providing a general measure of the model’s performance. Sensitivity, or the true positive rate, is the ratio of correct glaucoma predictions to the total number of glaucoma cases, assessing the model’s ability to correctly identify positive instances. Specificity, or the true negative rate, is the ratio of correct normal predictions to the total number of normal cases, evaluating the model’s ability to correctly identify negative instances. The Receiver Operating Characteristic (ROC) curve is important for understanding the performance of a binary classification (glaucoma/non-glaucoma). It visualizes the trade-off between the true positive rate (sensitivity) and the false positive rate (1 − specificity) across different classification thresholds. The area under the ROC curve (AUC) offers an aggregate measure of performance across all possible classification thresholds, reflecting the model’s ability to distinguish between classes. An AUC of 0.5 indicates random guessing and 1.0 is for perfect classification. The optimum threshold value for classification decisions is often determined using the Youden Index, which maximizes the difference between the true positive rate and the false positive rate, effectively balancing sensitivity and specificity. A higher value of the index indicates better model performance, with 1 being the best possible value (achievable when TPR is 1 and FPR is 0).

To systematically assess differences in correlation patterns between glaucoma and normal groups within our datasets and determine confidence intervals (CIs), we developed a robust, Python-based analytical framework using “SciPy” for statistical analysis74. Our primary objective was to understand variability in different metrics (AUC, sensitivity, specificity, and Youden Index) and identify potential biomarkers or risk factors associated with glaucoma. For computing CIs around these metrics, we employed a bootstrap resampling technique using Python. This technique involves generating multiple bootstrap samples from the original data, calculating the desired metrics for each sample, and then determining the empirical distribution of these metrics. The CIs are subsequently extracted from these distributions by selecting the appropriate percentile values based on the desired confidence level (typically 95%), ensuring a robust estimation that accounts for the inherent variability in the data. Separate correlation matrices were generated for each group, employing the Seaborn library for heatmap visualizations. To ensure the appropriateness of our correlation analysis, we conducted preliminary tests of normality on our datasets using SciPy74. Based on these assessments, Pearson correlation was chosen for datasets that conformed to normal distribution, while Spearman rank correlation was utilized for datasets that did not, accommodating both linear and monotonic relationships. A critical aspect of our methodology is leveraging bootstrap techniques to estimate the variability and statistical significance of the observed correlation differences. By generating thousands of bootstrap samples, we constructed empirical distributions of correlation differences, deriving statistical estimates and CIs. This non-parametric approach is advantageous as it does not assume a specific underlying distribution, enhancing the robustness of our findings. Bootstrap-derived estimates were integrated into heatmap visualizations, offering a statistically grounded comparison of correlation patterns characteristic between different datasets. p values play a crucial role in our statistical analysis, serving as a tool to quantify the evidence against a null hypothesis. Conventional significance levels (***p < 0.001, **p < 0.01, *p < 0.05) were adopted to annotate heatmaps. ***p < 0.001 to indicate an extremely strong likelihood that the observed correlation difference is not due to random chance.

Related Articles

Photovoltaic bioelectronics merging biology with new generation semiconductors and light in biophotovoltaics photobiomodulation and biosensing

This review covers advancements in biosensing, biophotovoltaics, and photobiomodulation, focusing on the synergistic use of light, biomaterials, cells or tissues, interfaced with photosensitive dye-sensitized, perovskite, and conjugated polymer organic semiconductors or nanoparticles. Integration of semiconductor and biological systems, using non-invasive light-probes or -stimuli for both sensing and controlling biological behavior, has led to groundbreaking applications like artificial retinas. From fusion of photovoltaics and biology, a new research field emerges: photovoltaic bioelectronics.

Iron homeostasis and ferroptosis in muscle diseases and disorders: mechanisms and therapeutic prospects

The muscular system plays a critical role in the human body by governing skeletal movement, cardiovascular function, and the activities of digestive organs. Additionally, muscle tissues serve an endocrine function by secreting myogenic cytokines, thereby regulating metabolism throughout the entire body. Maintaining muscle function requires iron homeostasis. Recent studies suggest that disruptions in iron metabolism and ferroptosis, a form of iron-dependent cell death, are essential contributors to the progression of a wide range of muscle diseases and disorders, including sarcopenia, cardiomyopathy, and amyotrophic lateral sclerosis. Thus, a comprehensive overview of the mechanisms regulating iron metabolism and ferroptosis in these conditions is crucial for identifying potential therapeutic targets and developing new strategies for disease treatment and/or prevention. This review aims to summarize recent advances in understanding the molecular mechanisms underlying ferroptosis in the context of muscle injury, as well as associated muscle diseases and disorders. Moreover, we discuss potential targets within the ferroptosis pathway and possible strategies for managing muscle disorders. Finally, we shed new light on current limitations and future prospects for therapeutic interventions targeting ferroptosis.

The risk effects of corporate digitalization: exacerbate or mitigate?

This study elaborates on the risk effects of corporate digital transformation (CDT). Using the ratio of added value of digital assets to total intangible assets as a measure of CDT, this study overall reveals an inverse relationship between CDT and revenue volatility, even after employing a range of technical techniques to address potential endogeneity. Heterogeneity analysis highlights that the firms with small size, high capital intensity, and high agency costs benefit more from CDT. It also reveals that advancing information infrastructure, intellectual property protection, and digital taxation enhances the effectiveness of CDT. Mechanism analysis uncovers that CDT not only enhances financial advantages such as bolstering core business and mitigating non-business risks but also fosters non-financial advantages like improving corporate governance and ESG performance. Further inquiries into the side effects of CDT and the dynamics of revenue volatility indicate that CDT might compromise cash flow availability. Excessive digital investments exacerbate operating risks. Importantly, the reduction in operating risk associated with CDT does not sacrifice the potential for enhanced company performance; rather, it appears to augment the value of real options.

How digital technologies have been applied for architectural heritage risk management: a systemic literature review from 2014 to 2024

This systematic literature review critically examines the application of digital technologies in architectural heritage risk management from 2014 to 2024, focusing exclusively on English-language publications. As the significance of architectural heritage continues to be recognized globally, there is an increasing shift towards integrating digital solutions to ensure its preservation and management. This paper explores the evolution and application of digital technologies such as Building Information Modeling (BIM), Geographic Information Systems (GIS), and advanced imaging techniques within the field. It highlights how these technologies have facilitated the non-destructive evaluation of heritage sites and enhanced accessibility and interaction through virtual and augmented reality applications. By synthesizing data from various case studies and scholarly articles, the review identifies current trends and the expanding scope of digital interventions in heritage conservation. It discusses the interplay between traditional conservation approaches and modern technological solutions, providing insights into their complementary roles. The analysis also addresses the challenges and limitations encountered in the digital preservation of architectural heritage, such as data integration, the compatibility of different technologies, and the need for more comprehensive frameworks to guide the implementation of digital tools in heritage conservation practices. Ultimately, this review underscores the transformative impact of digital technology in managing architectural heritage risks, suggesting directions for future research and the potential for innovative applications in the field.

Responses

Your email address will not be published. Required fields are marked *