This page is for health professionals only.

NO
I AM NOT
A HEALTHCARE PROFESSIONAL.
Diabetic Retinopathy Screening Approaches in Developing Countries: A Systematic Review and Meta-Analysis
PDF
Cite
Share
Request
Meta-Analysis
VOLUME: 55 ISSUE: 5
P: 260 - 275
October 2025

Diabetic Retinopathy Screening Approaches in Developing Countries: A Systematic Review and Meta-Analysis

Turk J Ophthalmol 2025;55(5):260-275
1. Widya Mandala Catholic University Faculty of Medicine, Surabaya, Indonesia
2. Widya Mandala Catholic University Faculty of Medicine, Department of Ophthalmology, Surabaya, Indonesia
3. Udayana University Faculty of Medicine, Department of Ophthalmology, Denpasar, Indonesia
4. Udayana University Faculty of Medicine, Denpasar, Indonesia
No information available.
No information available
Received Date: 08.05.2025
Accepted Date: 15.08.2025
Online Date: 27.10.2025
Publish Date: 27.10.2025
PDF
Cite
Share
Request

Abstract

Objectives

Diabetic retinopathy (DR) is one of the primary causes of vision loss among people living with diabetes and is expected to rise globally in the coming years. Effective screening strategies are essential, particularly in developing countries where resources and access to specialized care are limited. Our objective was to assess how accurately different screening methods detect DR, specifically artificial intelligence (AI)-based tools, portable fundus cameras, and trained non-ophthalmologist personnel, implemented in a developing country.

Materials and Methods

A literature search was conducted in ScienceDirect, PubMed, and the Cochrane Library. Study quality was assessed using the Quality Assessment of Diagnostic Accuracy Studies 2 tool. While all included studies were reviewed qualitatively, only those evaluating AI-based screening tools were included in the meta-analysis. Meta-analysis was performed using MetaDisc 2.0 to calculate pooled sensitivity, specificity, diagnostic odds ratio, and likelihood ratios for any DR, referable DR, and vision-threatening DR.

Results

A total of 25 studies were included, with 21 AI-based studies eligible for the meta-analysis. The pooled sensitivity and specificity respectively were 0.890 (95% confidence interval [CI]: 0.845-0.924) and 0.900 (95% CI: 0.832-0.942) for any DR, 0.933 (95% CI: 0.890-0.960) and 0.903 (95% CI: 0.871-0.928) for referable DR, and 0.891 (95% CI: 0.393-0.990) and 0.936 (95% CI: 0.837-0.977) for vision-threatening DR. Meta-regression identified camera type as a significant factor. Portable fundus cameras and general physicians showed good agreement with the gold standards.

Conclusion

These findings support the potential of AI-assisted DR screening in low-resource settings and highlight the complementary roles of portable imaging and task-shifting to trained non-specialists.

Keywords:
Diabetic retinopathy screening, artificial intelligence, portable fundus camera, non-specialist, developing countries

Introduction

Diabetes mellitus (DM) is a long-term metabolic disorder that may result in microvascular and macrovascular complications. As living standards have improved significantly, changes in dietary habits and lifestyles have contributed to a steady rise in the prevalence of DM. The primary microvascular complication associated with DM is diabetic retinopathy (DR). It is the leading cause of vision impairment among adults and older individuals.1 The global incidence of DR is expected to rise significantly, increasing from approximately 103 million people in 2020 to an estimated 130 million by 2030 and nearly 161 million by 2045.2, 3 Meanwhile, cases of vision-threatening diabetic retinopathy (VTDR) are projected to grow by 26.3%, reaching 36 million by 2030 and 44.82 million by 2045.3

The ideal method for diagnosing DR is a thorough eye examination with pupil dilation, performed by an ophthalmologist utilizing either an indirect ophthalmoscope or a slit lamp biomicroscope. However, various obstacles limit optimal DR screening, such as limited healthcare access, time limitations, substantial personnel costs, insufficient awareness and comprehension, and inadequate care coordination.4 In clinical trials, the Early Treatment Diabetic Retinopathy Study (ETDRS) seven-standard field protocol, comprising 7 stereoscopic 30-degree fundus photographs, has long been considered the benchmark for DR assessment. Nevertheless, single-field fundus imaging is a practical and effective alternative, particularly considering the logistical, financial, and time-related limitations that make the ETDRS approach unsuitable for routine screening.5

The current recommended guidelines for DR management strategies strongly focus on screening and fundus evaluation. Recent technological advancements, including improved camera technology and artificial intelligence (AI), are becoming more affordable and accessible in low- and middle-income countries. Digitizing health records for individuals with DR would support the creation of a registry, allowing for efficient patient tracking, monitoring disease progression, and assessing referral and treatment outcomes.6 Therefore, this study aimed to present an overview of the implementation of DR screening modalities in developing countries, including using AI, fundus camera technology, and other community-based screening, and compare them to opportunistic-based screening approaches.

Materials and Methods

Data Sources and Search Strategy

This review followed the guidelines outlined in the Preferred Reporting Items for Systematic Reviews and Meta-Analyses.7, 8 The study was registered in the International Prospective Register of Systematic Reviews (CRD420251007510). Seven reviewers independently searched for studies published in PubMed, ScienceDirect, and the Cochrane database for relevant articles. The following search terms were used to identify potentially relevant articles: “diabetic retinopathy” AND “screening” AND “community based” OR “telemedicine” OR “teleophthalmology” OR “artificial intelligence” OR “camera” AND “developing countries” OR “low-income countries” OR “middle income countries.” The terms from each category were independently compared and cross-referenced with those from other categories.

Selection Criteria and Selection

This systematic review and meta-analysis included studies conducted in developing countries (i.e., low- and middle-income countries) that involved participants with type 1 or type 2 DM, and provided data on the sensitivity, specificity, or agreement level of the screening methods used. The screening modalities included AI, telemedicine, camera technology, or other community-based programs. The selected studies must have also compared these interventional screening modalities with standard care screening. The “developing countries” in this research were classified based on World Bank data when the studies were conducted. Any country categorized as a low- or middle-income country was included under the term “developing countries.”

Studies were excluded if they lacked sufficient data, focused solely on the prevalence of DR or on comorbid eye diseases, or were case reports, guidelines, editorials, commentaries, opinions, or reviews. Titles and abstracts of the selected articles were screened by seven reviewers, with full texts of potentially eligible studies examined for final inclusion. Any disagreements were resolved through discussion.

Quality Assessment

The seven reviewers independently assessed the quality of all included studies using the Quality Assessment of Diagnostic Accuracy Studies 2 (QUADAS-2) tool.9 The QUADAS-2 scale comprised four bias risk assessment domains: patient selection, index test, reference standard, and flow and timing. Each domain included two or three individual questions. Risk in a domain was considered low if all questions were answered affirmatively. This scale also evaluated applicability of the study based on patient selection, index test, and reference standard.

Data Extraction and Analysis

After article selection, the seven reviewers summarized and extracted data related to the screening methods’ diagnostic accuracy. These data included total participants, the country where the study was conducted, interventional screening methods, technical characteristics (pupil dilation status, AI system, device), indicators measured, and outcomes such as DR type, sensitivity, specificity, and agreement. Since not all studies analyzed each of these indicators, our meta-analysis was further divided into subgroups based on the available uniform indicators. We used the web application MetaDisc 2.0 for the outcome variables of true positives, false positives, false negatives, and true negatives. We also generated a summary receiver operating characteristic (SROC) curve and forest plots to visualize the pooled results. The bivariate I2 test was used to assess heterogeneity resulting from a potential non-threshold effect in this meta-analysis. If I2 exceeds 50%, it is deemed considerable heterogeneity. MetaDisc 2.0 supports bivariate meta-analysis and provides global heterogeneity (bivariate I2), but does not compute subgroup-specific I2 directly.

Next, subgroup analysis and meta-regression techniques (pupil dilation status, AI algorithm, and camera device) were applied to diagnostic accuracy and heterogeneity to evaluate the possible impact of the covariates. This approach allowed us to maximize its diagnostic meta-analysis strength while acknowledging its limitations. To assess diagnostic accuracy, a bivariate random-effects model was employed to derive pooled sensitivity, specificity, diagnostic odds ratio (DOR), and likelihood ratios (LR+ and LR-). The area under the SROC curve reflected the AI’s performance in diagnosing DR.

Results

Study Selection and Characteristics

Figure 1 summarizes the literature search and selection process. Initially, a total of 3,216 relevant articles were identified from the specified databases using a structured retrieval approach. Studies that were duplicates, conference abstracts, or one of the article types specified in the exclusion criteria (case report, guideline, editorial, commentary, opinion, or review, including meta-analysis), those without available full texts, and those with titles or abstracts unrelated to our review were excluded. After this initial screening, 42 original studies remained. Further evaluation led to the exclusion of papers with unclear methodologies or incomplete or irrelevant targeted outcomes.

The characteristics of the remaining 25 studies are summarized in Table 1. These studies used several screening modalities for detecting DR: 21 evaluated the accuracy of AI-based/assisted screening, 2 assessed the accuracy of handheld/smartphone-based fundus cameras, and 2 reported about empowering trained general physicians to enhance the coverage of DR detection. The studies were performed in developing countries in Asia (China, India, Sri Lanka, Philippines, and Thailand), South America (Brazil and Mexico), and Africa (Zambia and Kenya). The primary goal of screening studies involving general physicians using an AI-based portable device was to evaluate and compare their accuracy to standard care for identifying any grade of DR, referable diabetic retinopathy (RDR), and VTDR. Most studies employed the International Clinical Diabetic Retinopathy Severity Scale classification system, where moderate nonproliferative diabetic retinopathy (NPDR) or worse was considered RDR, and severe NPDR or worse was considered VTDR. We included more-than-mild DR in the RDR group.

Quality Assessment

Twenty-five studies were reviewed for methodological quality and potential bias, following the QUADAS-2 guidelines. The evaluation revealed a risk of patient selection bias in approximately 40% of the studies (Figure 2). An overview of the quality assessment for each study is provided in Figure 3. For the remaining three domains (index test, reference standard, and flow and timing), the results suggested a generally low risk of bias and few issues with applicability. No studies were excluded following the quality assessment.

The Performance of Screening Modalities for Detecting Diabetic Retinopathy

Twenty-one studies were included in the final meta-analysis stage. These studies evaluated the performance of AI-based/assisted screening for DR in developing countries compared to standard/reference screening methods. They reported the performance of AI in detecting RDR (n=18), VTDR (n=3), and DR of any severity (n=11). We further evaluated AI’s performance in detecting any DR and RDR based on pupil dilation status (mydriatic or non-mydriatic), algorithm (convolutional neural network [CNN] or deep learning [DL]), and camera device (smartphone-based/portable retinal camera or retinal fundus camera). Studies where pupil dilation was performed only when necessary were classified under the non-mydriatic group, whereas those employing combined methods were included in the mydriatic group. Most studies excluded ungradable images, while some performed analyses with and without the ungradable images. In this review, we only included the results for gradable images (see Table 2).

We used MetaDisc 2.0 to analyze the performance of AI-based screening in the included studies. Table 2 presents the pooled sensitivity, specificity, DOR, LR+, LR-, and I2 for any DR, RDR, and VTDR. The forest plots of sensitivity, specificity, and SROC curve are shown in Figures 4, 5, and 6, respectively. The SROC curves illustrate the overall diagnostic performance of AI models for detecting any DR and RDR. The SROC curve for RDR demonstrates a more concentrated confidence ellipse, indicating greater consistency across studies. In contrast, the wider prediction ellipse in the any-DR SROC suggests higher variability in diagnostic accuracy. This variability may reflect differences in study populations, image quality, or AI model architectures. Overall, the AI models exhibited more stable and reliable performance in detecting RDR, whereas their effectiveness in identifying any DR appears more heterogeneous.

The I2 values were high overall, with 0.809 for any DR and 0.82 for RDR, indicating substantial heterogeneity. We also performed a meta-regression with MetaDisc 2.0 using the subgroup analysis parameters to explore potential sources of heterogeneity. The outcomes are presented in Tables 3 and 4. For the meta-regression analysis, only studies utilizing CNN or DL algorithms were included for the AI algorithm covariate because one study employed a machine learning approach, which was insufficient to form a meaningful subgroup or allow for reliable meta-regression. In addition, one study was also excluded from the meta-regression evaluating pupil dilation status and camera type because it did not clearly state whether images were obtained using a mydriatic or non-mydriatic method, nor did it specify the type of camera used.

These exclusions were made to maintain consistency in covariate classification and preserve the validity of the meta-regression analysis. However, all excluded studies were still included in the overall pooled analysis of diagnostic accuracy. We found that for any DR, none of the covariates significantly explained the heterogeneity. In contrast, for RDR detection, p values indicated statistical significance for camera device (p<0.05), suggesting that variations in the type of camera used for RDR detection could contribute to the heterogeneity across studies.

Two studies evaluated handheld or smartphone-based fundus imaging (SBFI) as a portable device alternative to standard fundus photography. Wintergerst et al.10 compared four SBFI modalities, three using direct and one using indirect ophthalmoscopy. The images were compared against reference standards of 7-field color fundus photography. Meanwhile, indirect ophthalmoscopy conducted by a specialist was evaluated for image clarity, coverage area, duration of examination, and accuracy in diagnosing DR.

Among 381 eyes of 193 subjects, all SBFI methods produced clear images, but direct SBFI had more artifacts and lower contrast than indirect SBFI. Across different smartphone-based imaging systems, sensitivity for any DR detection ranged from 67% to 79% while specificity remained high, between 98% and 100%. For RDR (moderate NPDR or worse), sensitivity varied between 76% and 87%, with specificity between 96% and 100%. Detection of severe DR (severe NPDR or PDR) achieved 100% sensitivity and specificity with some devices. For diabetic maculopathy, sensitivity ranged from 79% to 83%, while specificity was consistently 100%. The authors concluded that indirect ophthalmoscopy-based SBFI provided the highest diagnostic accuracy, with a strong agreement with the reference standard (Cohen’s kappa: 0.868).10

Salongcay et al.11 evaluated non-mydriatic and mydriatic handheld retinal imaging versus ETDRS 7-standard field fundus photography in 225 eyes of 116 patients. For detection of any DR, non-mydriatic devices demonstrated sensitivities ranging from 80% to 89% and specificities between 88% and 97%. Sensitivity for RDR was 87%-93%, while specificity varied from 76% to 92%. For VTDR (severe NPDR or worse, including PDR and DME), sensitivity ranged from 83% to 88% but specificity was lower, ranging from 69% to 86%. Smartscope NM and Aurora/RetinaVue-700 MD images achieved 80% sensitivity and 95% specificity for detecting DR, meeting thresholds for RDR and DME. However, no device met the 95% specificity requirement for VTDR. Non-mydriatic imaging also had higher ungradable rates (15.1%-38.3% for DR) than mydriatic imaging (0%-33.8%).11

Next, two studies evaluated the agreement and diagnostic accuracy of non-ophthalmologists in DR screening. Cunha et al.12 assessed the efficacy of non-mydriatic fundus photography in DR screening by analyzing the diagnostic agreement across qualified family physicians (FP), general ophthalmologists (GO), and a retinal specialist. A total of 397 eyes of 200 individuals with diabetes were examined. The retinal specialist diagnosed DR in 41.8% of eyes, whereas GO1 and GO2 diagnosed DR in 28.7% and 45.8% of cases, respectively. Diagnostic agreement between the FPs and the retinal specialist for DR diagnosis varied from modest to considerable, with kappa values as follows: FP1 = 0.56, FP2 = 0.69, FP3 = 0.73, FP4 = 0.71. Similarly, agreement in DR severity grading was moderate to substantial (FP1 = 0.51, FP2 = 0.66, FP3 = 0.69, FP4 = 0.64). However, the agreement for DME diagnosis was lower, varying from fair (FP1 = 0.33, FP2 = 0.39, FP3 = 0.37) to moderate (FP4 = 0.51).12

Furthermore, Piyasena et al.13 evaluated the diagnostic accuracy of a handheld non-mydriatic fundus camera in Sri Lanka, where nine general physicians were trained by ophthalmologists to perform DR screening. Two physicians with the highest agreement with the retinal specialist (k = 0.8-0.9) were selected as final graders. For any DR, sensitivity in non-mydriatic imaging ranged from 78.3% to 82.7%, while specificity ranged from 70.4% to 76.2%. With pupil dilation, sensitivity ranged from 78.0% to 79.3%, and specificity improved to 89.2%-91.5%. The kappa agreement value with a retinal specialist for any DR improved from 0.42-0.47 in non-mydriatic imaging to 0.66-0.68 after pupil dilation. For RDR, sensitivity in non-mydriatic imaging ranged from 84.9% to 86.8%, while specificity ranged from 71.7% to 77.3%. With pupil dilation, sensitivity improved to 88.7%-92.5% and specificity increased to 94.9%-96.4%. The kappa agreement values for RDR detection were 0.23-0.29 in non-mydriatic imaging and increased to 0.68-0.76 in mydriatic imaging. For maculopathy detection, sensitivity in non-mydriatic imaging was 89.2%, specificity was 70.1%, and the kappa agreement with the reference standard was 0.29. The percentage of ungradable images was 43.4% in non-mydriatic imaging and decreased to 12.8% after pupil dilation.13

Discussion

This study assessed the diagnostic effectiveness of different DR detection methods to increase screening availability in developing countries. Recent technological advancements hold significant potential to enhance healthcare services, especially in developing countries. This research analyzed 25 studies, of which 21 were included in the meta-analysis and 4 were included in the qualitative review.

Among the 21 meta-analyzed studies, the diagnostic performance of AI-based/assisted screening demonstrated strong diagnostic ability with a pooled sensitivity of 0.890, specificity of 0.900, and DOR of 72.680 for detecting DR. Similarly, the diagnostic performance of AI-based/assisted screening for detecting RDR had a pooled sensitivity of 0.933, specificity of 0.903, and an even higher DOR of 130.617, demonstrating high accuracy for identifying more severe cases requiring referral.

Meanwhile, although only three studies evaluated VTDR, the pooled results still suggest encouraging performance, with pooled sensitivity at 0.891 and specificity at 0.936, though the limited data warrant careful interpretation. These results exceeded the Food and Drug Administration established 85% sensitivity and 82.5% specificity endpoints.14 They are also consistent with those found in earlier systematic reviews and meta-analyses that evaluated the diagnostic accuracy of AI algorithms in DR screening.15, 16, 17 Our results are also comparable to those of meta-analyses on AI-based detection for other eye disease such as glaucoma, pathologic myopia, and dry eye disease.18, 19, 20

Furthermore, we conducted a subgroup analysis to investigate the factors influencing AI performance in detecting any DR and RDR. AI exhibited similar accuracy in detecting DR from both non-mydriatic and mydriatic images. Mydriatic photographs have slightly better sensitivity but slightly lower specificity than non-mydriatic photographs. This result may be because mydriasis produces more detailed images. False positives occur due to subtle lesions or certain non-DR retinal abnormalities including drusen, atrophy or hypertrophy of the retinal pigment epithelium, telangiectatic vessels near the macula, tessellated fundus, and retinal vein occlusion.21, 22, 23 However, retinal lesions unrelated to DR still indicate that the patient must consult an ophthalmologist or retina specialist. Therefore, they cannot be considered false positives and of no concern in terms of clinical implications. Meanwhile, in non-mydriatic photographs, the retinal images tend to be darker, may not capture all subtle DR lesions, and could result in a higher percentage of ungradable images.24

For the AI algorithm architecture, there was minimal difference in pooled performance between CNN-based models and broader DL algorithms. The pooled sensitivity for DL models was slightly higher than for CNN models, but CNN models achieved better specificity. However, these differences were not statistically significant. Our results suggested that choosing between DL and a CNN architecture did not contribute substantially to diagnostic performance. DL is an advanced branch of machine learning that utilizes multi-layered neural networks to analyze extensive datasets, allowing systems to identify complex visual patterns autonomously. CNN, a specific DL variant, is optimized for image analysis, especially in medical diagnostics.25 CNN-based models utilize convolutional layers to accurately recognize and categorize retinal abnormalities, including microaneurysms, hemorrhages, and exudates, essential for DR detection.26 Since DL-based models demonstrate greater sensitivity, they may be more appropriate for initial screening to minimize missed cases. On the other hand, CNN models could be utilized as reliable confirmation tools, helping to reduce unnecessary referrals due to false positives. Joseph et al.27 also reported in their meta-analysis that the DL algorithm, which included CNN, demonstrated high accuracy compared to machine learning. Only one study in our review used machine learning. When this study was excluded, the pooled sensitivity and specificity increased to 90% and 91%, respectively, for detecting any DR. The improved efficiency and diagnostic accuracy of DL over traditional machine learning has revolutionized the ability to detect DR using fundus images.25, 26, 27

Three studies incorporated AI-generated heatmaps to enhance interpretability in DR screening. Bellemo et al.28 used heatmaps to highlight specific areas in the retinal fundus images that most significantly influence CNN determination. These visualizations illustrate the AI system’s decision-making process and explain features that may encourage trust in AI models.28 The heatmaps of the lesions provided by the AI can also be utilized for patient education.21 Noriega et al.29 also showed that incorporating attention heatmaps highlighted DR lesions and improved grader sensitivity when used in an assistive screening approach. Sayres et al.30 further investigated the heatmaps’ impact on ophthalmologists’ grading accuracy and confidence. They found that while heatmaps improved sensitivity for RDR, they also led to overdiagnosis in cases with no DR, increasing false positives for mild NPDR. This result might be because heatmaps can highlight pathological features but cannot effectively indicate the absence of disease. Despite this initial increase in overdiagnosis, grader accuracy improved over time, suggesting that clinicians adapted to interpreting heatmaps with experience.30

Moreover, although 7-field ETDRS group stereoscopic color fundus photography remains the gold standard for DR, its high cost and time demands have led to the use of handheld and smartphone-based cameras, especially in community-based screening initiatives. Regarding camera type, smartphone-based or portable fundus cameras demonstrated higher sensitivity than desktop fundus cameras. However, they exhibited a slight decrease in specificity, particularly for RDR detection. In our meta-analysis, camera type emerged as a significant source of heterogeneity, which suggested that hardware differences, including image quality and field of view, directly influence AI performance, especially in detecting more severe disease stages. These results align with those reported by Tan et al.31, who found a pooled sensitivity and specificity of 87% and 94% for any DR and 91% and 89% for RDR, respectively. However, while they observed a progressive increase in sensitivity and specificity as DR severity advanced (pooled sensitivity and specificity were 39% and 95% for mild NPDR, 71% and 95% for moderate NPDR, and 80% and 97% for PDR), our meta-analysis did not specifically assess the accuracy for each DR stage. Such an analysis was not possible due to differences in study methods, reference standards, and DR classification approaches.

Furthermore, we examined studies that specifically evaluated smartphone-based and handheld fundus imaging for DR detection to understand the impact of device type on diagnostic performance. Wintergerst et al.10 found that SBFI, especially when using indirect ophthalmoscopy, offered the highest-quality images, the widest field of view, and demonstrated excellent sensitivity and specificity (0.79-0.99 for any DR and 1.0-1.0 for severe DR), and excellent agreement with the reference standard (Cohen’s kappa 0.868). Salongcay et al.11 also reported that non-mydriatic and mydriatic handheld retinal imaging obtained good to excellent kappa agreement values with the ETDRS 7-standard field photography. However, the non-mydriatic method was linked to higher rates of ungradable images and lower levels of agreement.11 Similarly, Prathiba et al.32 found that the non-mydriatic retinal camera demonstrated good agreement with standard tabletop fundus photography. Nevertheless, as with other non-mydriatic approaches, a higher proportion of ungradable images was observed, reinforcing the need for selective pupil dilation to improve image quality and reduce screening errors.32 These findings suggest that for community-based DR screening programs, device selection should consider the trade-off between portability, image quality, and the need for pupil dilation to optimize diagnostic accuracy and reduce false positives.

Although this review focuses on diagnostic accuracy, real-world factors like patient adherence are crucial for successful DR screening programs. The RAIDERS trial in Rwanda evaluated how AI-assisted screening influenced follow-up adherence. Mathenge et al.33 found that immediate AI feedback increased referral adherence by 30.1% (51.5% vs. 39.6%, p=0.048) and a faster median time to follow-up (4 vs. 8 days) compared to human grading. Similarly, Liu et al.34 reported a threefold improvement in adherence (55.4% vs. 18.7%) after implementing AI-based screening in a low-income primary care setting. These findings highlight the potential benefits of AI-assisted screening beyond its diagnostic performance. It also aligns with findings from public perception studies where patients demonstrated high confidence in AI-generated medical diagnoses, suggesting that trust in AI may positively influence screening adherence.35 AI-based/assisted screening may also improve real-world patient engagement by reducing delays and enhancing adherence to follow-up care.

Expanding DR screening by task-shifting to non-ophthalmologists is an important strategy, especially in resource-limited settings where access to specialists is scarce. Two studies evaluated the diagnostic agreement between non-ophthalmologists (FPs/general physicians) and retinal specialists in DR screening. Cunha et al.12 evaluated FP performance in DR screening, comparing it with retinal specialists. They found that FPs achieved moderate to substantial agreement with a retinal specialist (k=0.56-0.73), though agreement on macular edema was fair to moderate (k=0.33-0.51). However, similar agreement was also demonstrated between GOs and the retinal specialist, which suggests that FPs and GOs had similar diagnostic skills.12

Similarly, Piyasena et al.13 reported that general physicians achieved high agreement for any DR detection (k=0.42-0.47 in non-mydriatic imaging, improving to 0.66-0.68 in mydriatic imaging) and for RDR (k=0.23-0.29 non-mydriatic, improving to 0.68-0.76 mydriatic). However, the kappa agreement value for maculopathy detection was lower (k=0.29 non-mydriatic). The study also highlighted that ungradable images were high (43.4%) in non-mydriatic imaging but decreased to 12.8% after pupil dilation, reinforcing the importance of image quality for accurate DR screening.13 Both studies suggest that trained non-ophthalmologists can effectively detect RDR, but challenges remain in maculopathy detection and handling ungradable images. These findings underscore the need for further training and calibration of primary care providers if task-shifting strategies are to be effectively deployed in low-resource settings.

Our review has several strengths. One of the key strengths is its focus on DR screening in developing countries, where access to ophthalmologists is often limited. By including various screening modalities, such as AI-based/assisted identification, smartphone-based or portable fundus imaging, and trained non-ophthalmologist-assisted screening, this review incorporates a wider range of diagnostic methods, allowing for a broader comparison of different screening approaches and providing valuable insights into practical alternatives for resource-limited settings. We also performed a meta-regression analysis incorporating multiple relevant factors, offering important insights. Additionally, most of the included studies reflect real-world screening conditions, enhancing the applicability of these findings to national DR screening programs and public health initiatives.

Nevertheless, this review has several limitations. First, the included studies cover a range of study designs, including retrospective, prospective, cross-sectional, and randomized controlled experiments. The heterogeneity in study design may introduce variability in the reported diagnostic performance of the AI models. Second, the meta-regression analysis identified camera type as a significant source of heterogeneity, suggesting that differences in imaging hardware, such as resolution and field of view, impact diagnostic accuracy. However, mydriatic status and AI algorithm type did not significantly contribute to heterogeneity, indicating that other unaccounted factors may still influence screening accuracy. Another limitation is the unequal distribution of studies across subgroups. Moreover, this meta-analysis focused primarily on diagnostic accuracy, without assessing whether earlier detection through AI-assisted or non-ophthalmologist screening improves patient outcomes such as treatment adherence and vision preservation.

Conclusion

This review highlights the growing feasibility of integrating AI-based and portable imaging technologies into DR screening programs in developing countries. Portable fundus cameras integrated with AI-based software can potentially lower the workload of ophthalmologists while reducing missed or incorrect diagnoses, ultimately helping to prevent vision loss caused by DR. Our findings suggest that both non-mydriatic and mydriatic imaging perform well, making them promising options for large-scale screening. However, pupil dilation should be considered for patients with ungradable retinal images to improve sensitivity without compromising specificity, as it can enhance image quality and reduce missed diagnoses. Ideally, this approach should be conducted under the supervision of trained physicians to maintain screening accuracy, reduce unnecessary referrals, and provide timely and appropriate care. These findings also emphasize the importance of quality assurance measures, including regular training, structured feedback loops, and possibly integrating AI decision support to assist non-specialist graders. Standardizing grading criteria, improving image quality, and refining AI models will be essential to developing reliable and scalable DR screening solutions, particularly in resource-limited settings. Our study demonstrated diagnostic accuracy across modalities, which can guide the development of more inclusive, scalable, and economical national screening programs. This insight might help policymakers choose the appropriate technologies based on workforce availability and local infrastructure. Future research to improve diagnostic performance should assess how these screening techniques could affect clinical outcomes including early intervention, treatment adherence, and long-term vision preservation. These outcome-based studies are essential to fully demonstrate the public health benefits of integrating AI-assisted screening into routine diabetes care.

Ethics

Ethics Committee Approval: Not applicable.
Informed Consent: Not applicable.

Authorship Contributions

Surgical and Medical Practices: K.A.H., A.A., N.M.A.S., T.E., Concept: Y.Y., K.A.H., Design: Y.Y., K.A.H., Data Collection or Processing: Y.Y., A.C.V.G., N.P.K.M.D., Analysis or Interpretation: Y.Y., K.A.H., A.A., N.M.A.S., T.E., A.C.V.G., N.P.K.M.D., Literature Search: Y.Y., K.A.H., A.A., N.M.A.S., T.E., A.C.V.G., N.P.K.M.D., Writing: Y.Y., K.A.H.
Conflict of Interest: No conflict of interest was declared by the authors.
Financial Disclosure: The authors declared that this study received no financial support.

References

1
Lin KY, Hsih WH, Lin YB, Wen CY, Chang TJ. Update in the epidemiology, risk factors, screening, and treatment of diabetic retinopathy. J Diabetes Investig. 2021;12:1322-1325.
2
Tan TE, Wong TY. Diabetic retinopathy: looking forward to 2030. Front Endocrinol (Lausanne). 2023;13:1077669.
3
Teo ZL, Tham YC, Yu M, Chee ML, Rim TH, Cheung N, Bikbov MM, Wang YX, Tang Y, Lu Y, Wong IY, Ting DSW, Tan GSW, Jonas JB, Sabanayagam C, Wong TY, Cheng CY. Global prevalence of diabetic retinopathy and projection of burden through 2045: systematic review and meta-analysis. Ophthalmology. 2021;128:1580-1591.
4
Avidor D, Loewenstein A, Waisbourd M, Nutman A. Cost-effectiveness of diabetic retinopathy screening programs using telemedicine: a systematic review. Cost Eff Resour Alloc. 2020;18:16.
5
Abou Taha A, Dinesen S, Vergmann AS, Grauslund J. Present and future screening programs for diabetic retinopathy: a narrative review. Int J Retina Vitreous. 2024;10:14.
6
Takkar B, Das T, Thamarangsi T, Rani PK, Thapa R, Nayar PD, Rajalakshmi R, Choudhury N, Hanutsaha P. Development of diabetic retinopathy screening guidelines in South-East Asia region using the context, challenges, and future technology. Semin Ophthalmol. 2022;37:97-104.
7
Page MJ, McKenzie JE, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, Shamseer L, Tetzlaff JM, Akl EA, Brennan SE, Chou R, Glanville J, Grimshaw JM, Hróbjartsson A, Lalu MM, Li T, Loder EW, Mayo-Wilson E, McDonald S, McGuinness LA, Stewart LA, Thomas J, Tricco AC, Welch VA, Whiting P, Moher D. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021;372:n71.
8
Page MJ, Moher D, Bossuyt PM, Boutron I, Hoffmann TC, Mulrow CD, Shamseer L, Tetzlaff JM, Akl EA, Brennan SE, Chou R, Glanville J, Grimshaw JM, Hróbjartsson A, Lalu MM, Li T, Loder EW, Mayo-Wilson E, McDonald S, McGuinness LA, Stewart LA, Thomas J, Tricco AC, Welch VA, Whiting P, McKenzie JE. PRISMA 2020 explanation and elaboration: updated guidance and exemplars for reporting systematic reviews. BMJ. 2021;372:n160.
9
QUADAS-2. https://www.bristol.ac.uk/population-health-sciences/projects/quadas/quadas-2/.
10
Wintergerst MWM, Mishra DK, Hartmann L, Shah P, Konana VK, Sagar P, Berger M, Murali K, Holz FG, Shanmugam MP, Finger RP. Diabetic retinopathy screening using smartphone-based fundus imaging in India. Ophthalmology. 2020;127:1529-1538.
11
Salongcay RP, Aquino LAC, Salva CMG, Saunar AV, Alog GP, Sun JK, Peto T, Silva PS. Comparison of handheld retinal imaging with ETDRS 7-standard field photography for diabetic retinopathy and diabetic macular edema. Ophthalmol Retina. 2022;6:548-556.
12
Cunha LP, Figueiredo EA, Araújo HP, Costa-Cunha LVF, Costa CF, Neto JMC, Matos AMF, de Oliveira MM, Bastos MG, Monteiro MLR. Non-mydriatic fundus retinography in screening for diabetic retinopathy: agreement between family physicians, general ophthalmologists, and a retinal specialist. Front Endocrinol (Lausanne). 2018;9:251.
13
Piyasena MMPN, Yip JLY, MacLeod D, Kim M, Gudlavalleti VSM. Diagnostic test accuracy of diabetic retinopathy screening by physician graders using a hand-held non-mydriatic retinal camera at a tertiary level medical clinic. BMC Ophthalmol. 2019;19:89.
14
Abràmoff MD, Lavin PT, Birch M, Shah N, Folk JC. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. NPJ Digit Med. 2018;1:39.
15
Wang Z, Li Z, Li K, Mu S, Zhou X, Di Y. Performance of artificial intelligence in diabetic retinopathy screening: a systematic review and meta-analysis of prospective studies. Front Endocrinol (Lausanne). 2023;14:1197783.
16
Wang S, Zhang Y, Lei S, Zhu H, Li J, Wang Q, Yang J, Chen S, Pan H. Performance of deep neural network-based artificial intelligence method in diabetic retinopathy screening: a systematic review and meta-analysis of diagnostic test accuracy. Eur J Endocrinol. 2020;183:41-49.
17
Nielsen KB, Lautrup ML, Andersen JKH, Savarimuthu TR, Grauslund J. Deep learning-based algorithms in screening of diabetic retinopathy: a systematic review of diagnostic performance. Ophthalmol Retina. 2019;3:294-304.
18
Shi NN, Li J, Liu GH, Cao MF. Artificial intelligence for the detection of glaucoma with SD-OCT images: a systematic review and Meta-analysis. Int J Ophthalmol. 2024;17:408-419.
19
Heidari Z, Hashemi H, Sotude D, Ebrahimi-Besheli K, Khabazkhoob M, Soleimani M, Djalilian AR, Yousefi S. Applications of artificial intelligence in diagnosis of dry eye disease: a systematic review and meta-analysis. Cornea. 2024;43:1310-1318.
20
Zhang Y, Li Y, Liu J, Wang J, Li H, Zhang J, Yu X. Performances of artificial intelligence in detecting pathologic myopia: a systematic review and meta-analysis. Eye (Lond). 2023;37:3565-3573.
21
Jain A, Krishnan R, Rogye A, Natarajan S. Use of offline artificial intelligence in a smartphone-based fundus camera for community screening of diabetic retinopathy. Indian J Ophthalmol. 2021;69:3150-3154.
22
Penha FM, Priotto BM, Hennig F, Przysiezny B, Wiethorn BA, Orsi J, Nagel IBF, Wiggers B, Stuchi JA, Lencione D, de Souza Prado PV, Yamanaka F, Lojudice F, Malerbi FK. Single retinal image for diabetic retinopathy screening: performance of a handheld device with embedded artificial intelligence. Int J Retina Vitreous. 2023;9:41.
23
Pawar B, Lobo SN, Joseph M, Jegannathan S, Jayraj H. Validation of artificial intelligence algorithm in the detection and staging of diabetic retinopathy through fundus photography: an automated tool for detection and grading of diabetic retinopathy. Middle East Afr J Ophthalmol. 2021;28:81-86.
24
Yang Y, Pan J, Yuan M, Lai K, Xie H, Ma L, Xu S, Deng R, Zhao M, Luo Y, Lin X. Performance of the AIDRScreening system in detecting diabetic retinopathy in the fundus photographs of Chinese patients: a prospective, multicenter, clinical study. Ann Transl Med. 2022;10:1088.
25
Ting DSW, Pasquale LR, Peng L, Campbell JP, Lee AY, Raman R, Tan GSW, Schmetterer L, Keane PA, Wong TY. Artificial intelligence and deep learning in ophthalmology. Br J Ophthalmol. 2019;103:167-175.
26
Ting DSW, Peng L, Varadarajan AV, Keane PA, Burlina PM, Chiang MF, Schmetterer L, Pasquale LR, Bressler NM, Webster DR, Abramoff M, Wong TY. Deep learning in ophthalmology: the technical and clinical considerations. Prog Retin Eye Res. 2019;72:100759.
27
Joseph S, Selvaraj J, Mani I, Kumaragurupari T, Shang X, Mudgil P, Ravilla T, He M. Diagnostic accuracy of artificial intelligence-based automated diabetic retinopathy screening in real-world settings: a systematic review and meta-analysis. Am J Ophthalmol. 2024;263:214-230.
28
Bellemo V, Lim ZW, Lim G, Nguyen QD, Xie Y, Yip MYT, Hamzah H, Ho J, Lee XQ, Hsu W, Lee ML, Musonda L, Chandran M, Chipalo-Mutati G, Muma M, Tan GSW, Sivaprasad S, Menon G, Wong TY, Ting DSW. Artificial intelligence using deep learning to screen for referable and vision-threatening diabetic retinopathy in Africa: a clinical validation study. Lancet Digit Health. 2019;1:e35-e44.
29
Noriega A, Meizner D, Camacho D, Enciso J, Quiroz-Mercado H, Morales-Canton V, Almaatouq A, Pentland A. Screening diabetic retinopathy using an automated retinal image analysis system in independent and assistive use cases in Mexico: Randomized Controlled Trial. JMIR Form Res. 2021;5:e25290.
30
Sayres R, Taly A, Rahimy E, Blumer K, Coz D, Hammel N, Krause J, Narayanaswamy A, Rastegar Z, Wu D, Xu S, Barb S, Joseph A, Shumski M, Smith J, Sood AB, Corrado GS, Peng L, Webster DR. Using a deep learning algorithm and integrated gradients explanation to assist grading for diabetic retinopathy. Ophthalmology. 2019;126:552-564.
31
Tan CH, Kyaw BM, Smith H, Tan CS, Tudor Car L. Use of smartphones to detect diabetic retinopathy: scoping review and meta-analysis of diagnostic test accuracy studies. J Med Internet Res. 2020;22:e16658.
32
Prathiba V, Rajalakshmi R, Arulmalar S, Usha M, Subhashini R, Gilbert CE, Anjana RM, Mohan V. Accuracy of the smartphone-based nonmydriatic retinal camera in the detection of sight-threatening diabetic retinopathy. Indian J Ophthalmol. 2020;68(Suppl 1):42-46.
33
Mathenge W, Whitestone N, Nkurikiye J, Patnaik JL, Piyasena P, Uwaliraye P, Lanouette G, Kahook MY, Cherwek DH, Congdon N, Jaccard N. Impact of artificial intelligence assessment of diabetic retinopathy on referral service uptake in a low-resource setting: The RAIDERS randomized trial. Ophthalmol Sci. 2022;2:100168.
34
Liu J, Gibson E, Ramchal S, Shankar V, Piggott K, Sychev Y, Li AS, Rao PK, Margolis TP, Fondahn E, Bhaskaranand M, Solanki K, Rajagopal R. Diabetic retinopathy screening with automated retinal image analysis in a primary care setting improves adherence to ophthalmic care. Ophthalmol Retina. 2021;5:71-77.
35
Stai B, Heller N, McSweeney S, Rickman J, Blake P, Vasdev R, Edgerton Z, Tejpaul R, Peterson M, Rosenberg J, Kalapara A, Regmi S, Papanikolopoulos N, Weight C. Public perceptions of artificial intelligence and robotics in medicine. J Endourol. 2020;34:1041-1048.
36
Natarajan S, Jain A, Krishnan R, Rogye A, Sivaprasad S. Diagnostic accuracy of community-based diabetic retinopathy screening with an offline artificial intelligence system on a smartphone. JAMA Ophthalmol. 2019;137:1182-1188.
37
Nunez do Rio JM, Nderitu P, Bergeles C, Sivaprasad S, Tan GSW, Raman R. Evaluating a deep learning diabetic retinopathy grading system developed on mydriatic retinal images when applied to non-mydriatic community screening. J Clin Med. 2022;11:614.
38
Gulshan V, Rajan RP, Widner K, Wu D, Wubbels P, Rhodes T, Whitehouse K, Coram M, Corrado G, Ramasamy K, Raman R, Peng L, Webster DR. Performance of a deep-learning algorithm vs manual grading for detecting diabetic retinopathy in India. JAMA Ophthalmol. 2019;137:987-993.
39
Ruamviboonsuk P, Tiwari R, Sayres R, Nganthavee V, Hemarat K, Kongprayoon A, Raman R, Levinstein B, Liu Y, Schaekermann M, Lee R, Virmani S, Widner K, Chambers J, Hersch F, Peng L, Webster DR. Real-time diabetic retinopathy screening by deep learning in a multisite national screening programme: a prospective interventional cohort study. Lancet Digit Health. 2022;4:e235-e244.
40
Sosale B, Aravind SR, Murthy H, Narayana S, Sharma U, Gowda SGV, Naveenam M. Simple, mobile-based artificial intelligence algo r ithm in the detection of diabetic retinopathy (SMART) study. BMJ Open Diabetes Res Care. 2020;8:e000892.
41
Pei X, Yao X, Yang Y, Zhang H, Xia M, Huang R, Wang Y, Li Z. Efficacy of artificial intelligence-based screening for diabetic retinopathy in type 2 diabetes mellitus patients. Diabetes Res Clin Pract. 2022;184:109190.
42
Dong X, Du S, Zheng W, Cai C, Liu H, Zou J. Evaluation of an artificial intelligence system for the detection of diabetic retinopathy in chinese community healthcare centers. Front Med (Lausanne). 2022;9:883462.
43
Ming S, Xie K, Lei X, Yang Y, Zhao Z, Li S, Jin X, Lei B. Evaluation of a novel artificial intelligence-based screening system for diabetic retinopathy in community of China: a real-world study. Int Ophthalmol. 2021;41:1291-1299.
44
Zhang Y, Shi J, Peng Y, Zhao Z, Zheng Q, Wang Z, Liu K, Jiao S, Qiu K, Zhou Z, Yan L, Zhao D, Jiang H, Dai Y, Su B, Gu P, Su H, Wan Q, Peng Y, Liu J, Hu L, Ke T, Chen L, Xu F, Dong Q, Terzopoulos D, Ning G, Xu X, Ding X, Wang W. Artificial intelligence-enabled screening for diabetic retinopathy: a real-world, multicenter and prospective study. BMJ Open Diabetes Res Care. 2020;8:e001596.
45
Li N, Ma M, Lai M, Gu L, Kang M, Wang Z, Jiao S, Dang K, Deng J, Ding X, Zhen Q, Zhang A, Shen T, Zheng Z, Wang Y, Peng Y. A stratified analysis of a deep learning algorithm in the diagnosis of diabetic retinopathy in a real-world study. J Diabetes. 2022;14:111-120.
46
Bawankar P, Shanbhag N, K SS, Dhawan B, Palsule A, Kumar D, Chandel S, Sood S. Sensitivity and specificity of automated analysis of single-field non-mydriatic fundus photographs by Bosch DR Algorithm-Comparison with mydriatic fundus photography (ETDRS) for screening in undiagnosed diabetic retinopathy. PLoS One. 2017;12:e0189854.
47
Al Turk L, Wang S, Krause P, Wawrzynski J, Saleh GM, Alsawadi H, Alshamrani AZ, Peto T, Bastawrous A, Li J, Tang HL. Evidence based prediction and progression monitoring on retinal images from three nations. Transl Vis Sci Technol. 2020;9:44.
48
Rajalakshmi R, Subashini R, Anjana RM, Mohan V. Automated diabetic retinopathy detection in smartphone-based fundus photography using artificial intelligence. Eye (Lond). 2018;32:1138-1144.
49
Hansen MB, Abràmoff MD, Folk JC, Mathenge W, Bastawrous A, Peto T. Results of automated retinal image analysis for detection of diabetic retinopathy from the Nakuru Study, Kenya. PLoS One. 2015;10:e0139148.
50
Malerbi FK, Nakayama LF, Melo GB, Stuchi JA, Lencione D, Prado PV, Ribeiro LZ, Dib SA, Regatieri CV. Automated identification of different severity levels of diabetic retinopathy using a handheld fundus camera and single-image protocol. Ophthalmol Sci. 2024;4:100481.