Abstract
Background: Artificial intelligence (AI) algorithms offer an effective solution to alleviate the burden of diabetic retinopathy (DR) screening in public health settings. However, there are challenges in translating diagnostic performance and its application when deployed in real-world conditions.
Objective: This study aimed to assess the technical feasibility of integration and diagnostic performance of validated DR screening (DRS) AI algorithms in real-world outpatient public health settings.
Methods: Prior to integrating an AI algorithm for DR screening, the study involved several steps: (1) Five AI companies, including four from India and one international company, were invited to evaluate their diagnostic performance using low-cost nonmydriatic fundus cameras in public health settings; (2) The AI algorithms were prospectively validated on fundus images from 250 people with diabetes mellitus, captured by a trained optometrist in public health settings in Chandigarh Tricity in North India. The performance evaluation used diagnostic metrics, including sensitivity, specificity, and accuracy, compared to human grader assessments; (3) The AI algorithm with better diagnostic performance was integrated into a low-cost screening camera deployed at a community health center (CHC) in the Moga district of Punjab, India. For AI algorithm analysis, a trained health system optometrist captured nonmydriatic images of 343 patients.
Results: Three web-based AI screening companies agreed to participate, while one declined and one chose to withdraw due to low specificity identified during the interim analysis. The three AI algorithms demonstrated variable diagnostic performance, with sensitivity (60%-80%) and specificity (14%-96%). Upon integration, the better-performing algorithm AI-3 (sensitivity: 68%, specificity: 96, and accuracy: 88·43%) demonstrated high sensitivity of image gradability (99.5%), DR detection (99.6%), and referral DR (79%) at the CHC.
Conclusions: This study highlights the importance of systematic AI validation for responsible clinical integration, demonstrating the potential of DRS to improve health care access in resource-limited public health settings.
doi:10.2196/67529
Keywords
Introduction
The global prevalence of diabetes is growing [
], leading to increased vision loss and blindness associated with it [ ]. There is an urgent need for diabetic retinopathy (DR) screening programs to identify vision-threatening DR to enable timely treatment [ , ]. However, this rising prevalence is straining health care systems already struggling to improve care and manage health care costs [ ].Despite its critical role, DR screening (DRS) remains limited in many low-resource settings. Conventional screening with trained human graders is often costly, time-consuming, and challenging to scale [
, ]. The gap between eye care needs and ophthalmologist availability exacerbates public health challenges [ ]. In India, DR management faces challenges due to limited screening programs, low public awareness, and poor understanding of routine retinal exams [ ]. Artificial intelligence-driven DRS enables faster, more affordable, and efficient screening, especially in underserved areas [ , ]. It allows noneye health professionals to conduct screenings and refer without specialists [ ].AI algorithms have demonstrated performance comparable to or exceeding human experts in DR classification [
, ]. However, concerns exist about their suboptimal performance in real-world settings and across diverse populations [ ]. Real-world validation is critical to ensure AI algorithms perform accurately in diverse settings, as disease prevalence, image quality, and patient-related factors may differ from the training dataset [ - ]. Prospective studies are essential for evaluating AI systems in the contexts where they will be deployed [ ]. Hence, integrating AI into clinical practice requires alignment with clinical workflows and stronger evidence on its real-world accuracy and user experiences [ , ].This is the first Indian study to validate multiple commercial AI algorithms for DRS and to assess the feasibility of implementing a validated AI system in public health settings.
Methods
Study Design
The study prospectively validated three DR detection AI algorithms (validation phase) and assessed the technical feasibility of implementing a validated AI algorithm in public health settings (implementation phase). The STARD (Standards for Reporting of Diagnostic Accuracy Studies) checklist [
] was used to report the completeness and transparency of the diagnostic accuracy study and the iCHECK-DH framework to enhance the completeness and transparency of reporting related to the digital health implementation components [ ]. summarizes the study design.
Ethical Considerations
The study received approval from the Postgraduate Institute of Medical Education and Research (PGIMER) Institutional Ethics Committee (PGI/IEC/2020/001342) and followed the recommendations of the Declaration of Helsinki. The study was prospectively registered with the Clinical Trials Registry India (CTRI/2022/10/046185). Individuals aged > 30 years with a history of diabetes mellitus were screened for DR, following the National Program for Prevention and Control of Non-Communicable Diseases [
]. Informed consent was obtained, and their routine care remained unchanged. All data used in this study were anonymized prior to analysis, with no personal identifiers retained.Study Site (Real-World Settings)
Validation was conducted from March to June 2021 at the Department of Endocrinology, PGIMER, Chandigarh, and a Primary Health Center (PHC) in Khizrabad, District Mohali, Punjab. The AI-enabled DRS was implemented between February 2022 and June 2022 at a community health center (CHC) in Badhani Kalan, Moga District, Punjab, India.
Sample Size
Sample sizes (validation=256; implementation=348) were calculated assuming DR prevalence of 17%, nongradable image rates of 18.4% in validation and 30.3% during implementation, 70% sensitivity, 86% specificity, 95% CI, and nonresponse rates of 10% and 15% in validation and implementation, respectively [
, - ].AI Algorithms
Based on a scoping review, five AI companies that used cloud-based AI (four Indian, one international) were invited. Before validation, they received the study objectives and camera or image specifications.
Fundus Image Acquisition
Training of Optometrists
Two experienced optometrists were trained for 15 days at the Advanced Eye Centre (AEC) and PHC settings before validation (Table S1 in
) [ ]; two distinct optometrists received similar training for implementation in Moga district. Training covered identifying good-quality images [ ] and capturing additional images when needed. They were supervised until proficient in independent data collection and imaging.Camera
All participants underwent nonmydriatic, two-field (macula and disc-centered), 45° FOV fundus photography using a 3Netra Classic Portable Benchtop Fundus Camera (Forus Health) [
].Recruitment Process
During validation, participants were recruited through Accredited Social Health Activist (ASHA) workers at PHCs and by research staff in the endocrinology department . The study did not alter patients’ routine care during the recruitment process. During implementation, recruitment at the CHC was assisted by a nurse from the medicine clinic.
Darkroom
[To minimize ungradable images from nonmydriatic cameras that are a key challenge for real-world DRS programs, [
]dark rooms with sealed windows and ventilators were set up at PHC Khijrabad and at CHC Badhani Kalan (Figure S1 in ), with participants seated in the dark for ≥2 minutes to achieve physiological mydriasis before imaging.Image Grading
Image identifiers were removed, and images remained unprocessed before AI analysis. The research optometrist sorted and uploaded them by eye laterality (ie, right and left eye). AI companies had no direct access; only the uploading optometrist handled the images per the agreement with the companies. A separate, role-based account was created for image upload and grading, accessible only to the optometrist. Audit logs tracked access, and secure file transfer protocols protected the transmission of images and the human grader had similar restricted access. Validation images were not used for training or testing the AI.
AI Grading Protocol
The AI algorithms’ screening outcomes were inconsistent in providing DR stages (mild nonproliferative diabetic retinopathy (NPDR)), moderate NPDR, severe NPDR, and proliferative diabetic retinopathy (PDR)), diabetic macular edeme (DME) (yes/no), and referable DR (RDR) (Figure S2 in
). Hence, the DR grade (yes/no) was chosen to validate the AI algorithms discussed elsewhere [ ].Reference Standard Grading
All AI screening outputs were manually graded by human graders using the International Classification for Diabetic Retinopathy (ICDR) classification system [
]. Images were labeled as “gradable” or “non-gradable,” and DR as “present” or “absent.” DR was classified as mild, moderate, or severe NPDR and PDR [ ]. Microaneurysms that are early signs of DR, indicate mild NPDR [ ]. Any DR mentioned above, including moderate NPDR were considered referable DR (RDR) [ ]. DME was defined as hard exudates with or without foveal involvement [ ]. Images with ≥80% visibility and clear view up to the third vascular branch were considered gradable and assessed for DR [ ].- Validation phase: Two masked human graders (HG1 and HG2), a trained optometrist with three years of grading experience, and a retina-trained ophthalmologist with seven years of grading experience, with different institutional affiliations, independently graded all the fundus images. A senior vitreoretinal expert with 25 years of grading experience re-evaluated 224 images from 56 participants (75%) where HG1 and HG2 disagreed on DR presence. A strong level of agreement [ ] (κ=0.85) was observed between HG2 and the senior retina specialist; hence, HG2 grading was considered the reference standard (RS) for AI validation.
- Implementation phase: All images were independently graded by two masked human graders (HG1 and HG2), who were different from those involved in the validation phase. In cases of disagreement, the grades were reviewed and adjudicated by a senior vitreoretinal specialist with over eight years of experience in grading. The final consensus-based, arbitrated dataset served as the RS and was used for all statistical analyses.
AI Integration and Implementation
The better-performing AI algorithm was integrated into the 3Netra Classic and pilot-tested for two weeks at AEC and District Hospital (DH), Mohali. This phase ensured hardware-software compatibility, assessed internet connectivity, and included dummy tests to validate outputs before implementation. The two-week testing at PGIMER and DH Mohali validated the full AI workflow, identified technical issues, assessed reliability in clinical settings, and highlighted data and reporting bottlenecks.
This feedback led to key refinements, including better internet connectivity, mandatory clinical variables, faster result turnaround, and local data storage on the National Institute of Transforming India (NITI) server for compliance. Final adjustments were implemented at Moga and Mohali sites (Figure S3 in
). The AI algorithm did not undergo additional training during the implementation period. Testing was conducted by optometrists, supported by a research optometrist, and supervised by the data scientist and principal investigator. The research team coordinated with the AI company to address feedback, and the machine learning scientist assisted with preimplementation adjustments.Poststudy, a follow-up mechanism was established to ensure continued service delivery (6 mo), with the optometrist monitored for adherence to screening protocols. Notably, the retinal camera and adjustable stand were retained at the CHC rather than being reclaimed as study assets, reinforcing sustainability through local ownership and continuity of service.
Referral Recommendation and Telephonic Follow-Up
Referral were based on cases diagnosed as moderate or more severe, as well as those with DME [
]. Participants requiring follow-up were contacted by phone one month after their DRS appointment to assess compliance, with up to three contact attempts made to gather this information.Data Analysis
The study data was collected using Research Electronic Data Capture (REDCap) [
]. Deidentified data were downloaded from REDCap and imported into Stata/IC (version 15.1; StataCorp) [ ] for analysis.The AI platform’s sensitivity, specificity, positive predictive value (PPV), and negative predictive value (NPV) were estimated with 95% exact binomial CIs. A P value of <·05 was considered significant for all statistical tests. A κ value was calculated monthly during the implementation phase to assess the optometrists learning in image acquisition and the AI algorithm’s performance in image quality and DR diagnosis. This measured the agreement between AI and RSs for image quality and DR diagnosis. The folders with missing bilateral or macula-centered images and people without diabetes mellitus were excluded from the analysis. All analyses were conducted on an eye-wise basis.
Results (Implementation)
Three AI algorithm companies Leben Care Health Services, Retinal AI Diagnostic Software, and SigTuple Technologies. Each AI platform was masked from the others and randomly assigned labels such as AI 1, AI 2, and AI 3. One Indian AI algorithm declined participation, and an Food and Drug Administration (FDA)-approved algorithm was excluded due to low specificity in interim analysis. Two Indian AI companies approached during the implementation phase but were not included.
Sociodemographic Details of the Study Participants
Among the 250 participants in the validation phase, 182 were recruited at the PHC Khijrabad and 68 from the endocrinology clinic. The mean age of the participants was 53.5 (SD 7) years in the PHC and 47.2 (SD 4) years in the Endocrinology clinic, respectively. Overall, 87/250 (48.3%) participants in the PHC group were men and 95 (51.7%) were women, and those in the Endocrinology group included 38 (55.8%) men and 30 (44.1%) women. In the implementation phase, 343 participants were recruited. The participants’ mean age was 58.48 (SD 10.41) years, with 202/343 (59%) women and 140 (41%) men (
).Characteristics | Validation phase (n=250) | Implementation phase (n=343) |
Gender n (%) | ||
Men | 121(48.4%) | 141 (41) |
Women | 129(51.6%) | 202 (59) |
Age (years), mean (SD) | 54.4 (13.8) | 58.4 (10.4) |
Duration of diabetes (years), mean (SD) | 7.1 (6.1) | 6.2 (5.9) |
Education, n (%) | ||
Illiterate | 72(28.8) | 135(39.4) |
Primary | 79(31.6) | 125(36.4) |
Matriculation | 70(28) | 63(18.4) |
Secondary | 14(6.8) | 8(2.3) |
Graduation and above | 15(6) | 12(3.5) |
Validation Phase
The analysis included 1099 fundus images of 500 eyes from 250 participants. According to the RS, 484 (96·8%) of eyes were gradable, with 16 (3·2%) being ungradable. AI 1, AI 2, and AI 3 achieved excellent gradability (100%, 93.8%, and 100%, respectively). The RS detected DR in 133 (27.48%) of the eyes; AI 1 detected notably more cases of DR in 446 (89.2%) eyes compared with AI 2, 122 (26.22%), and AI 3, 106 (21.2%) (
).AI 3 showed the best performance with specificity of detecting DR, 96.01% (93·24‐97·72); sensitivity, 68.42% (59·71‐76.05); PPV, 86.67% (78.31‐92.26); accuracy, 88.43%; and agreement with the RS (κ=0·65) and was selected for implementation (Table S2 in
). The validation results have been presented elsewhere [ ].Based on our study findings and recommendations from the project technical oversight committee which comprised ophthalmologists, public health experts, academicians, technologists, and research scientists providing technical expertise and strategic direction to evaluate, guide, and strengthen the project’s scientific rigor. The AI company was recommended to train its algorithm for DR stages, DME grading, and referral guidelines, incorporating inputs on camera integration, connectivity, and platform functionality. Changes were completed within four months, and implementation began at CHC, with testing images excluded from the final set (Table S3 in
). Changes were incorporated within four months and implemented at CHC.
Implementation Phase
During the implementation phase, 1372 fundus images were captured from 686 eyes of 343 participants. AI-3 exhibited slightly higher image gradability (682, 99.5%) than the RS (336, 92.71%). AI-3 identified 124 (18.2%) of participants with DR, while the RS detected 189 (28.9%). Moderate NPDR was most common (AI-3 at 87 (12.8%) and RS at 90 (14.1%). DME detection was lower for AI-3, 11 (1.6%) than the RS, 34 (5.3%). Referral rates were almost similar, (AI-3: 99; 14.5%) and RS (109, 17.1%) (
).Among the 50/686 (7.3%) ungradable eyes graded by HG, 21 (42%) had cataracts, while none labeled as ungradable by AI had cataract. The sensitivity, specificity, PPV, and NPV of AI-3 algorithm for DR detection were 99.6%, 64.7%, 87.4%, and 98.3%, respectively. For DME, specificity, PPV, and NPV were 99.7%, 81.8%, and 96%, respectively. However, sensitivity is relatively low (26.5%); for detecting RDR, sensitivity and specificity were 78.9% and 98.1%. For image gradability, the AI’s sensitivity was excellent at 100%, but specificity was low (8%) (
).
Outcome variables | Image gradability | DR | gradeDME | gradeRDR |
Sensitivity (%) | 100 | 99·6 | 26·5 | 78·9 |
Specificity (%) | 8 | 64·7 | 99·7 | 98·1 |
PPV (%) | 93·2 | 87·4 | 81·8 | 89·6 |
NPV (%) | 100 | 98·3 | 96 | 95·7 |
κ value | 0.69 | 0.72 | 0.38 | 0.81 |
aPPV: Positive predictive value.
bNPV: Negative predictive.
cAI: artificial intelligence.
dDR: Diabetic retinopathy.
eDME: Diabetic macular edeme.
fRDR: referable diabetic retinopathy.
κ Statistic Variation Across the Study Period
shows κ variations for image quality and DR grades over 4.5 months when the health system optometrist captured the images. κ values of image quality increased from 0 in February to 0.74 in June, and for DR grade from 0 to 0.71 in June. This steady improvement could be attributed to the enhanced image quality captured by the optometrist over the study period and higher; in DR, this increase can be linked to better image quality and higher detection, with slight dips due to variation in DR severity or case complexity.

Participant Referral and Follow-Up
Of 64 referred participants, 28 (43.8%) were contacted; only 9 (14%) adhered to referral advice and visited an ophthalmologist for a review. Of these, 1 received an Optical Coherence Tomography referral, 5 went to facilities without eye care services, 1 received eye drops, 1 was advised antivascular endothelial growth factor injection, 1 left after receiving laser treatment, and 1 had a follow-up visit.
Reasons for nonadherence included harvesting season (n=11), lack of family support (n=2), time constraints (n=2), extended absence from home (n=1), financial dependence on family (n=2), and comorbidities (n=1).
Discussion
Principal Findings
This study is among the first in India to validate multiple AI algorithms for DRS, critically assessing their technical feasibility for integration into real-world public health settings. We evaluated the best-performing AI algorithm using images captured by a skilled research optometrist. This study outlines the Government of India’s efforts to foster an ecosystem that ensures the integration of responsible AI technologies before application to end users [
].In our study, AI system performance during validation varied significantly: sensitivity (59.7%97.74%), specificity (14.25%96.01%), PPV (30.16%86.67%), NPV (85%94.34%), and accuracy (37.19%88.43%). A multicenter study by Lee et al systematically compared seven AI-based DRS algorithms, revealing high NPVs (82.72%93.69%) but widely varying sensitivities (50.98%85.90%) in real-world performance [
]. In pivotal clinical trials for the IDX AI algorithm, the FDA’s benchmark for superiority was set at 85% sensitivity and 82.5% specificity [ ]. Although AI-3 sensitivity (68.3%) was below the FDA’s >85% threshold, it outperformed other AIs in other diagnostic metrics: specificity (96.01%), PPV (86.67%), NPV (88.92%), and accuracy (88.43%). Prospective validation of a DRS algorithm at two Indian tertiary eye care hospitals demonstrated 89% sensitivity and 92% specificity on nonmydriatic images [ ].However, there are challenges in comparing algorithms using published results due to the variations in study methodologies [
]. One challenge is the dependence of an AI algorithm’s accuracy on the quality of the retinal images obtained [ ]. Our study included ungradable images in validation, contrasting with many studies that preprocess or exclude lower-quality images in their training data sets [ , , - ] (Figure S6 in ). Excluding these images fails to reflect real-world settings, potentially lowering the algorithm’s performance [ , ]. To date, validation studies demonstrate that most AI algorithms using mydriatic fundus images achieve high diagnostic accuracy [ ]. In the validation phase, the dark room environment facilitated nonmydriatic conditions, yielding 0% ungradability for AI-1 and AI-3, 6.2% for AI-2, and 3.2% by human grader, contrasting with 1830% in LMIC settings per a systematic review [ , ].Cataracts are a leading cause of ungradable images [
]. However, no images with cataract or other media opacities were classified as ungradable during implementation, potentially affecting AI-3 specificity ( ). Notably, the AI-3 algorithm failed to detect 21 (42%) eyes with cataracts, categorized as gradable. The AI-3 algorithm’s sensitivity increased significantly from 68.4% to 99.6%, while specificity decreased from 96% to 64.7% between the validation and implementation phases. The sensitivity likely improved post validation due to algorithm training (Table S2 in ). The presence of cataracts, affecting media opacity and gradability, likely could have decreased specificity by increasing false positives [ ]. After algorithmic training, AI-3 detected DME ( ) and sensitivity (26.5%), highlighting the need for improved training to enhance DME sensitivity and subsequently achieve higher RDR sensitivity (78.9%).Optometrists serve as frontline eye care providers globally and are ideally positioned for task-sharing in DRS. Their integration into DRS and care pathways is well-established in diverse models worldwide, supporting sustainable and scalable eye care delivery [
- ]. In this study, a health system optometrist was trained and engaged in DRS at the CHC, with oversight from the research team. Over the study period, image quality and DR detection improved, reflecting a positive learning curve ( ). These findings highlight the potential of optometrist-led AI-assisted screening in strengthening task-shifting models and expanding access to DR care in resource-limited settings [ - ]. Implementing an opportunistic AI-enabled DRS model holds promise for enhancing detection rates [ ]. However, our study’s adherence to referral recommendations remains suboptimal, with approximately 14% of participants attending recommended follow-up visits at an eye care facility. Key barriers to referral adherence include awareness gaps, logistical challenges (ie, travel, DR related cost), and persisting health system limitations, including weak referral pathways and poor patient tracking [ ]. Personalized approaches, such as phone calls, voicemails, and detailed result letters, are shown to be effective in improving referral adherence [ ]. Low referral adherence underscores the need for effective, coordinated referral pathways before introducing new screening models, while recognizing the vital role of teleophthalmology in such systems [ , ].A notable strength of this study is its real-time implementation, conducted during regular clinics by a health system optometrist within a public health care setting. This marks a significant milestone that is likely to substantially enhance DRS services. The study showed that training and monitoring fundus image quality could significantly improve the effectiveness of the AI-enabled DRS program.
However, several limitations should be considered when interpreting the results. Using a single fundus camera may limit generalizability, as AI performance may vary across camera models and imaging conditions. Additionally, the study relied on a specific AI algorithm, which may not account for variations in other models or evolving AI systems. Setting-specific constraints such as workflow integration and infrastructure availability could also impact scalability. Furthermore, potential biases in real-world data, such as variations in patient demographics, image quality, and disease prevalence, may influence AI performance and limit broader applicability. Integrating DRS was feasible; however, we could monitor the optometrist only for three months poststudy. Regular monitoring is crucial for program sustainability and healthcare provider motivation.
Lessons Learnt
Operational challenges in DRS included limited patient access due to long travel distances, poor transport, and low awareness; inadequate infrastructure; and ergonomic barriers affecting both patients and screeners. Uncontrolled lighting and power issues led to 26% ungradable images. Adaptive measures, such as transport support, ergonomic adjustments, darkroom setups, and equipment reinforcements, raised image gradeability to 95.6% and improved efficiency. Addressing these barriers through infrastructure upgrades, controlled environments, and community facilitation is essential for sustainable DRS in primary health care systems.
In conclusion, this study highlights the essential role of systematic AI validation in integrating technology responsibly into clinical workflows. By demonstrating AI’s feasibility for DRS in Indian public health settings, our findings support scalable solutions to improve health care accessibility in resource-constrained contexts across the Global South. However, long-term sustainability and large-scale implementation will require ongoing funding, a robust workforce, and effective policy integration. Further research is needed to evaluate the large-scale deployment of AI-driven screening strategies, examining their clinical effectiveness, cost-effectiveness, and the challenges of implementing them in real-world settings.
Acknowledgments
We extend our gratitude to the Punjab Health Department, Civil Surgeons of Mohali and Moga, Senior Medical Officers of Boothgarh and Badhani Kalan, and the Deputy Commissioner of Moga for their support in conducting this study. The study was funded by the National Institution for Transforming India (NITI) Aayog, the Government of India.
Parts of this study were presented in abstract form at the Global Partnership on Artificial Intelligence (GPAI) Summit, New Delhi, India, 12–14 December 2023 [
].Authors' Contributions
Conceptualization: AC, A Kankaria, GM, MD, VG
Data curation: DB, GK, LK, NM, PV
Formal analysis: AC, VM
Methodology: AC, A Kankaria, GM, MD, VG
Validation: AC, MD, VM
Writing – original draft: AC, A Kankaria, MD
Writing – review & editing: A Kernohan, GM, LV, NS
Conflicts of Interest
None declared.
Supplementary materials.
DOCX File, 792 KBiCHECK-DH: Guidelines and Checklist for the Reporting on Digital Health Implementations.
PDF File, 160 KBReferences
- Cho NH, Shaw JE, Karuranga S, et al. IDF Diabetes Atlas: global estimates of diabetes prevalence for 2017 and projections for 2045. Diabetes Res Clin Pract. Apr 2018;138:271-281. [CrossRef] [Medline]
- International Diabetes Federation. IDF Diabetes Atlas. 7th ed. International Diabetes Federation; 2015.
- Scanlon PH. The English National Screening Programme for diabetic retinopathy 2003-2016. Acta Diabetol. Jun 2017;54(6):515-525. [CrossRef] [Medline]
- Alowais SA, Alghamdi SS, Alsuhebany N, et al. Revolutionizing healthcare: the role of artificial intelligence in clinical practice. BMC Med Educ. Sep 22, 2023;23(1):689. [CrossRef] [Medline]
- Padhy SK, Takkar B, Chawla R, Kumar A. Artificial intelligence in diabetic retinopathy: a natural step to the future. Indian J Ophthalmol. Jul 2019;67(7):1004-1009. [CrossRef] [Medline]
- Grzybowski A, Brona P, Lim G, et al. Artificial intelligence for diabetic retinopathy screening: a review. Eye (Lond). Mar 2020;34(3):451-460. [CrossRef] [Medline]
- Bali J, Bali O, Sahu A, Boramani J, Senthil T, Deori N. State of the nation survey on cataract surgery in India. Indian J Ophthalmol. Nov 2022;70(11):3812-3817. [CrossRef] [Medline]
- Ramasamy K, Raman R, Tandon M. Current state of care for diabetic retinopathy in India. Curr Diab Rep. Aug 2013;13(4):460-468. [CrossRef] [Medline]
- Beede E, Baylor E, Hersch F, et al. A human-centered evaluation of a deep learning system deployed in clinics for the detection of diabetic retinopathy. 2020. Presented at: Conference on Human Factors in Computing Systems - Proceedings; Apr 21, 2020; Honolulu HI USA. [CrossRef]
- Chassagnon G, Vakalopoulou M, Paragios N, Revel MP. Artificial intelligence applications for thoracic imaging. Eur J Radiol. Feb 2020;123:108774. [CrossRef] [Medline]
- Keel S, Lee PY, Scheetz J, et al. Feasibility and patient acceptability of a novel artificial intelligence-based screening model for diabetic retinopathy at endocrinology outpatient services: a pilot study. Sci Rep. Mar 12, 2018;8(1):4330. [CrossRef] [Medline]
- Gulshan V, Peng L, Coram M, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. Dec 13, 2016;316(22):2402-2410. [CrossRef] [Medline]
- Abràmoff MD, Lavin PT, Birch M, Shah N, Folk JC. Pivotal trial of an autonomous AI-based diagnostic system for detection of diabetic retinopathy in primary care offices. NPJ Digit Med. 2018;1(1):39. [CrossRef] [Medline]
- Romero-Brufau S, Wyatt KD, Boyum P, Mickelson M, Moore M, Cognetta-Rieke C. A lesson in implementation: a pre-post study of providers’ experience with artificial intelligence-based clinical decision support. Int J Med Inform. May 2020;137:104072. [CrossRef] [Medline]
- Kohner EM, Aldington SJ, Stratton IM, et al. United Kingdom Prospective Diabetes Study, 30: diabetic retinopathy at diagnosis of non-insulin-dependent diabetes mellitus and associated risk factors. Arch Ophthalmol. Mar 1998;116(3):297-303. [CrossRef] [Medline]
- Topol EJ. High-performance medicine: the convergence of human and artificial intelligence. Nat Med. Jan 2019;25(1):44-56. [CrossRef] [Medline]
- Kanagasingam Y, Xiao D, Vignarajan J, Preetham A, Tay-Kearney ML, Mehrotra A. Evaluation of artificial intelligence-based grading of diabetic retinopathy in primary care. JAMA Netw Open. Sep 7, 2018;1(5):e182665. [CrossRef] [Medline]
- Kelly CJ, Karthikesalingam A, Suleyman M, Corrado G, King D. Key challenges for delivering clinical impact with artificial intelligence. BMC Med. Oct 29, 2019;17(1):195. [CrossRef] [Medline]
- Cohen JF, Korevaar DA, Altman DG, et al. STARD 2015 guidelines for reporting diagnostic accuracy studies: explanation and elaboration. BMJ Open. Nov 14, 2016;6(11):e012799. [CrossRef] [Medline]
- Perrin Franck C, Babington-Ashaye A, Dietrich D, et al. iCHECK-DH: Guidelines and Checklist for the Reporting on Digital Health Implementations. J Med Internet Res. 2023;25:e46694. [CrossRef]
- National program for prevention and control of cancer, diabetes, cardiovascular diseases and stroke (NPCDCS)—handbook for counselors reducing risk factors for noncommunicable diseases risk factors for NCDs. Directorate General of Health Services Ministry of Health and Family Welfare Government of India. URL: https://nhm.gov.in/index1.php?lang=1&level=2&sublinkid=1048&lid=604 [Accessed 2023-10-11]
- Arenas-Cavalli JT, Abarca I, Rojas-Contreras M, Bernuy F, Donoso R. Clinical validation of an artificial intelligence-based diabetic retinopathy screening tool for a national health system. Eye (Lond). Jan 2022;36(1):78-85. [CrossRef] [Medline]
- Rêgo S, Dutra-Medeiros M, Soares F, Monteiro-Soares M. Screening for diabetic retinopathy using an automated diagnostic system based on deep learning: diagnostic accuracy assessment. Ophthalmologica. 2021;244(3):250-257. [CrossRef] [Medline]
- Sivaprasad S, Netuveli G, Wittenberg R, et al. Complex interventions to implement a diabetic retinopathy care pathway in the public health system in Kerala: the Nayanamritham study protocol. BMJ Open. Jun 28, 2021;11(6):e040577. [CrossRef] [Medline]
- Sivaprasad S, Raman R, Conroy D, et al. The ORNATE India Project: United Kingdom-India Research Collaboration to tackle visual impairment due to diabetic retinopathy. Eye (Lond). Jul 2020;34(7):1279-1286. [CrossRef] [Medline]
- Darwish DY, Patel SN, Gao Y, et al. Diagnostic accuracy and reliability of retinal pathology using the Forus 3Nethra fundus camera compared to ultra wide-field imaging. Eye (Lond). May 2019;33(5):856-857. [CrossRef] [Medline]
- Shi C, Lee J, Wang G, Dou X, Yuan F, Zee B. Assessment of image quality on color fundus retinal images using the automatic retinal image analysis. Sci Rep. Jun 2022;12(1):10455. [CrossRef]
- Yip MYT, Lim G, Lim ZW, et al. Technical and imaging factors influencing performance of deep learning systems for diabetic retinopathy. NPJ Digit Med. 2020;3:40. [CrossRef] [Medline]
- Duggal M, Chauhan A, Kankaria A, et al. Responsible adoption of cloud-based artificial intelligence in health care: a validation case study of multiple artificial intelligence algorithms for diabetic retinopathy screening in public health settings. In: Advancing Responsible AI in Public Sector Application. Taylor and Francis; 2024.
- Wong TY, Sun J, Kawasaki R, et al. Guidelines on Diabetic Eye Care: The International Council of Ophthalmology Recommendations for Screening, Follow-up, Referral, and Treatment Based on Resource Settings. Ophthalmology. Oct 2018;125(10):1608-1622. [CrossRef] [Medline]
- Wang Z, Li X, Yao M, Li J, Jiang Q, Yan B. A new detection model of microaneurysms based on improved FC-DenseNet. Sci Rep. Dec 2022;12(1):1. [CrossRef]
- McHugh ML. Interrater reliability: the kappa statistic. Biochem Med. 2012:276-282. [CrossRef]
- Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap)--a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. Apr 2009;42(2):377-381. [CrossRef] [Medline]
- Stata statistical software: release 17. StataCorp. 2021. URL: https://www.stata.com/ [Accessed 2025-08-20]
- AI & emerging technologies group promoting the adoption of cutting-edge technologies to create significant economic and societal impact. Government of India, Ministry of Electronics and Information Technology. URL: https://tinyurl.com/5f4a3kr7 [Accessed 2024-06-10]
- Lee AY, Yanagihara RT, Lee CS, et al. Multicenter, head-to-head, real-world validation study of seven automated artificial intelligence diabetic retinopathy screening systems. Diabetes Care. May 2021;44(5):1168-1175. [CrossRef] [Medline]
- Gulshan V, Rajan RP, Widner K, et al. Performance of a deep-learning algorithm vs manual grading for detecting diabetic retinopathy in India. JAMA Ophthalmol. Sep 1, 2019;137(9):987-993. [CrossRef] [Medline]
- Sosale B, Sosale AR, Murthy H, Sengupta S, Naveenam M. Medios- An offline, smartphone-based artificial intelligence algorithm for the diagnosis of diabetic retinopathy. Indian J Ophthalmol. Feb 2020;68(2):391-395. [CrossRef] [Medline]
- Abràmoff MD, Lou Y, Erginay A, et al. Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning. Invest Ophthalmol Vis Sci. Oct 1, 2016;57(13):5200-5206. [CrossRef] [Medline]
- Gargeya R, Leng T. Automated identification of diabetic retinopathy using deep learning. Ophthalmology. Jul 2017;124(7):962-969. [CrossRef] [Medline]
- Bhaskaranand M, Ramachandra C, Bhat S, et al. The value of automated diabetic retinopathy screening with the eyeart system: a study of more than 100,000 consecutive encounters from people with diabetes. Diabetes Technol Ther. Nov 2019;21(11):635-643. [CrossRef] [Medline]
- Piyasena M, Murthy GVS, Yip JLY, et al. Systematic review and meta-analysis of diagnostic accuracy of detection of any level of diabetic retinopathy using digital retinal imaging. Syst Rev. Nov 7, 2018;7(1):182. [CrossRef] [Medline]
- Scanlon PH, Foy C, Malhotra R, Aldington SJ. The influence of age, duration of diabetes, cataract, and pupil size on image quality in digital photographic retinal screening. Diabetes Care. Oct 2005;28(10):2448-2453. [CrossRef] [Medline]
- Melles RB, Conell C, Siegner SW, Tarasewicz D. Diabetic retinopathy screening using a virtual reading center. Acta Diabetol. Feb 2020;57(2):183-188. [CrossRef] [Medline]
- George PP, Yun OCS, Siow K, et al. Is there scope for expanding the optometrist’s scope of practice in Singapore? - a survey of optometrists, opticians in Singapore. Cont Lens Anterior Eye. Jun 2019;42(3):258-264. [CrossRef] [Medline]
- Baker H, Ratnarajan G, Harper RA, Edgar DF, Lawrenson JG. Effectiveness of UK optometric enhanced eye care services: a realist review of the literature. Ophthalmic Physiol Opt. Sep 2016;36(5):545-557. [CrossRef] [Medline]
- Scheetz J, Koca D, McGuinness M, et al. Real-world artificial intelligence-based opportunistic screening for diabetic retinopathy in endocrinology and indigenous healthcare settings in Australia. Sci Rep. Aug 4, 2021;11(1):15808. [CrossRef] [Medline]
- Padhy D, Marmamula S, Mettla AL, et al. Changing trends in barriers to accessing eye care services in rural South India: results from the longitudinal Andhra Pradesh Eye Disease Study III (APEDS III) cohort. Eye (Lond). Aug 2024;38(11):2209-2215. [CrossRef] [Medline]
- Bonilla-Escobar FJ, Eibel MR, Le L, Gallagher DS, Waxman EL. Follow-up in a point-of-care diabetic retinopathy program in Pittsburgh: a non-concurrent retrospective cohort study. BMC Ophthalmol. Aug 20, 2024;24(1):356. [CrossRef] [Medline]
- Xie Y, Nguyen QD, Hamzah H, et al. Artificial intelligence for teleophthalmology-based diabetic retinopathy screening in a national programme: an economic analysis modelling study. Lancet Digit Health. May 2020;2(5):e240-e249. [CrossRef] [Medline]
Abbreviations
AI: artificial intelligence |
CHC: community health center |
DME: diabetic macular edeme |
DR: diabetic retinopathy |
DRS: diabetic retinopathy screening |
NPDR: nonproliferative diabetic retinopathy |
NPV: negative predictive value |
PDR: proliferative diabetic retinopathy |
PHC: primary health center |
PPV: positive predictive value |
RDR: referable diabetic retinopathy |
RS: reference standard |
STARD: Standards for Reporting of Diagnostic Accuracy Studies |
Edited by Caroline Perrin; submitted 14.10.24; peer-reviewed by Gilbert Lim, Jocelyn J Drinkwater, Sanatosh Vipparthi; final revised version received 08.06.25; accepted 08.06.25; published 09.09.25.
Copyright© Mona Duggal, Anshul Chauhan, Vishali Gupta, Ankita Kankaria, Deepmala Budhija, Priyanka Verma, Vaibhav Miglani, Preeti Syal, Gagandeep Kaur, Lakshay Kumar, Naveen Mutyala, Rishabh Bezbaruah, Nayanshi Sood, Ashleigh Kernohan, Geeta Menon, Luke Vale. Originally published in JMIR Medical Informatics (https://medinform.jmir.org), 9.9.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR Medical Informatics, is properly cited. The complete bibliographic information, a link to the original publication on https://medinform.jmir.org/, as well as this copyright and license information must be included.