Introduction
Age-related macular degeneration (AMD) is a progressive retinal disorder affecting millions of people worldwide []. In its advanced stages, characterized by neovascularization and geographic atrophy (GA), it can lead to significant vision loss, although symptoms may be subtle during the early and intermediate phases []. The Classification of Atrophy Meetings group has defined atrophy lesion development as incomplete retinal pigment epithelium (RPE) and outer retinal atrophy and complete RPE and outer retinal atrophy (cRORA) based on imaging methods []. GA, also known as cRORA, is the endpoint of dry AMD and is characterized by the loss of photoreceptors, RPE, and choriocapillaris [,]. With the advent of 2 approved therapies for GA secondary to AMD in 2023, namely pegcetacoplan (Syfovre) [] and avacincaptad pegol [], the treatment of GA represents a significant breakthrough. However, the effectiveness of these therapies relies heavily on early detection and the ability to monitor treatment response—a significant unmet need in current clinical practice. The recent approval of complement inhibitors underscores the necessity for precise, reproducible, and practical tools to not only identify GA at its earliest stages but also to objectively track morphological changes over time, thereby evaluating therapeutic efficacy [,]. Artificial intelligence (AI) is uniquely positioned to address this gap by enabling precise, reproducible, and automated quantification of GA progression and treatment response using noninvasive imaging modalities []. Unlike conventional methods that rely on subjective and time-consuming manual assessments, AI algorithms can detect subtle subclinical changes in retinal structures—such as photoreceptor integrity loss, RPE atrophy, and hyperreflective foci—long before they become clinically apparent. Thus, AI-based retinal imaging offers a critical foundation for early detection and timely intervention in GA.
Various imaging techniques, both invasive and noninvasive, can directly visualize GA lesions. Invasive methods, such as fluorescence angiography, often result in a poor patient experience and entail high costs due to pupil dilation and sodium fluorescein injection. While it remains the gold standard for assessing neovascular AMD and offers significant diagnostic insights for retinal vascular diseases, in most cases, noninvasive fundus images are used for GA diagnosis and management []. Color fundus photography (CFP), fundus autofluorescence (FAF), and near-infrared reflectance (NIR) are based on 2D images, which can generally produce results to quantify the atrophic area but fail to identify the retinal structure axially []. Compared with fundus imaging, optical coherence tomography (OCT) provides high-resolution, noninvasive 3D images of retinal structures for macular assessment. In addition, conventional B-scan (axial direction) OCT images can be integrated with en-face scans, facilitating the identification of atrophy borders similar to FAF [,]. Nonetheless, manual labeling is tedious, time-consuming, and impractical in a clinical setup []. There is an urgent and unmet need for early detection and management of GA using retinal image modalities. Recent advancements in AI, especially deep learning (DL), present a promising opportunity for enhancing GA detection, classification, segmentation, quantification, and prediction.
In the 1950s, AI referred to computer systems capable of performing complex tasks that historically only a human could do. So what is AI? How is it used in medicine today? And what may it do in the future? AI refers to the theory and development of computer systems capable of performing tasks that historically required human intelligence, such as recognizing speech, making decisions, and identifying patterns. AI is an umbrella term that encompasses a wide variety of technologies, including machine learning (ML) and DL []. ML is a subfield of AI that uses algorithms trained on datasets to create self-learning models capable of predicting outcomes and classifying information without human intervention []. ML refers to the general use of algorithms and data to create autonomous or semiautonomous machines. DL, meanwhile, is a subset of ML that layers algorithms into “neural networks” with 3 or more layers. Thus, it somewhat resembles the human brain, enabling machines to perform increasingly complex tasks []. DL algorithms generally have high and clinically acceptable diagnostic accuracy across different areas (ophthalmology, respiratory, breast cancer, etc) in radiology []. Within ophthalmology, DL algorithms showed reliable performance for detecting multiple findings in macular-centered retinal fundus images []. Therefore, automatic GA segmentation plays a vital role in the diagnosis and management of advanced AMD and its application in the clinical setting.
Given the rapid evolution of AI applications in ophthalmology and the growing clinical importance of GA, this study aimed to systematically review the current evidence on AI-based approaches for the detection and management of GA secondary to dry AMD using noninvasive imaging modalities. We aimed to evaluate diagnostic accuracy relative to reference standards and examine methodological challenges to inform the design of future research and clinical implementation.
Methods
Protocol and Registration
Before starting this systematic review and meta-analysis, we registered a protocol on the PROSPERO website. This review adhered to the PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) and PRISMA-DTA (PRISMA of Diagnostic Test Accuracy) checklists [,].
Eligibility Criteria
We included studies using AI algorithms to detect, classify, identify, segment, quantify, or predict GA secondary to AMD from CFP, OCT, OCT angiography, FAF, or NIR. The data were from participants, with or without symptoms, who were diagnosed with GA (or cRORA) secondary to nonexudative AMD. Study designs were not restricted; multicenter or single-center, prospective or retrospective, post hoc analysis, clinical study, or model development studies were all accepted. Eyes with neovascular complications or macular atrophy from causes other than AMD, any previous anti-vascular endothelial growth factor treatment, any confounding retinopathy, or poor image quality were excluded.
Electronic Search Strategy
Two consecutive searches were conducted on PubMed, Embase, Web of Science, Scopus, Cochrane Library, and CINAHL. Because this review required the extraction of baseline data and items, considering the completeness of the data, we did not conduct any in press or print source searches and excluded conference proceedings and similar materials. The initial search was completed from the date of entry to December 1, 2024; the updated search, from December 1, 2024, to October 5, 2025. We used a search strategy for patient (GA) and index tests (AI and retinal images) that had been used in previous Cochrane Review without any search peer review process []. There were no restrictions on the date of publication. The language was limited to English. In , detailed search strategies for each database are provided. During this process, no filters were used. During the search process, we adhered to the PRISMA-S (Preferred Reporting Items for Systematic reviews and Meta-Analyses literature search extension) reporting guidelines [].
Selection Process
All relevant literature was imported into EndNote (version 20; Clarivate Analytics) software, and literature screening was conducted independently by 2 researchers (NS and JL) who specialize in ophthalmology. Duplicates were removed from the software, and the titles and abstracts of the literature were reviewed to identify those relevant to the topic. Finally, the full texts were downloaded and examined, leading to the selection of literature that met the inclusion criteria. In cases of inconsistencies in the final inclusion decisions made by the 2 researchers, a third professional (LL) was consulted to resolve the dispute.
Data Collection Process
Using standardized data items, the data were extracted independently from the included studies by 2 researchers (NS and JL). A third review author (LL) confirmed or adjudicated any discrepancies through group discussion. We retrieved the following data items: (1) study characteristics (author, year, study design, region, and theme), (2) dataset characteristics (databases, source of databases, training/validation/testing ratio, patient number, number of images or volumes, scan number, mean age, clinical registration number, and model evaluation method), (3) image and algorithm characteristics (devices, metrics, image modality, image resolution, and AI algorithms), (4) performance metrics (outcomes, performance of models, ground truth, and performance of the ophthalmologists), and (5) main results. All the information was retrieved from the main text and the tables provided in . Therefore, we did not seek additional data by contacting the authors or experts. In some studies, the authors reported multiple sets of performance data based on a subset of a single dataset. For example, they may have reported results such as sensitivity, specificity, accuracy, and so forth, conducted on the cross-validation set, the test set, or the development set. We referred to the relevant literature to select the optimal set of test performance results []. However, when the primary study provided performance results based on a single test, the development dataset was used to train the AI model, and an external validation set ultimately was used to determine the performance of the optimal model. We extracted the external validation set performance data [].
Risk of Bias and Application
We worked in pairs to assess the risk of bias and the applicability of the studies, which involved detection, classification, identification, segmentation, and quantification using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS)-AI [] and the modified QUADAS-2 tool [], while predictive studies used the Prediction Model Risk of Bias Assessment Tool (PROBAST) [].
In the current context, QUADAS-AI has not yet established a complete specification of items. Therefore, we referenced the examples provided by QUASAS-AI and the published literature to compile the revised QUADAS-AI items, which included 4 domains and 9 leading questions (Table S4 in ). The PROBAST tool comprises participants, predictors, outcomes, and analysis, containing 20 signaling questions across 4 domains (Table S5 in ). We also evaluated the applicability of the study based on the leading or signaling questions in the first 3 domains. A study with “yes” answers to all index questions was considered to have a low risk of bias. If the answer to any of the informational questions was “no,” there was a potential for bias, leading the authors to rate the risk of bias as high. “Indeterminate” grades were applied when detailed content was not provided in the literature, making it difficult for the evaluator to reach a judgment. They were used only when the reported data were insufficient. Throughout the process, disagreements between the 2 reviewers (NS and JL) were resolved by consulting the senior reviewer (LL).
Data Synthesis
As very few studies reported the number of true positives, true negatives, false positives, and false negatives, we restricted the quantitative analysis to determine the diagnostic accuracy of AI as a triaging tool for GA secondary to nonexudative AMD. However, a meta-analysis was not performed due to significant methodological heterogeneity across studies, arising from diverse AI architectures, imaging modalities, outcome metrics, and validation protocols. Instead, a systematic review was performed to qualitatively summarize performance trends. This approach allowed for a comprehensive evaluation of the AI capabilities in the detection and management of GA via noninvasive images.
Results
Study Selection
A total of 979 records related to the topic of this systematic review were searched across 6 different databases using a combination of subject terms and free-text terms. After removing duplicates, 335 records remained and were examined for titles and abstracts. Excluding studies not relevant to the research topic resulted in 200 reports. The full texts were then downloaded and reviewed in detail based on the eligibility criteria for the studies. In the final qualitative analysis, 41 studies were included. Of these, 10 focusing on GA diagnosis, 20 on GA assessment and progression, and 11 on GA prediction. presents the detailed flow diagram of the literature selection.
AI in Detecting the Presence of GA
Ten of the 41 included studies focused on AI-based detection of GA using noninvasive retinal images (Table S1 in ). As listed in , the studies were published from 2018 to 2025. Four of the studies [-] focused on model development, 3 [-] were retrospective studies, and 3 [-] were prospective studies (1 multicenter cohort study, 1 multicenter and low-interventional clinical study, and 1 clinical study). Geographically, half were from the United States, with others from Israel, Italy, Switzerland, Germany, and a multicenter European collaboration. The studies addressed several detection-related tasks: 5 focused solely on GA detection [-,,], 2 covered detection and classification [,], and others integrated detection with quantification or segmentation [,,].
| Author | Study design | Region | Purpose of the study | Source of datasets | Number of patients | Number of images or scans | Model evaluation method | Image modality (image resolution) | AI algorithms | Outcomes | Performance of models |
| Fineberg et al [] | Retrospective cohort study | Israel (Petah Tikva) | Detection and classification (GA) | Rabin Medical Center | 113 | 659 | 10-fold cross-validation | NIR (640*640 pixels) | CNNs: ResNet50, EfficientNetB0, ViT_B_16, and YOLOv8 variants. | ACC, P, SEN, SPE, F1, IoU, and DSC |
|
| Kalra et al [] | Retrospective clinical study | United States (Cleveland) | Detection, quantification, and segmentation (presence of GA and pixel-wise GA area measurement) | the Cole Eye Institute of the Cleveland Clinic | 341 | 900 | triple-fold cross-validation | SD-OCT (256*256 pixels) | CNN: U-Net | F1, ACC, P, R, SEN, and SPE |
|
| Derradji et al [] | Retrospective clinical study | Switzerland (Lausanne) | Detection and segmentation (RORA) | An existing image database of the Medical Retina Department at Jules-Gonin Eye Hospital | 57 | 62 | 5-fold cross-validation | SD-OCT (NR) | CNN: U-Net | SEN, DSC, P, and Kappa |
|
| de Vente et al [] | Prospective multicenter and low-interventional clinical study (including cross-sectional and longitudinal study part) | 20 sites in 7 European countries | Detection and quantification (cRORA) | The MACUSTAR Study Cohort | 168 | 143 (ZEISS); 167 (Spectrails) | NR | SD-OCT (512*650 pixels) | CNN: U-Net | SEN, SPE, PPV, NPV, and Kappa |
|
| Sarao et al [] | Prospective clinical study | Italy (Udine) | Detection (presence of GA) | the Istituto Europeo di Microchirurgia Oculare (IEMO) study | 180 | 540 | NR | CFP (NR) | CNN: Efficientnet_b2 | SEN, SPE, ACC, F1, R, AUROC, and AUPRC |
|
| Keenan et al [] | Multicenter and prospective cohort study | United States (Maryland) | Detection (presence of GA) | Age-Related Eye Disease Study (AREDS) dataset | 4582 | 59,812 | 5-fold cross-validation | CFP (512 pixels) | CNN: inception v3 | ACC, SEN, SPE, P, AUC, and Kappa |
|
| Yao et al [] | Model development and evaluation | United States (California) | Detection (presence of nGA) | the Early Stages of AMD (LEAD) study | 140 | 1884 | 5-fold cross-validation | SD-OCT (512*496 pixels) | CNN: ResNet18 | SEN, SPE, ACC, P, and F1 |
|
| Chiang et al [] | Model development | United States (California) | Detection (complete retinal pigment epithelial and outer retinal atrophy (cRORA) in eyes with AMD) | University of Pennsylvania, University of Miami, and Case Western Reserve University; (2) Doheny Image Reading Research Laboratory, Doheny-UCLA (University of California Los Angeles Eye Centers) | 71 (training); 649 (testing #1); 60 (testing #2) | 188 (training); 1117 (testing #1) | 5-fold cross-validation | SD-OCT (256*256 pixels) | CNN: ResNet18 | SEN, SPE, PPV, NPV, AUROC, and AUPRC |
|
| Elsawy et al [] | Model development | United States (Maryland) | Detection (explain decision making and compare methods) | The Age-Related Eye Disease Study 2 (AREDS2) Ancillary SD-OCT study from Devers Eye Institute, Emory Eye Center, Duke Eye Center, and the National Eye Institute | 311 | 1284 scans | 10-fold cross-validation | SD-OCT (128*128 or 224* pixels) | 3D CNN: deep-GA-Net | ACC, P, R, F1, Kappa, AUROC, and AUPRC |
|
| Treder et al [] | Model development | Germany (Muenster) | Detection and classification (GA) | Public database: ImageNet | 400 (training); 60 (test set) | 400 (training); 60 (test set) | NR | FAF (NR) | Deep CNN: self-learning algorithm | SEN, SPE, and ACC |
|
aAI: artificial intelligence.
bACC: accuracy.
cAUPRC: area under the precision-recall curve.
dCNN: convolutional neural network.
eCFP: color fundus photography.
fcRORA: complete retinal pigment epithelium and outer retinal atrophy.
gDSC: dice similarity coefficient.
hFAF: fundus autofluorescence.
iIoU: intersection over union.
jNR: not reported.
kOCT: optical coherence tomography.
lPPV: positive predictive value.
mP: precision.
nR: recall.
oSD-OCT: spectral domain OCT.
pSEN: sensitivity.
qSPE: specificity.
rAUROC: area under the receiver operating characteristic curve.
sAMD: age-related macular degeneration.
tNPV: negative predictive value.
Dataset configurations varied: 6 studies used training, validation, and test sets [-,,]; 3 used only training and test sets [,,]; and 1 included a tuning set []. Collectively, these studies involved at least 7132 participants, with ages ranging from 50 to 85 years. Three studies were registered with ClinicalTrials.gov (NCT00734487, NCT01790802, and NCT03349801) [,,]. Cross-validation methods included 5-fold (40% of studies) [,,,], 10-fold (20%) [,], and triple-fold (10%) []; 30% did not report validation details.
Spectral-domain (SD)–OCT was the most frequently used imaging modality (6/10 of studies) [-,,,], followed by CFP (2/10) [,], and FAF or NIR (2/10 each) [,]. Most studies applied image preprocessing techniques—such as size standardization, orientation adjustment, intensity normalization, and noise reduction—to improve model performance. DL-based algorithms for GA detection have been developed for multiple image modalities. For example, Derradji et al [] trained a convolutional neural networks (CNNs), U-Net, to predict atrophic signs in the retina, based on the EfficientNet-b3 architecture. Kalra et al [] and de Vente et al [] also trained a DL model based on U-Net. Yao et al [] applied 3D OCT scans with ResNet18 pretrained on the ImageNet dataset, and Chiang et al [] developed CNN (ResNet18) to improve computational efficiency. Elsawy et al [] proposed Deep-GA-Net, a 3D backbone CNN with a 3D loss-based attention layer, and evaluated the effectiveness of using attention layers. Sarao et al [] used a deep CNN, the EfficientNet_b2 model, which was pretrained on the ImageNet dataset and is well-known for its high efficiency and small size. Keenan et al [] established their model using Inception v3, while Treder et al [] performed a deep CNN, a self-learning algorithm, processing input data with FAF images.
A total of 14 performance sets were extracted from the 10 studies. Key metrics included sensitivity, specificity, accuracy, positive predictive value, negative predictive value, intersection over union, area under the receiver operating characteristic curve, area under the precision-recall curve, F1-score, precision, recall, Kappa, and dice similarity coefficient. Six OCT-based studies showed that DL models could detect GA with high accuracy, comparable to human graders [-,,,]. Two studies using CFP also reported strong performance [,], while FAF- and NIR-based approaches demonstrated excellent repeatability and reliability [,].
We conducted a thorough evaluation of the 10 diagnostic studies’ methodological quality for the “participant selection,” “index test,” “reference standard,” and “flow and timing” domains at the study level (). None of the studies had an overall low or unclear risk of bias; instead, every study had a high risk of bias in at least 1 of the 4 domains. Regarding “patient selection,” only 4 studies [,,,] described the eligibility criteria; the rest did not report them. One study [] used an open dataset (ImageNet) and did not include a test set. The small sample size of 4 studies [,,,] may have resulted in overfitting. In addition, 3 studies [,,] did not report image formats and resolutions. Five studies [,,-] had a high risk of bias in participant selection because the included participants were not only GA secondary to dry AMD but also had other unrelated diseases. Regarding the “Index test,” only 1 algorithm was externally validated using a different dataset []; all other items were evaluated as low risk.
| Study | Risk of bias | Concerns regarding applicability | |||||
| Patient selection | Index test | Reference standard | Flow and timing | Patient selection | Index test | Reference standard | |
| Chiang et al [] | High risk | Low risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| Elsawy et al [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| Kalra et al [] | High risk | High risk | Low risk | Low risk | High risk | Low risk | Low risk |
| Keenan et al [] | High risk | High risk | Low risk | Low risk | High risk | Low risk | Low risk |
| Sarao et al [] | High risk | High risk | Low risk | Low risk | High risk | Low risk | Low risk |
| Yao et al [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| Treder et al [] | High risk | High risk | Low risk | Low risk | High risk | Low risk | Low risk |
| Vente et al [] | High risk | High risk | Low risk | Low risk | High risk | Low risk | Low risk |
| Derradji et al [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| Fineberg et al [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
AI in GA Assessment and Progression
Twenty studies explored AI for GA assessment and progression using noninvasive imaging, published between 2019 and 2025 (Table S2 in ). As shown in , these studies covered 11 segmentation [,,-], 2 algorithm optimization [,], 3 AMD progression classification [-], and 3 combined tasks such as identification, segmentation, and quantification [-]. One study focused solely on GA quantification []. Retrospective analyses accounted for 9 studies [,,,,,,,,], while 7 were model development [,-,,,], and the remainder were prospective [,], comparative [], or cross-sectional []. Geographically, contributions came from China (6/20), the United States (7/20), the United Kingdom (2/20), Australia (2/20), France (1/20), Israel (1/20), and Austria (1/20).
| Author | Study design | Region | Purpose of the study | Source of datasets | Number of patients | Number of images or scans | Model evaluation method | Image modality (Image resolution) | AI algorithms | Outcomes | Performance of models |
| Pramil et al [] | Retrospective review of images | United States (Boston) | Segmentation (GA lesions) | The “SWAGGER” cohort of the non-Exudative Age-Related Macular Degeneration (from New England Eye Center at Tufts Medical Center) | 90 | 126 | 5-fold cross-validation | SS-OCT (500*500 pixels) | CNN: U-Net | SEN, SPE, and DICE |
|
| Siraz et al [] | Retrospective comparative study | United States (North Carolina) | Classification (central and noncentral GA) | Atrium Health Wake Forest Baptist | 104 | 355 | NR | SD-OCT (224*224 pixels) | CNNs: ResNet50, MobileNetV2, and ViT-B/16 | AUROC, F1, and ACC |
|
| Arslan et al [] | Retrospective cohort clinical study | Australia (Victoria) | Segmentation (GA lesion area) | The Center for Eye Research Australia or a private ophthalmology practice diagnosed with GA | 51 | 702 | 5-fold cross-validation | FAF (768*768 or 1536*1536 pixels) | CNN: U-Net | DSC, DSC loss, SEN, SPE, MAE, ACC, R, and P |
|
| Hu et al [] | Retrospective clinical study | China (Shenyang) | Classification (dry AMD progression phases) | Shenyang Aier Eye Hospital | 338 | 3401 | 5-fold cross-validation | SD-OCT (NR) | CNNs: EfficientNetV2, DenseNet169, Xception, and ResNet50NF | ACC, SEN, SPE, F1, Macro-f1, and Kappa |
|
| Spaide et al [] | Retrospective analysis and model comparison | United States (Washington) | Segmentation (GA lesion area) | The SWAGGER cohort from the New England Eye Center at Tufts Medical Center | 87 | 126 scans | 5-fold cross-validation | SS-OCT (NR) | CNN: U-Net | DSC |
|
| Vogl et al [] | Retrospective analysis | Austria (Vienna) | Identification (GA progression after pegcetacoplan treatment) | The FILLY trial | 156 | NR | NR | SD-OCT (512*512 pixels) | CNN: U-Net | LPR |
|
| Szeskin et al [] | Retrospective analysis | Israel (Jerusalem) | Identification, quantification (GA lesion) | Datasets D1 and D2 from the Hadassah University Medical Center | D1: 18; D2: 16 | NR | 4-fold cross-validation | SD-OCT (496*1024 pixels and 496*1536 pixels) | CNN: the custom column classification CNN | AUROC, P, R, and F1 |
|
| Spaide et al [] | Retrospective analysis | United States (California) | Segmentation (GA lesion area) | Proxima A and B | Proxima A: 154; Proxima B: 183 | Proxima A: 497; Proxima B: 940 | NR | FAF, NIR (768 *768 pixels) | Multimodal DL: U-Net; YNet | DSC and r2 |
|
| AI-khersan et al [] | Retrospective analysis | United States (Texas) | Segmentation (GA) | The Retina Consultants of Texas and Retina Vitreous Associates | 33; 326 | 367; 348 | 5-fold cross-validation | SD-OCT (512*496pixels; 200*1024pixels) | CNN: 3D-to-2D U-Net | DSC and r2 |
|
| Chu et al [] | Prospective study | United States (Washington) | Identification, segmentation, and quantification (GA) | The University of Miami | 70; 20; 25 | NR | NR | SS-OCT (512*512 pixels) | CNN: U-Net | DSC, SEN, and SPE |
|
| Merle et al [] | Prospective observational study | Australia (Victoria) | Quantification (GA) | The Center for Eye Research Australia | 50 | NR | NR | SD-OCT; FAF (NR) | CNN: U-Net | Spearman correlation coefficient and SEN |
|
| Yang et al [] | Model development | China (Shenyang) | Classification (stage of dry AMD progression) | Shenyang Aier Excellence Eye Hospital | 1310 | 16,384 | 3-fold cross-validation | SD-OCT (NR) | CNNs: ResNet50, EfficientNetB4, MobileNetV3, Xception | ACC, SEN, SPE, and F1 |
|
| Ji et al [] | Model development | China (Nanjing) | Segmentation (GA lesion area) | Dataset1 and dataset2 | 8; 54 | NR | NR | SD-OCT (224*224 pixels) | Weakly supervised multitask learning: Mirrored X-Net | DSC, IoU, AAD, and CC |
|
| Ma et al [] | Model development | China (Jinan) | Segmentation (GA lesion area) | Dataset1 and dataset2 | 62 | NR | 5-fold cross-validation | SD-OCT (224*224 pixels) | Weakly supervised model: VGG16 | DSC, OR, AAD, CC, and AUROC |
|
| Royer et al [] | Model development | France (Issy-Les-Moulineaux) | Segmentation (GA lesion area) | the Clinical Imaging Center of the Quinze-Vingts Hospital | 18 | 328 | 8 different random combinations of 12 series to train the model and 6 for the tests | NIR (256*256 pixels) | Unsupervised neural networks: W-net | F1, P, and R |
|
| Xu et al [] | Model development | China (Jinan) | Segmentation (GA lesion area) | dataset1 and dataset2 | 8 (test I); 56 (test II) | 55 (dataset1); 56 (dataset2) | NR | SD-OCT (1024*512*128pixels; 1024*200*200pixels) | Self-learning algorithm | OR, AAD, and CC |
|
| Zhang et al [] | Model development | United Kingdom (London) | Segmentation and quantification (GA) | The FILLY study | 200 | 984 | NR | SD-OCT (NR) | CNN: U-Net | DSC, ICC, ACC, SEN, SPE, and F1 |
|
| Liu et al [] | Model development | China (Wuhan) | Segmentation (GA) | Wuhan Aier Eye Hospital; the public dataset OCTA500 | 300 | 2923 | 5-fold cross-validation | SD-OCT (512*512 pixels) | Self-learning algorithm (dual-branch image projection network) | Jaccard index, DSC, ACC, P, and R |
|
| Williamson et al [] | Cross-sectional study | United Kingdom (London) | Segmentation (GA lesion area) | INSIGHT Health Data Research Hub at Moorfields Eye Hospital | 9875 (OCT); 81 (FAF) | NR | NR | 3D-OCT; FAF (512*512 pixels) | Self-learning algorithm | PPV | |
| Safai et al [] | Comparative analysis | United States (Wisconsin) | Identification (the best AI framework for segmentation of GA) | AREDS2 study; the GlaxoSmithKline (GSK) study | 271(AREDS2); 100(GSK) | 601 (AREDS2); 156 (GSK) | 5-fold cross-validation | FAF (512*512 pixels) | CNNs: UNet, FPN, PSPNet, EfficientNet, ResNet, VGG, mViT | CC and DSC |
|
aSS-OCT: swept-source OCT.
bCNN: convolutional neural network.
cSEN: sensitivity.
dSPE: specificity.
eDSC: dice similarity coefficient.
fNR: not reported.
gSD-OCT: spectral domain OCT.
hAUROC: area under the receiver operating characteristic curve.
iACC: accuracy.
jCGA: central geographic atrophy.
kNCGA: noncentral geographic atrophy.
lFAF: fundus autofluorescence.
mMAE: mean absolute error.
nR: recall.
oP: precision.
pAMD: age-related macular degeneration.
qLPR: local progression rate.
rNIR: near-infrared reflectance.
sDL: deep learning.
tr2: Pearson correlation coefficient.
uOCT: optical coherence tomography.
vIoU: intersection over union.
wAAD: absolute area difference.
xCC: correlation coefficient.
yOR: overlap ratio.
zICC: intraclass coefficient.
aaPPV: positive predictive value.
abAREDS2: Age-Related Eye Disease Study 2.
acFPN: Feature Pyramid Network.
adVGG: Visual Geometry Group.
aemViT: Mix Vision Transformer.
Dataset configurations varied: 9 out of 20 studies used training, validation, and test sets [,,-,-]; 11 studies used training and test sets [,,-,]; 2 studies used training and validation sets [,]; 1 study comprised training, tuning, and internal validation sets []; and 2 studies did not specify [,]. Across studies, at least 14,064 participants provided image data for analysis. Less than half of the studies (9/20, 45%) provided demographic information, with the average age of participants ranging from 55 to 94 years. Six studies were registered with ClinicalTrials.gov (NCT01342926, NCT02503332, NCT02479386, NCT02399072, and NCT04469140 [,,,,,]). To assess the generalization ability of the DL model, cross-validation methods included 5-fold (8/20 studies [,,,-,]), 4-fold (1/20 study []), 3-fold (1/20 study []), and other approaches (1/20 study []). Nine studies did not report validation specifics.
Multiple imaging modalities supported GA assessment: spectral domain optical coherence tomography (SD-OCT) was most common, followed by swept-source OCT (SS-OCT), 3D-OCT, FAF, and NIR. Preprocessing techniques were widely applied to standardize images and improve model performance. Algorithm architectures varied, with U-Net being the most frequently used. Other approaches included custom CNNs, self-learning algorithms, weakly supervised models, and multimodal networks. For example, Hu et al [] trained the DL models (ResNet-50, Xception, DenseNet169, and EfficientNetV2), evaluating them on a single fold of the validation dataset, with all F1-scores exceeding 90%. Yang [] proposed an ensemble DL architecture that integrated 4 different CNNs, including ResNet50, EfficientNetB4, MobileNetV3, and Xception, to classify dry AMD progression stages. GA lesions on FAF were automatically segmented using multimodal DL networks (U-Net and Y-Net) fed with FAF and NIR images []. In contrast to the multimodal algorithms mentioned above (ie, the examples of DL models), Safai [] investigated 3 distinct segmentation architectures along with 4 commonly used encoders, resulting in 12 different AI model combinations to determine the optimal AI framework for GA segmentation on FAF images.
From 20 studies, 42 performance sets were collected. Common metrics included correlation coefficient, mean absolute error, Spearman correlation coefficient, intraclass coefficient, overlap ratio, Pearson correlation coefficient (r2), Kappa, specificity (SPE), sensitivity (SEN), accuracy, positive predictive value (PPV), F1-score, P, R, intersection over union, and dice similarity coefficient (DSC). Regarding the segmentation, classification, identification, and quantification of GA in SD-OCT, 12 studies demonstrated performance comparable to that of clinical experts [,,,,,-,,]. AI was also capable of efficiently detecting, segmenting, and measuring GA in SS-OCT, 3D-OCT, and FAF images, according to 4 studies [,,,]. AI for GA segmentation in FAF and NIR images, with clinical data showing good segmentation performance [,,].
We performed a comprehensive assessment of the methodological quality of 16 GA assessment and progression studies encompassing 4 domains (). Only 8 studies detailed the eligibility criteria in the “patient selection” category, while the others had not been published. Three of the studies [-] lacked complete datasets, and 3 others [,,] had small datasets or limited volumes of data. In addition, 3 studies [,,] failed to provide information on image formats or resolutions. Two studies [,] were ranked as high risk regarding patient selection since the participants included other types of dry AMD (drusen, nascent GA). In terms of applicability, 18 studies were classified as low risk, while 2 were deemed high risk concerning patient selection. Concerning the “Index test,” only 3 algorithms underwent external validation with a different dataset [,,]. All other items were evaluated as low risk.
| Study | Risk of bias | Concerns regarding applicability | |||||
| Patient selection | Index test | Reference standard | Flow and timing | Patient selection | Index test | Reference standard | |
| M Hu [] | High risk | High risk | Low risk | Low risk | High risk | Low risk | Low risk |
| JK Yang [] | High risk | High risk | Low risk | Low risk | High risk | Low risk | Low risk |
| A Safai [] | Low risk | Low risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| WD Vogl [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| A Szeskin [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| ZD Chu [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| ZX Ji [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| X Ma [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| C Royer [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| T Spaide [] | High risk | Low risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| T Spaide [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| DJ Williamson [] | Low risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| RB Xu [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| J Arslan [] | Low risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| V Pramil [] | Low risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| GY Zhang [] | High risk | Low risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| DA Merle [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| H AI-khersan [] | Low risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| S Siraz [] | Low risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
| XM Liu [] | High risk | High risk | Low risk | Low risk | Low risk | Low risk | Low risk |
AI in Predicting GA Lesion Area and Progression
Eleven studies used AI for predicting GA lesion growth and progression using noninvasive imaging (Table S3 in ). These studies were published between 2021 and 2025, with some information provided in . The study designs consisted of 6 retrospective studies [-], 2 model development studies [,], 2 post hoc analyses [,], and 1 clinical evaluation of a DL algorithm []. Participants or images came from various regions: 6 studies were based in the United States [,-,], 3 in Australia [-], 1 in Switzerland [], and another involving multiple centers in China and the United States []. Research aims focused on GA growth prediction [,,-,,], combined prediction and evaluation of lesion features [], treatment response assessment [], and integrated segmentation-prediction tasks [,].
| Author | Study design | Region | Purpose of the study | Source of datasets | Number of patients | Number of images or scans or cubes | Model evaluation method | Image modality (resolution) | AI algorithms | Outcomes | Performance of models |
| Gigon et al [] | Retrospective monocentric study | Switzerland (Lausanne) | Prediction (RORA progression) | Jules Gonin Eye Hospital | 119 | NR | NR | SD-OCT (384*384 pixels) | CNN: EfficientNet-b3 | DSC |
|
| Dow et al [] | Retrospective cohort study | United States (Atlanta, Georgia, Portland, Oregon, North Carolina; Maryland, Raleigh, Morrisville, Cary); United Kingdom (Durham, South Durham) | Prediction (iAMD to GA within 1 year) | 3 independent datasets from AREDS2 and a tertiary referral center and associated satellites | 316; 53; 48 | 1085; 53; 48 | 5-fold cross-validation | SD-OCT (512 *1000 pixels) | CNN: Inception v3 | SEN, SPE, PPV, NPV, ACC |
|
| Cluceru et al [] | Retrospective clinical study; observation study | United States (California) | Prediction and evaluation (GA growth rate and GA features related to shape and size) | The lampalizumab phase 3 clinical trials and an accompanying observational study | 1041; 255 | NR | 5-fold cross-validation | FAF (384 * 384 pixels) | CNN: VGG16 | r2 |
|
| Anegondi et al [] | Retrospective clinical study; observation study | United States (California) | Prediction and prognosis (GA lesion area and GA growth rate after lampalizumab treatment) | The lampalizumab phase 3 clinical trials and an accompanying observational study | 1279; 443; 106; 169 | NR | 5-fold cross-validation | SD-OCT, FAF (512*512 pixels) | CNN: Inception v3 | r2 | GA prediction:
GA growth rate:
|
| Salvi et al [] | Retrospective analysis | United States (California) | Prediction (the 1 year region of growth of GA lesions) | The following lampalizumab clinical trials and prospective observational studies | 597 | NR | NR | FAF (768*768 pixels or 1536*1536 pixels) | CNN: U-Net | P, R, DSC, r2 | Whole lesion:
|
| Yoshida [] | Retrospective analysis | United States (California) | Prediction (GA progression) | Three prospective clinical trials | 1219; 442 | NR | 5-fold cross-validation | 3D OCT (496*1024*49 voxels) | CNNs: (1) en-face intensity maps; (2) SLIVER-net; (3) a 3D CNN; and (4) en-face layer thickness and between-layer intensity maps from a segmentation model | r2 |
|
| GS Reiter [] | Post hoc analysis | Austria (Vienna) | Prediction (GA lesions progression) | the phase II randomized controlled trial FILLY | 134 | 268 scans | 5-fold cross-validation | FAF, NIR, SD-OCT (NR) | CNN: PSC-UNet | ACC, Kappa, concordance index |
|
| J Mai [] | Post hoc analysis | Austria (Vienna) | Segmentation, quantification, and prediction (GA lesion and progression) | The phase 2 FILLY clinical trial and the Medical University of Vienna (MUV) | 113; 100 | 226; 967 | 5-fold cross-validation | SD-OCT, FAF (768*768 and 1536*1536 pixels) | CNN: U-Net | DSC, Hausdorff distance, ICC |
|
| YH Zhang [] | Model development | China (Nanjing); United States (California) | Prediction (GA growth) | The Byers Eye Institute of Stanford University; the Jiangsu Provincial People’s Hospital | 22; 3 | 86 cubes; 33 cubes | Leave-one-out cross-validation | SD-OCT (178*270 pixels) | Recurrent neural network: the bi-directional long-short term memory network; CNN: 3D-UNet | DSC, CC |
|
| SX Wang [] | Model development | United States (California) | Segmentation and prediction (GA lesion area and GA progression) | The University of California—Los Angeles | 147 | NR | 8-fold cross-validation | SD-OCT, FAF (512*512 pixels) | CNN: U-Net | SEN, SPE, ACC, OR |
|
| J Mai [] | Clinical evaluation of a DL-based algorithm | Austria (Vienna) | Prediction (GA lesions progression) | The Medical University of Vienna | 100 | 967 | 5-fold cross-validation | SD-OCT, FAF (NR) | CNN: PSC-UNet | DSC, MAE, and r2 |
|
aRORA: retinal pigment epithelial and outer retinal atrophy.
bNR: not reported.
cOCT: optical coherence tomography.
dCNN: convolutional neural network.
eDSC: dice similarity coefficient.
fAMD: age-related macular degeneration.
gAREDS2: Age-Related Eye Disease Study 2.
hSEN: sensitivity.
iSPE: specificity.
jPPV: positive predictive value.
kNPV: negative predictive value.
lACC: accuracy.
mFAF: fundus autofluorescence.
nr2: Pearson correlation coefficient.
oP: precision.
pR: recall.
qEZ: ellipsoid zone.
rRPE: retinal pigment epithelium.
sNIR: near-infrared reflectance.
tICC: intraclass coefficient.
uCC: correlation coefficient.
vOR: overlap ratio.
wMAE: mean absolute error.
Dataset structures varied: 3 out of 11 studies used training-validation-test splits [,,]; 2 out of 11 studies used training-test sets [,]; 3 out of 11 studies used training-validation sets [,,]; and the rest adopted development–holdout [,] or development-holdout-independent test configurations []. In total, 6706 participants were included across studies. Fewer than half of the studies (4/11, 36.4%) reported demographic information, with mean age ranges spanning from 74 to 83 years [,,,]. Six studies [-,,] were ethically approved and registered on ClinicalTrials.gov under the following identifiers: NCT02503332, NCT02247479, NCT02247531, NCT02479386, NCT01229215, and NCT02399072. The DL model’s generalizability was assessed using leave-one-out cross-validation in 1 study [], 5-fold cross-validation in 7 studies [,,,,-], and 8-fold cross-validation in 1 study []. The remaining 2 studies [,] did not specify the cross-validation methodology.
Studies used 3D-OCT, SD-OCT, NIR, and FAF images, primarily sourced from Heidelberg, Zeiss, and Bioptigen devices. While most reported image metrics, 2 studies did not specify resolution details [,]. Commonly used DL architectures included Inception v3 [,], PSC-UNet [,], U-Net [,,], EfficientNet-b3 [], and VGG16 []. In addition, some studies introduced novel approaches, such as en-face intensity maps, SLIVER-net, 3D CNN, and a recurrent neural network, for improved GA progression forecasting.
According to various image modalities, datasets, and follow-up durations, we gathered 31 sets of performance data from 11 studies. The performance metrics included the Hausdorff distance, concordance index, overlap, SEN, SPE, accuracy, mean absolute error, Kappa, DSC, P, PPV, R, r2, and negative predictive value. The findings for a single image modality (3D-OCT, SD-OCT, or FAF) demonstrated the development of DL algorithms to predict GA growth rate and progression with excellent performance characteristics comparable to trained experts [-,-]. Multimodal approaches combining FAF, NIR, and SD-OCT further showed feasibility for individualized lesion growth prediction and localization [,-].
In this systematic review, we used the PROBAST tool to rigorously evaluate prediction models across 4 domains, addressing 20 signaling questions for each paper reviewed. Within the “participants” domain, all studies used appropriate data sources; however, only 6 studies [-,,] clearly outlined their inclusion and exclusion criteria for participants, leaving the others unclear. In terms of “predictors,” these were defined and evaluated similarly for all participants, having no connection to outcome data and being available at baseline. All studies evaluated “yes” to the questions on outcome measurement methods, definitions, interference factors, and measurement time intervals. Concerning “analysis,” Dow [] and Zhang [] applied a small dataset with an insufficient number of participants. While Zhang performed internal validation, the lack of external validation notably limits the model’s generalizability, which was constructed with bi-directional long-short term memory networks and CNN frameworks. Two studies by Salvi [] and Yoshida [] lacked independent and external validation. Gigon [] failed to explicitly mention missing data handling, complex problems, and model overfitting. Conversely, all other items were evaluated as low risk, and the applications of the studies were universally ranked as low risk (Table S1 in ).
Discussion
Principal Findings
This systematic review evaluated the performance of AI, particularly DL algorithms, in detecting and managing GA secondary to dry AMD using noninvasive imaging modalities. Our findings demonstrate that AI models exhibit strong capabilities in accurately detecting, segmenting, quantifying, and predicting GA progression from OCT, FAF, CFP, and NIR imaging, achieving diagnostic accuracy comparable to that of human experts. However, this review also identified several methodological challenges, such as limited sample sizes, inconsistent annotation standards, and a general lack of external validation, which may hinder the clinical generalizability and practical application of these models. Despite these limitations, AI-based tools show significant potential for future use by both specialists and nonspecialists in primary and specialty care settings.
AI in Detecting GA With OCT, FAF, NIR, and CFP Images
Ten studies published between 2018 and 2025 were included, involving at least 7132 participants aged 50 to 85 years. Half of the studies were conducted in the United States, while others originated from European countries. SD-OCT was the most frequently used imaging modality (6/10 studies), followed by CFP (2/10 studies), NIR (1/10 studies), and FAF (1/10 studies). Image preprocessing techniques, such as standardization of size, orientation, and intensity, as well as noise reduction, were consistently applied to enhance model stability and training efficiency. However, 3 studies did not report critical image parameters, such as resolution, potentially limiting reproducibility. DL-based algorithms, including CNNs, were the primary methodologies used for GA detection. Cross-validation techniques, such as 5-fold and 10-fold methods, were used in half of the studies to assess model robustness, though 3 studies did not report validation strategies. AI, particularly DL algorithms, holds significant promise for the detection of GA using noninvasive imaging modalities. OCT, CFP, NIR, and FAF each demonstrated robust diagnostic potential, with performance metrics rivaling or exceeding human expertise.
AI for GA Management With OCT, FAF, and NIR Images
A total of 20 studies (14,064 participants) were published between 2019 and 2025, focusing on themes such as GA segmentation, classification, quantification, and progression prediction. The research designs and geographic regions are diverse. The studies included retrospective analysis (9/20), model development (7/20), and prospective, comparative, or cross-sectional studies (4/20). Significant contributions came from China (6/20) and the United States (7/20), with additional studies from the United Kingdom (2/20), Australia (2/20), France (1/20), Israel (1/20), and Austria (1/20). The studies used a variety of imaging modalities to assess GA, including SD-OCT, FAF, NIR, SS-OCT, and 3D-OCT. DL algorithms demonstrated remarkable performance in GA management tasks. U-Net was the most commonly used architecture. Multimodal approaches combined FAF and NIR images with DL networks to improve segmentation accuracy. Performance metrics, such as DSC, Kappa, SEN, SPE, and accuracy, consistently showed strong diagnostic accuracy, with several studies achieving performance comparable to clinical experts.
Eleven studies with 6706 participants, published between 2021 and 2025, concentrated on the application of AI for predicting and segmenting GA lesions, as well as their growth and progression. The methodologies were diverse, including retrospective studies, model development studies, post hoc analyses, and clinical algorithm assessment. Participants or images were gathered from regions such as the United States, Australia, Switzerland, and various centers in China and the United States, ensuring broad geographic representation. Demographic information was reported in fewer than half of the studies, with a mean age ranging from 74 to 83 years. Imaging modalities, such as 3D-OCT, SD-OCT, NIR, and FAF, were obtained from devices including Bioptigen, Heidelberg Spectralis HRA+OCT, and Cirrus OCT. While the image preprocessing parameters were consistent across most studies, some did not specify image resolution. Multiview CNN architectures and advanced frameworks, such as the bi-directional long-short term memory networks, were used. DL algorithms exhibited excellent predictive capabilities, with multimodal approaches enabling individualized GA lesion growth prediction.
Noninvasive Image Analysis Techniques for GA
GA, a late-stage form of dry AMD, is marked by the irreversible loss of photoreceptors, RPE, and choriocapillaris [,]. The application of noninvasive imaging modalities has revolutionized the detection and management of GA. A comparative summary of AI performance across these modalities is provided in Table S2 in . CFP serves as a standard initial assessment tool, useful for screening and early detection. It identifies GA lesions as visible underlying choroidal vessels and well-defined regions of RPE hypopigmentation []. FAF imaging using a blue excitation wavelength (488 nm) visualizes metabolic changes at the level of photoreceptor or RPE complex and is practical in assessing GA lesion size and progression with hypo-autofluorescence []. In contrast to nonatrophic areas, GA lesions on NIR (787-820 nm, longer than FAF) typically appear brighter and less harmful to the eye []. In addition, NIR can help detect the boundaries of foveal lesions, where image contrast is lower on FAF []. Recently, the Classification of Atrophy Meeting group recommended that atrophy in both patients with and those without neovascular AMD be defined based on specific drusen characteristics and other anatomical features, and it is most easily characterized by OCT [,]. OCT stands out as the gold standard for GA detection and classification, providing high-resolution, cross-sectional, and en face images of the retina and choroid. SD-OCT is widely used in research and clinical trials, offering precise measurement of GA area and growth rates, while SS-OCT and 3D-OCT offer superior structural insights and potential for AI-driven automation [,,]. Despite the higher cost and technical complexity of advanced OCT technologies, their detailed GA assessment capabilities make them indispensable tools in both clinical practice and research. Furthermore, OCT provides volumetric (3D) structural data, unlike the 2D en face projections of FAF, CFP, and NIR. It allows AI to learn not just the surface appearance of atrophy but also the cross-sectional structure alterations that define and precede GA []. As technology advances, the integration of AI and further developments in imaging techniques are expected to enhance the utility of these modalities, overcoming current limitations and expanding their applications in ophthalmology.
Advantages and Challenges of AI Architectures in Clinical Workflow
AI addresses critical limitations of traditional GA monitoring, such as labor-intensive manual grading and intergrader variability []. Therefore, automated algorithms enable rapid, standardized analysis of large fundus image datasets, reducing clinician workload and enhancing reproducibility []. Furthermore, our review revealed a clear trend in the choice of model architectures tailored to specific clinical tasks. A critical analysis of these architectures is provided in Table S3 in . Interestingly, with the advancement of AI algorithm architectures, numerous studies have emerged that use these technologies to identify atrophy caused by various retinal diseases and to evaluate treatment outcomes through image analysis. Miere et al [] pretrained a DL-based classifier to automatically distinguish GA from atrophy secondary to inherited retinal diseases on FAF according to etiology, using 2 approaches (a trained and validated method and a 10-fold cross-validation method), achieving good accuracy and excellent area under the receiver operating characteristic (AUROC) values. In addition, a study examined the association between treatment and changes in photoreceptor lamina thickness in patients with GA secondary to AMD. The effect of pegcetacoplan on photoreceptors in OCT was supported by this post hoc analysis, which demonstrated that treatment with the drug was linked with reduced outer retinal thinning []. Similarly, DL-based OCT image analysis assessed the therapeutic effectiveness of complement component 3 inhibition in delaying GA progression, with findings indicating decreased photoreceptor thinning and loss []. Recent studies demonstrating the application of AI algorithms in imaging further validate their potential as reliable supplements to human expertise in the diagnosis and management of GA.
Technical Challenges and Limitations
Despite the promising advancements in AI for GA detection and management, several technical challenges and limitations persist. A significant limitation of OCT-based AI models is their difficulty in distinguishing GA secondary to AMD from other forms of retinal atrophy; thus, the findings may not generalize to broader AMD cases or other retinal diseases, which limits their clinical applicability. In addition, images from different OCT devices show significant variability and imprecision, not offering good enough data acquisition []. Another major challenge is the variability in algorithm performance caused by differences in training data, image acquisition protocols, and disease definitions. These differences reduce reproducibility and limit practical deployment. For instance, the absence of standardized reporting in AI studies can result in discrepancies when interpreting results and hinder comparisons between different models. Moreover, despite the high-performance metrics (eg, SEN, SPE, DSC>0.85, and AUROC>0.95) reported by many studies, methodological limitations remain. All diagnostic studies included in this review were assessed as high risk in at least 1 domain (10/10), only 1 GA assessment study (1/20) was evaluated as low risk across all domains, and several prediction studies (7/11) were ranked as high or unclear risk in at least 1 domain, primarily due to small or nonrepresentative datasets and a lack of detailed reporting on image preprocessing and external validation. These methodological shortcomings may lead to an overestimation of AI model performance and reduced overall robustness, thereby decreasing the generalizability of the findings and limiting confidence in their real-world applicability. Future studies should prioritize the use of larger, more diverse datasets and implement rigorous validation frameworks to enhance performance metrics (including detection, segmentation, quantification, and prediction accuracy) and conduct prospective, multicenter validation studies to improve clinical applicability and generalizability. Furthermore, adherence to established reporting guidelines for AI studies (such as the Standards for Reporting Diagnostic Accuracy-AI and Checklist for Artificial Intelligence in Medical Imaging [,]) would improve comprehension and transparency, allow for more meaningful comparisons between systems, and facilitate meta-analyses.
Real-World Implications and Research Contributions
Overall, despite some limitations, AI is constantly evolving and holds great potential for transformation in the health care sector []. AI has the potential to accelerate existing forms of medical analysis; however, its algorithms require further testing to be fully trusted. Clinically, AI-based automated tools show strong potential to facilitate early detection, precise quantification, progression, and prediction of GA, thereby reducing the burden on retinal specialists and improving diagnostic consistency. Furthermore, DL algorithms have demonstrated effectiveness in identifying retinal image features associated with cognitive decline, dementia, Parkinson disease, and cardiovascular risk factors []. These findings indicate that AI-based retinal images hold promise for transforming primary care and systemic disease management. Although most AI applications remain in the validation phase, the integration of AI with multimodal imaging, novel biomarkers, and emerging therapeutics holds promise for transforming clinical management paradigms in GA and advancing personalized medicine. Future efforts should focus on developing standardized datasets, improving algorithmic generalizability, and conducting real-world validation studies to fully integrate AI into routine ophthalmic practice.
Conclusion
AI, especially DL-based algorithms, holds considerable promise for the detection and management of GA secondary to dry AMD, with performance comparable to trained experts. This systematic review synthesizes and critically appraises the current evidence, highlighting that AI’s capabilities extend across GA management—from initial detection and precise segmentation to the forecasting of lesion progression, which informs future research directions. Meanwhile, with the development of C5 inhibitors, AI-based noninvasive fundus image analysis is expected to detect, identify, and monitor GA at an early stage, thereby increasing the window of opportunity in the future. AI has strong potential to augment and streamline clinical workflows by offering automated, reproducible analysis that can assist clinicians in managing large volumes of imaging data; however, more studies are needed to further validate its effectiveness, repeatability, and accuracy.
The authors declared that artificial intelligence (AI) or AI-assisted technologies were not used in the writing process of this manuscript.
This research was funded by the Central High-Level Traditional Chinese Medicine Hospital Project of the Eye Hospital, China Academy of Chinese Medical Sciences (grant no GSP5-82); the National Natural Science Foundation of China (grant no 82274589); the Science and Technology Innovation Project, China Academy of Chinese Medical Sciences (grant no CI2023C008YG); the Institute-level Research Launch Fund of the Eye Hospital, China Academy of Chinese Medical Sciences (grant no kxy-202402); and the Special Project for the Director of the Business Research Office (grant no 2020YJSZX-2).
All data generated or analyzed during this study are included in this published article and its multimedia appendix files.
None declared.
Edited by Amaryllis Mavragani, Stefano Brini; submitted 26.Jul.2025; peer-reviewed by Jiale Zhang, Xiaolong Liang; final revised version received 11.Oct.2025; accepted 11.Oct.2025; published 21.Nov.2025.
© Nannan Shi, Jiaxian Li, Mengqiu Shang, Weidao Zhang, Kai Xu, Yamin Li, Lina Liang. Originally published in the Journal of Medical Internet Research (https://www.jmir.org), 21.Nov.2025.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in the Journal of Medical Internet Research (ISSN 1438-8871), is properly cited. The complete bibliographic information, a link to the original publication on https://www.jmir.org/, as well as this copyright and license information must be included.
