Open Access
Review  |   June 2025
Current Applications of Artificial Intelligence for Fuchs Endothelial Corneal Dystrophy: A Systematic Review
Author Affiliations & Notes
  • Siyin Liu
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
  • Lynn Kandakji
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
  • Aleksander Stupnicki
    University College London Medical School, London, UK
  • Dayyanah Sumodhee
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
  • Marcello T. Leucci
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0002-3998-6415
  • Scott Hau
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0001-6913-6107
  • Shafi Balal
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0001-7683-2741
  • Arthur Okonkwo
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
  • Ismail Moghul
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0003-3653-2327
  • Sandor P. Kanda
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0009-0006-8694-8653
  • Bruce D. Allan
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0002-8503-4482
  • Dan M. Gore
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
  • Kirithika Muthusamy
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0001-6009-0923
  • Alison J. Hardcastle
    University College London Institute of Ophthalmology, London, UK
    https://orcid.org/0000-0002-0038-6770
  • Alice E. Davidson
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0002-1816-6151
  • Petra Liskova
    Department of Paediatrics and Inherited Metabolic Disorders, First Faculty of Medicine, Charles University and General University Hospital, Prague, Czech Republic
    Department of Ophthalmology, First Faculty of Medicine, Charles University and General University Hospital in Prague, Prague, Czech Republic
    https://orcid.org/0000-0001-7834-8486
  • Nikolas Pontikos
    University College London Institute of Ophthalmology, London, UK
    Moorfields Eye Hospital NHS Foundation Trust, London, UK
    https://orcid.org/0000-0003-1782-4711
  • Correspondence: Nikolas Pontikos, University College London, Institute of Ophthalmology, 11-43 Bath Street, London EC1V 9EL, UK. e-mail: [email protected] 
Translational Vision Science & Technology June 2025, Vol.14, 12. doi:https://doi.org/10.1167/tvst.14.6.12
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Siyin Liu, Lynn Kandakji, Aleksander Stupnicki, Dayyanah Sumodhee, Marcello T. Leucci, Scott Hau, Shafi Balal, Arthur Okonkwo, Ismail Moghul, Sandor P. Kanda, Bruce D. Allan, Dan M. Gore, Kirithika Muthusamy, Alison J. Hardcastle, Alice E. Davidson, Petra Liskova, Nikolas Pontikos; Current Applications of Artificial Intelligence for Fuchs Endothelial Corneal Dystrophy: A Systematic Review. Trans. Vis. Sci. Tech. 2025;14(6):12. https://doi.org/10.1167/tvst.14.6.12.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose: Fuchs endothelial corneal dystrophy (FECD) is a common, age-related cause of visual impairment. This systematic review synthesizes evidence from the literature on artificial intelligence (AI) models developed for the diagnosis and management of FECD.

Methods: We conducted a systematic literature search in MEDLINE, PubMed, Web of Science, and Scopus from January 1, 2000, to June 31, 2024. Full-text studies utilizing AI for various clinical contexts of FECD management were included. Data extraction covered model development, predicted outcomes, validation, and model performance metrics. We graded the included studies using the Quality Assessment of Diagnostic Accuracies Studies 2 tool. This review adheres to the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) recommendations.

Results: Nineteen studies were analyzed. Primary AI algorithms applied in FECD diagnosis and management included neural network architectures specialized for computer vision, utilized on confocal or specular microscopy images, or anterior segment optical coherence tomography images. AI was employed in diverse clinical contexts, such as assessing corneal endothelium and edema and predicting post-corneal transplantation graft detachment and survival. Despite many studies reporting promising model performance, a notable limitation was that only three studies performed external validation. Bias introduced by patient selection processes and experimental designs was evident in the included studies.

Conclusions: Despite the potential of AI algorithms to enhance FECD diagnosis and prognostication, further work is required to evaluate their real-world applicability and clinical utility.

Translational Relevance: This review offers critical insights for researchers, clinicians, and policymakers, aiding their understanding of existing AI research in FECD management and guiding future health service strategies.

Introduction
Fuchs endothelial corneal dystrophy (FECD) is a common, age-related cause of visual dysfunction characterized by the degeneration of the corneal endothelium, a critical monolayer of cells on the inner surface of the cornea.1,2 The prevalence of FECD varies widely among different ethnic groups, as it is estimated to affect 9.2% of White individuals,3 6.7% of Chinese Singaporeans,4 and 4.1% of Japanese.5 A recent meta-analysis estimated a global prevalence of 7.33%.6 Corneal endothelial cell density declines prematurely and progressively with age in FECD, eventually leading to corneal edema, resulting in loss of corneal transparency, glare, and impaired vision.7 Corneal transplantation becomes the sole definitive treatment option at this stage, with FECD being the primary indication for posterior lamellar keratoplasty in the United Kingdom.8 
Historically, grading the severity of FECD involved quantifying guttae and detecting corneal edema through slit-lamp biomicroscopic examination.9 However, this approach has limited clinical utility due to significant interobserver variability in counting and describing guttae. Although end-stage corneal edema is evident on clinical examination, subclinical edema, by definition, is undetectable to the naked eye. Corneal densitometry and specific tomographic features derived from Scheimpflug camera tomography have been suggested as alternative methods for detecting and quantifying corneal edema.10,11 Anterior segment optical coherence tomography (AS-OCT) offers high-resolution cross-sectional scans of the cornea, but subtle localized swelling might be challenging to detect by humans from these images.12 Current criteria for diagnosing and classifying the severity of FECD have been summarized recently.13 
The rapid advances in artificial intelligence (AI) have sparked transformative changes across various medical fields, including ophthalmology. The ophthalmic community is uniquely positioned to harness AI strategies due to the widespread use of imaging tools in clinical practice. Over the past decade, AI has made significant strides in screening and detection of retinal diseases such as diabetic retinopathy, age-related macular degeneration, and inherited retinal diseases.1417 This progress has been made possible by the availability of large and diverse labeled datasets. More recently, promising results have also been reported in anterior segment conditions—namely, keratoconus18 and refractive surgery screening.19 Despite the prevalence of FECD, application of AI in this area has been relatively limited. 
This systematic review aimed to synthesize and evaluate the quality of reporting, risk of bias, the different types of AI algorithms, and their performance for FECD. Our goal was to accurately portray the current state of development, identifying the opportunities and challenges to implementing AI in routine clinical practice. 
Methods
Search Strategy
We conducted a literature review of AI applied to the diagnosis and management of FECD published between January 1, 2000, and August 31, 2023. The Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) Statement 2020 criteria20 were followed by searching bibliographic databases (PubMed, MEDLINE, EMBASE, Web of Science, Scopus, Cochrane Library, IEEE Explorer, and Digital Library) using keyword searches on their title, abstract, and keywords. It has been registered a priori in the International Prospective Register of Systematic Reviews (PROSPERO ID: CRD42023454518). The search terms included a combination of the following terms: (“Fuchs endothelial corneal dystrophy” OR “Fuchs dystrophy” OR “endothelial dystrophy” OR “guttata”) AND (“algorithm” OR “artificial intelligence” OR “machine learning” OR “deep learning”) AND (“detection” OR “diagnosis” OR “grading” OR “classification” OR “segmentation”). In each database, the same search terms were used, and a filter was set to include human studies and exclude animal studies. Relevant studies were also identified via a search in Google Scholar or from the bibliographies of the included studies. 
Eligibility Criteria
Studies were included if they investigated the application of AI algorithms for the detection of FECD, classification of FECD disease severity, identifying corneal edema, assessment of the corneal endothelium in FECD patients, diagnosis of graft detachment, and prediction of post-corneal transplantation outcomes. The studies should have reported the performance of the AI model using metrics appropriate for the task, such as sensitivity and specificity for classification or the Sørensen–Dice coefficient for segmentation. The performance metric should have been derived from a validation or test set separate from the training dataset. Finally, the full-text article must be available, and only papers published in English were considered. 
Studies were excluded if they were not conducted on humans, were focused on guttata and/or corneal edema caused by factors other than FECD, or involved eyes complicated by other ophthalmic conditions. Additionally, this review excluded articles classified as case reports, reviews, comments, qualitative studies, or protocols. 
Data Synthesis
The PRISMA flowchart (Fig. 1) provides an overview of the study search and selection process. Initially, search results were extracted in a comma-separated value (CSV) format from the searched databases and imported into Rayyan, an established AI-powered tool that allows for a rigorous and semi-automated screening. A Rayyan inbuilt algorithm was then used to remove duplicate records, followed by a manual inspection and correction of missed duplicates. Subsequently, three reviewers (SL, LK, AS) manually screened the abstracts within Rayyan based on the inclusion and exclusion criteria. Disagreements in meeting the inclusion or exclusion criteria were resolved by discussion. Following abstract screening, full-text articles were carefully reviewed in the original source, with reasons for exclusion systematically documented (provided in Supplementary Table S1). 
Figure 1.
 
PRISMA flow diagram outlining the study selection process for the systematic review. This diagram illustrates the steps taken to include and exclude studies based on predefined eligibility criteria, highlighting the number of studies at each stage of the review process, including identification, screening, eligibility, and inclusion.
Figure 1.
 
PRISMA flow diagram outlining the study selection process for the systematic review. This diagram illustrates the steps taken to include and exclude studies based on predefined eligibility criteria, highlighting the number of studies at each stage of the review process, including identification, screening, eligibility, and inclusion.
Following this, data were extracted from studies that met the eligibility criteria: author and year, title, country, age, gender, number of eyes for each group, clinical utility, validation details, input data, algorithm, classification groups, performance metrics appropriate for the task (e.g., sensitivity, specificity, accuracy, precision, area under the receiver operating characteristic curve [AUC], Dice coefficient). Due to the heterogeneity in the described algorithms, study methodologies, and reported metrics, meta-analysis was considered inappropriate and was not performed. 
Bias Assessment
When evaluating bias across the included studies, we employed Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2),21 a tool designed to critically appraise the methodological quality and potential bias in studies that evaluate the diagnostic accuracy of medical tests. It assesses four key domains: patient selection, index test, reference standard, and flow/timing. Three independent reviewers (SL, LK, AS) conducted the assessments, and each study was reviewed by at least two reviewers for a reliable assessment. A meta-analysis of the adherence to QUADAS-2, measured by the percentage of “Yes” responses, was performed using the metafor package in R (R Foundation for Statistical Computing, Vienna, Austria). 
Results
Literature Search
The study selection process is outlined in Figure 1. The literature search initially identified 95 studies. Following the removal of 20 duplicate records, 25 studies were excluded based on relevance, including the exclusion of animal studies and those that did not investigate clinical applications to FECD, leaving 50 studies for full-text review. A detailed breakdown of the reasons for including or excluding each article is provided in Supplementary Table S1
Ultimately, 19 articles were included in our systematic review. The Table outlines the datasets and the developed AI pipelines used in the individual studies. Figure 2 presents a comparative heatmap of the evaluation metrics reported across studies, highlighting the variability in methodological choices and providing an overview of their performance based on accuracy, Dice coefficient, sensitivity, and specificity. 
Table.
 
Summary of the Published Studies That Included the Use of AI for the Assessment, Diagnosis, and Prognostication of FECD
Table.
 
Summary of the Published Studies That Included the Use of AI for the Assessment, Diagnosis, and Prognostication of FECD
Figure 2.
 
Heatmap of performance metrics across studies. Rows represent individual studies, and columns indicate accuracy, Dice coefficient, sensitivity, and specificity. Color intensity reflects performance, with darker shades indicating higher values. Metrics are categorized as ≤0.50 (lightest blue), 0.50 to 0.70 (light blue), 0.70 to 0.85 (medium blue), and ≥0.85 (darkest blue). Gray cells indicate that the corresponding study did not report this metric. High-performing models and variability in evaluation criteria can be readily observed in this figure. Comprehensive details on datasets, AI pipelines, and additional evaluation metrics are provided in the Table, which contextualizes these results within the broader systematic review.
Figure 2.
 
Heatmap of performance metrics across studies. Rows represent individual studies, and columns indicate accuracy, Dice coefficient, sensitivity, and specificity. Color intensity reflects performance, with darker shades indicating higher values. Metrics are categorized as ≤0.50 (lightest blue), 0.50 to 0.70 (light blue), 0.70 to 0.85 (medium blue), and ≥0.85 (darkest blue). Gray cells indicate that the corresponding study did not report this metric. High-performing models and variability in evaluation criteria can be readily observed in this figure. Comprehensive details on datasets, AI pipelines, and additional evaluation metrics are provided in the Table, which contextualizes these results within the broader systematic review.
Input Datasets for Model Development
The datasets utilized in these studies had a median sample size of 925 images (range, 30–11,340), with a median ratio of 1.09 FECD cases per control (range, 0.05–3.14). Primarily, algorithms were trained using single-center data (n = 15, 78.9%); one used a multicenter dataset (5.3%)29 and one study used a national registry dataset (5.3%).39 In three studies, the source of patients could not be determined.26,27,38 None of the studies utilized public datasets for model development. 
Imaging Modalities
Except for the one study utilizing national registry–derived clinical datasets for training algorithms predicting graft detachment,39 most of the models included in the review were trained using raw corneal images. Predominantly, cross-sectional images of the cornea were acquired via AS-OCT, which served as the primary imaging modality for model development. These images were generated from various devices, including Envisu (Bioptigen, Seattle, WA), SPECTRALIS (Heidelberg Engineering, Heidelberg, Germany), CASIA (Tomey Corporation, Aichi, Japan), and Avanti OCT (Optovue, Fremont, CA). These systems can produce numerical indices measuring corneal metrics such as thickness, keratometry, elevation, and aberrations, but the included studies train models using the raw images acquired instead of the processed indices. In addition, seven models focused on evaluating corneal endothelium, with seven employing images captured via specular microscopy with the CellCheck XL (Konan Medical, Hyogo, Japan), SP-1P (Topcon, Tokyo, Japan), and EM-3000 (Tomey Corporation), and one of those using confocal microscopy (HRT3 Rostock Cornea Module; Heidelberg Engineering). 
Algorithm Validation
Most of the studies included did not incorporate external validation (16/19); instead, they relied on internal validation methods such as k-fold cross-validation (n = 7) or random partitioning of the original dataset into training and validation sets (n = 9). Among these, one study by Hayashi et al.35 stands out for training the model using patient data from one surgeon to predict post-transplant graft detachment requiring rebubbling, then validating it with data from another surgeon. The authors reported an AUC of 87.5% when applying the trained algorithm to the second surgeon's dataset. However, because the performance of the model on the original dataset used for training and validating the algorithm was not provided, comparisons to assess overfitting are not feasible. 
To evaluate and augment a prior model developed by the same group31 in a preoperative setting, Bitton et al.32 trained the algorithm using AS-OCT images derived from 50 normal corneas and 240 edematous corneas of pre-Descemet membrane endothelial keratoplasty (DMEK) patients. These patients had indications for DMEK due to either FECD (144 eyes) or pseudophakic bullous keratopathy (96 eyes). The reported optimal edema fraction (EF) threshold, representing the ratio between pixels labeled as corneal edema and those representing normal cornea, was 0.143. At this threshold, the model had a sensitivity of 94.5%, specificity of 92%, and AUC of 96%, roughly in line with the performance of their internal validation using the original dataset. Although the model demonstrated promising performance, in some control cases a significant proportion of normal corneas were labeled as edematous. This occurrence may reflect global signal differences unrelated to the presence or absence of corneal edema. Caution is warranted, as convolutional neural network (CNN) models can be highly sensitive to subtle signal differences that might not be discernible to the human eye.41 
Recent studies have increasingly utilized external validation. For example, Qu et al.29 used a multicenter dataset with images from seven hospitals, and Foo et al.28 employed an unspecified independent external validation set. Both studies observed a decline in performance, with the algorithm of Qu et al.29 showing a more pronounced drop. This performance reduction underscores a generalization gap and highlights the limitations of clinical applicability in real-world settings. Retrospective training datasets often undergo extensive filtering and cleaning, which may not accurately represent the variability of real-world data, leading to overfitting and diminished performance on external datasets. 
Clinical Context, Algorithms, and Model Performance
The primary AI algorithms applied in FECD diagnosis and management encompass neural network architectures specialized for computer vision (i.e., image classification or segmentation tasks). Muijzer et al.39 pursued regression and decision tree–based approaches for identifying risk factors for graft detachment. This subsection is structured according to the clinical contexts where these AI algorithms were applied. Figure 3 is a summary diagram delineating the pertinent models utilized across different clinical scenarios. 
Figure 3.
 
Summary diagram presenting the various AI approaches used in the management of FECD. The diagram categorizes the AI algorithms employed in different clinical contexts, including the assessment of corneal endothelium, evaluation of corneal edema, and prognostication of post-corneal transplantation graft detachment and rejection. This overview highlights the application of AI methods within clinical workflow for managing FECD.
Figure 3.
 
Summary diagram presenting the various AI approaches used in the management of FECD. The diagram categorizes the AI algorithms employed in different clinical contexts, including the assessment of corneal endothelium, evaluation of corneal edema, and prognostication of post-corneal transplantation graft detachment and rejection. This overview highlights the application of AI methods within clinical workflow for managing FECD.
Assessment of Corneal Endothelium
The endothelium is critical in evaluating FECD due to specific morphological patterns that can help diagnose and stage disease. Imaging of the endothelium can be achieved using various technologies such as specular microscopy and in vivo confocal microscopy (IVCM) (Fig. 4).42 These technologies allow for semi-manual or automated quantification of endothelial cell density (ECD) and morphology. Automated corneal endothelial cell segmentation has long relied on marker-driven watershed segmentation, but this method is prone to both under- and over-segmentation, especially in areas with low signal-to-noise ratios.43,44 AI approaches have been applied to segment corneal endothelial cells more accurately.45 AI brings a significant advantage by adapting to variability in cell morphology and image quality, providing more precise segmentation where traditional methods fall short. Additionally, it can continuously improve through learning, allowing it to handle complex patterns and subtle changes that can be otherwise missed. 
Figure 4.
 
Slit-scanning in vivo confocal microscopy images of the corneal endothelium. (A) Healthy control cornea. Image of a normal corneal endothelium shows a uniform hexagonal cell pattern with consistent cell density and no visible guttae or abnormalities (white solid arrowhead). (B) Early-stage FECD. Image shows isolated guttae (white dashed arrowhead), characterized by dark, round bodies with central white hyperreflectivity, indicating early endothelial cell distress. (C) Late-stage FECD. Increased confluency of guttae (white dashed arrowhead), illustrating a more extensive presence of these structures (dashed-line highlighted area) with evidence of endothelial cell loss.
Figure 4.
 
Slit-scanning in vivo confocal microscopy images of the corneal endothelium. (A) Healthy control cornea. Image of a normal corneal endothelium shows a uniform hexagonal cell pattern with consistent cell density and no visible guttae or abnormalities (white solid arrowhead). (B) Early-stage FECD. Image shows isolated guttae (white dashed arrowhead), characterized by dark, round bodies with central white hyperreflectivity, indicating early endothelial cell distress. (C) Late-stage FECD. Increased confluency of guttae (white dashed arrowhead), illustrating a more extensive presence of these structures (dashed-line highlighted area) with evidence of endothelial cell loss.
Several studies have employed U-Net–based architectures for segmenting corneal endothelial cells from specular microscopy images. Shilpashree et al.25 used a modified U-Net with Watershed postprocessing to resolve merged cell borders in images of FECD patients and healthy subjects. Their model yielded 87.90% accuracy in correctly segmenting the percentage of pixels in the image, an F1 score of 82.27%, and an AUC of 96.70% for recognizing the cell borders. Sierra et al.26 approached the problem as a regression task of cell and gutta signed distance maps instead of pixel-level classification, which converged faster and yielded 83.79% accuracy in guttata identification. These studies demonstrate the effectiveness of U-Net variants in analyzing corneal endothelial images in FECD. 
Vigueras-Guillén et al.23 proposed a novel deep-learning method using Dense U-Net for segmenting specular microscopy images of the corneal endothelium with guttae, incorporating an attention mechanism referred to as feedback non-local attention (fNLA) to infer the cell edges in occluded areas. This binary segmentation enabled the estimation of ECD, coefficient of variation (CV), and hexagonality (HEX). Compared to a manually segmented ground truth of 1203 images, they found that Dense U-Net with fNLA produced the lowest error, with a mean absolute error of 23.16 cells/mm2 in ECD, 1.28% in CV, and 3.13% in HEX, which was 3 to 6 times smaller than the manufacturer's (Topcon) built-in software. Vigueras-Guillén et al.23 also utilized Dense U-Net to estimate corneal endothelium parameters from specular microscopy images acquired with a Topcon SP-1P Microscope of eyes 1, 3, 6, and 12 months after ultrathin Descemet stripping automated endothelial keratoplasty (UT-DSAEK).24 The proposed method outperformed the Topcon software, with a mean absolute error of 30.00 cells/mm2 (vs. 118.70 cells/mm2) for estimating ECD, 1.5% (vs. 4.6%) for CV, and 3.0% (vs. 10.3%) for HEX. Notably, this study reported the highest Dice coefficient among all segmentation tasks in the context of endothelium assessment. 
Karmakar et al.22 proposed a novel deep learning–based cell segmentation algorithm, Mobile-CellNet, to estimate ECD in specular microscopy images. The proposed algorithm uses two similar CNN-based image segmentation models working in parallel along with image postprocessing using classical image processing techniques to delineate the boundaries of the endothelial cells. When compared to the widely used U-Net architecture, Mobile-CellNet resulted in a mean absolute error of 4.06% for ECD on the test set, offering similar performance but greater computational efficiency. 
Foo et al.28 developed two classification algorithms using DenseNet-121 to detect FECD and to analyze corneal endothelium in specular microscopy images, respectively. The first model achieved an AUC of 0.96, 91% sensitivity, and 91% specificity on an internal dataset, but these metrics dropped to 0.77, 69%, and 68%, respectively, on an external dataset, performing worse than expert manual grading. Their second model identified widefield specular microscopy images with ECD > 1000 cells/mm2 in FECD eyes, achieving an AUC of 0.88, 79% sensitivity, and 78% specificity. 
The Heidelberg Rostock Cornea Module, commonly used for IVCM in clinical settings, lacks automated tools for assessment of corneal endothelium, prompting Qu et al.27 to develop a fully automated segmentation and morphometric parameter estimation system for assessing central corneal endothelial cells from IVCM images. The automated system calculated various morphometric parameters, including ECD, CV in cell area, and percentage of HEX. The automated ECDs were highly correlated with ECD measurements from the Topcon specular microscope (r = 0.932), which also correlated highly with ground-truth ECDs derived from manual calculation (r = 0.818). Given the ability of IVCM to provide cell-level visualization of corneal structure and superior optical penetration even in the presence of edema, an automated morphometric system for analyzing IVCM images could, as a proof of concept, offer a more effective diagnostic measure than specular microscopy. 
Qu et al.29 also developed an open-access system for recognizing IVCM images of various corneal endothelial diseases, including FECD, using the novel Enhanced Compact Convolutional Transformer (ECCT), which combines CNNs and transformers. When validated on a multicenter testing dataset from seven Chinese hospitals, the model achieved a sensitivity of 98.3%, specificity of 94.8%, and accuracy of 97% in recognizing FECD images, demonstrating comparable or superior performance to other robust AI models, including ResNet-34, EfficientNet-B5, DeiT-S, and Swin-T. 
Assessment of Corneal Edema
The presence of corneal edema (Fig. 5) signifies a crucial stage in the progression of FECD, as it indicates a decline in ECD below the threshold required for maintaining corneal dehydration (referred to as stromal deturgescence). Although corneal thickness is widely used in clinical settings to measure corneal edema, it can be misleading in naturally thin or thick corneas.46,47 Thus far, three studies have developed algorithms for assessing corneal edema, of which one aimed to classify the disease into early and late stages,30 and the other two quantified the severity of edema.32,48 
Figure 5.
 
Cross-sectional AS-OCT images of the cornea in FECD. (A) Early-stage FECD. AS-OCT image of the cornea shows no clinical signs of corneal edema. The corneal thickness remains within the normal range, and no significant epithelial or stromal changes are visible. (B) Late-stage FECD. AS-OCT image shows increased central corneal thickness, epithelial bullae (fluid-filled cystic spaces), and increased pixel intensity (backscatter) in the stroma, indicating corneal decompensation and edema.
Figure 5.
 
Cross-sectional AS-OCT images of the cornea in FECD. (A) Early-stage FECD. AS-OCT image of the cornea shows no clinical signs of corneal edema. The corneal thickness remains within the normal range, and no significant epithelial or stromal changes are visible. (B) Late-stage FECD. AS-OCT image shows increased central corneal thickness, epithelial bullae (fluid-filled cystic spaces), and increased pixel intensity (backscatter) in the stroma, indicating corneal decompensation and edema.
CNNs have demonstrated the ability to make inferences from imaging data structures through deep learning.49 Eleiwa et al.30 developed a deep learning model based on the Visual Geometry Group with 19 layers (VGG19) to classify AS-OCT images into normal, early-stage FECD (defined as without clinical edema), and late-stage FECD (presence of clinically evident edema), with sensitivity of 91%, 100%, and 99%; specificity of 97%, 92%, and 98%; and AUCs of 0.997, 0.974, and 0.998, respectively. Although this approach demonstrated the best performance among classification models for corneal edema assessment, it did not provide information on the precise location of edema on the images, which may limit its clinical application in detecting subclinical or focal edema. 
Zéboulon et al.48 developed a U-Net–based50 segmentation model to identify corneal edema and derive the EF from AS-OCT images. Trained on 111 scans of post-DMEK corneas with total corneal edema and 88 normal corneal scans, the model had Dice coefficients of 0.990 and 0.967 for normal and edema pixel predictions, respectively. With an optimal EF threshold of 0.068, the model had a sensitivity of 96.4%, a specificity of 100%, and an AUC of 0.994 for distinguishing edematous from normal corneas.48 Although this pilot study confirmed the potential of using a CNN model to detect corneal edema, the concurrent evaluation of the epithelium and stroma hindered its ability to detect mild stromal edema, which can be masked by a normal epithelium and lead to false negatives. The same research group later enhanced the model by incorporating additional layers to delineate the epithelium and stroma in the image analysis pipeline, enabling the new model to detect significant EF differences between normal corneas and those with mild edema,31 a capability not achieved by the original model.48 
Predicting Post-Corneal Transplantation Graft Detachment
Graft detachment, a complications specific to DMEK51 (Fig. 6), has reported rates ranging from 4% to 34.6%,51,52 with 0.2% to 75% of cases requiring postoperative rebubbling.53 AS-OCT is a superior modality for detecting subtle and shallow graft detachment compared to slit-lamp examination alone.54 Seven studies investigated the potential of utilizing AI algorithms to recognize,3336 quantify,37,38 and predict39 post-corneal transplantation graft detachment. 
Figure 6.
 
(A, B) Cross-sectional AS-OCT images of the cornea showing a small partial graft detachment after DMEK (A) and a more extensive graft detachment (B).
Figure 6.
 
(A, B) Cross-sectional AS-OCT images of the cornea showing a small partial graft detachment after DMEK (A) and a more extensive graft detachment (B).
Treder et al.33 utilized post-DMEK AS-OCT images to develop a CNN classifier based on a pretrained algorithm for distinguishing corneas with attached grafts from those with graft detachment. The input dataset was comprised of 1172 AS-OCT images (609 attached graft; 563 detached graft) from 111 eyes of 91 patients within the first 90-day postoperative period. The classifier achieved 98% sensitivity, 94% specificity, and 96% accuracy.33 This model demonstrated the highest sensitivity among classification models for corneal graft attachment assessment. However, irregular graft curvature, image artifacts, and flat or peripheral graft detachment were identified as potential sources of misclassification.33 
Rather than simply classifying graft detachment, Hayashi et al.34 compared the performance of multiple deep-learning classification models (VGG16, VGG19, ResNet50, InceptionV3, InceptionResNetV2, Xception, DenseNet121, DenseNet169, and DenseNet201) to discriminate patients requiring rebubbling after DMEK from those who did not. Training these models with a dataset of 496 images from 31 eyes requiring rebubbling and 496 images from 31 eyes not requiring rebubbling, all operated on by one surgeon, the VGG19 model exhibited the highest AUC of 0.964. The sensitivity and specificity of this model were reported as 0.967 and 0.915, respectively.34 Subsequently, the same research group developed the EfficientNet-based algorithm, trained on OCT images from patients operated on by one surgeon, which performed well in predicting the need for rebubbling in surgeries by another surgeon, with AUC of 0.875, Youden index of 0.214, and sensitivity and specificity of 78.9% and 78.6% respectively.35 
The previously described graft detachment classification models are primarily based on AS-OCT scans obtained after DMEK surgery. However, an algorithm capable of predicting the risk of graft detachments before surgery could significantly optimize surgical care by serving as a preoperative screening tool. Patefield et al.36 developed a novel multiple-instance learning artificial intelligence (MIL-AI) model55 using ResNet and preoperative AS-OCT images to distinguish between eyes with and without graft detachment after DMEK, with a sensitivity of 92%, a specificity of 45%, and an AUC of 0.63, outperforming the human labeling by ophthalmologists in sensitivity but demonstrating lower specificity. 
In contrast to binary classification approaches focusing on the presence or absence of graft detachment, deep learning–based segmentation techniques can quantify the length and location of the detachment. Heslinga et al.37 developed an automated method using AS-OCT images to locate and quantify graft detachment after DMEK. A ResNet-based56 model localized the scleral spur for anterior chamber measurements, followed by a U-Net architecture50 for semantic segmentation to identify image pixels corresponding to graft detachment. Trained and validated on 1280 AS-OCT images, of which 336 did not have graft detachment from 68 eyes post-DMEK, the model accurately estimated graft detachment lengths in 69% of cases, with a Dice score of 0.896.37 However, the performance of the model in identifying large central detachments was limited, likely due to insufficient training samples and confusion with intraocular gas. 
More recently, Glatz et al.38 developed a U-Net–based model to generate percentage detachment maps and three-dimensional volume representations of graft detachment after DMEK. Trained on 6912 manually labeled AS-OCT images from 27 eyes with FECD (including 14 eyes without graft detachment) after a median of 16 days post-DMEK, the model reported a Youden index of 0.850, a Dice coefficient of 0.729, a sensitivity of 0.854, and a specificity of 0.996.38 In an application set of 107 eyes, the model outperformed slit-lamp examination by detecting detachments missed or underestimated by human specialists. 
Rather than using an image-based computer vision system, Muijzer et al.39 leveraged data from the Netherlands Organ Transplant Registry and employed three machine learning models, namely L1 regularized logistic regression with least absolute shrinkage and selection operator (LASSO), classification tree algorithm (CTA), and random forest classification (RFC), to identify risk factors for graft detachment after posterior lamellar keratoplasty. Key risk factors identified included undergoing a DMEK procedure (rather than DSAEK), prior graft failure, and using sulfur hexafluoride gas during surgery. The RFC model performed best with an AUC of 0.72, sensitivity of 0.68, and specificity of 0.62, compared to LASSO (AUC, 0.70; sensitivity, 0.70; specificity, 0.65) and CTA (AUC, 0.65; sensitivity, 0.65; specificity, 0.62).39 
Predicting Graft Rejection and Survival
The assessment of ECD has been proposed as a predictor of graft rejection and survival.5759 Joseph et al.40 developed a deep learning pipeline for identifying eyes at risk of graft rejection within 1 to 24 months after DSAEK. The authors used a pretrained U-net60 model to segment endothelial cells and extract novel quantitative features relevant to cellular spatial arrangements and cell intensity values. These features were used to train random forest and logistic regression models, with over 0.80 accuracy in predicting post-DMEK graft rejection. Key predictors for graft rejection included cell-graph spatial arrangement, intensity, and shape features. However, research in this area is limited partly because many endothelial rejection episodes are asymptomatic. 
Study Quality
The risk of bias was assessed using the QUADAS-2 tool.21 A meta-analysis of the adherence rates (percentage of QUADAS-2 criteria answered “Yes”) across included studies yielded a pooled adherence rate of 55.1% (95% confidence interval [CI], 47.7–62.2), and heterogeneity analysis showed low variability between studies (Cochran's Q = 22.10; P = 0.2275; I2 = 18.6%). In general, patient selection was found to have a high risk of bias in over half of the included studies (11/19, 55%) because the majority of these studies were case–control studies and did not use consecutive or random samples. A high risk of bias was introduced by the index test in 26% of cases (5/19), mainly due to the result being interpreted with knowledge of the results of the reference standards. However, a significant proportion of the articles did not specify this and were marked as “Unclear.” Reference standards for many classification and segmentation models consisted of manual interpretation and annotation by human experts. Although no gold standards exist for these tasks, manual curation for reference results by trained specialists was considered a low risk of introducing bias. The individual assessment for each study is present in Supplementary Table S2
Discussion
Our systematic review identified 19 peer-reviewed manuscripts that presented AI algorithms focused on different stages of FECD management, using various imaging techniques and clinical data. Notably, this review is the first systematic examination of AI applications in assessing FECD. Predominant areas of investigation involve predicting graft detachments after posterior corneal transplantation and assessing corneal endothelium, followed by analyzing corneal edema and predicting graft survival. 
In the current literature, the predominant input data are raw pixel-level corneal imaging obtained from either AS-OCT or specular microscopy. Conversely, developing models for the detection of keratoconus, another prevalent cornea disease, often relies on computed parameters such as keratometry and elevation data rather than direct corneal images.18,61,62 This difference in approach arises because FECD assessment heavily hinges on clinical examination, lacking validated numerical indices to differentiate disease stages, unlike keratoconus. Moreover, the input imaging data were produced by devices of different manufacturers. Despite sharing the same underlying technology and appearing similarly to human eyes, whether computer vision models interpret these images uniformly at a pixel level across different systems remains unclear. Consequently, comparison or replication across systems poses challenges. Although several articles reported satisfactory performance by deep learning models, validation studies assessing model performance using images from different devices than those used in training are lacking. 
Interestingly, despite the Scheimpflug imaging system having a considerable clinical track record in FECD,63 our review did not uncover any studies attempting to train AI using Scheimpflug image data for FECD management. This is likely due to proprietary closed-source data formats, such as for the Pentacam (OCULUS, Wetzlar, Germany), which limit bulk export for AI training. Moreover, the newer AS-OCT technology produces higher-resolution cross-sectional images, enabling the differentiation of epithelial and stromal edema—a potentially appealing feature for training deep learning models. Due to the wavelengths used in AS-OCT, it generates clearer images for FECD patients with severe edema compared to Scheimpflug tomography, which is prone to interference by light scatter. These factors contribute to the preference for training computer vision models using images from AS-OCT over Scheimpflug. 
Successful AI model training relies on the availability of ground truth data as a reference standard and a sufficiently large sample size to reveal statistically meaningful patterns. Identifying a detached graft on AS-OCT and delineating endothelial cell edges on specular and confocal microscopy are relatively easy tasks for human eyes, enabling consistent manual labeling by trained experts as ground truth for model training. However, objective detection of corneal edema remains challenging. Corneal thickness is the most widely used surrogate parameter for assessing corneal edema in the clinic, although its reliability is limited due to natural variation in normal corneal thickness.64 Manual labeling of corneal edema on AS-OCT images presents greater challenges than detecting retinal pathologies such as macular edema, where features such as intraretinal/subretinal fluid are well defined as “intraretinal hyporeflective space surrounded by reflective septae” with clear borders, readily recognized by both machines and humans without clinical expertise.65 In contrast, detecting subclinical corneal edema is challenging, both on slit-lamp examination and on quasi-histological OCT images, as early edema in OCT often appears as subtle changes in reflectivity, requiring extensive subspecialist expertise for accurate delineation. Prognosticating post-corneal transplant graft survival poses equal challenges. Posterior lamellar transplant procedures generally show favorable outcomes. Using AI to uncover the subset of patients at risk of graft failure requires extensive long-term follow-up datasets, which are often unavailable. The lack of a universally accepted reference standard and dataset impede model training in these clinical scenarios. 
AI research concerning FECD management has predominantly concentrated on supervised learning approaches such as decision trees, regression analysis, and CNNs. Reported performance metrics across these studies exhibit variability, depending on the input data types, AI methodologies, and clinical contexts, making it challenging to compare the performance of different AI algorithms. For classification studies, sensitivity, specificity, and receiver operating characteristic analyses are commonly employed to gauge diagnostic capability. The vast majority of included studies (n = 14), however, did not report other metrics crucial for evaluation and systematic comparison of model performance, such as the F1 score or recall. Those are especially key in studies where the training dataset was unbalanced, which is often the case in reports for diseases of low prevalence such as FECD. Although many classification algorithms exhibited an AUC exceeding 90% when trained and validated on a single dataset, their performance significantly declined when tested on different input datasets,35 which potentially limits their clinical application. Dice coefficients have been used to evaluate the alignment between predicted and actual segmentation masks, revealing more variability in segmentation model performance even with data from the same imaging modality. This variability likely stems from the complexity of segmentation tasks, which require precise pixel-level accuracy and detailed boundary delineation, whereas classification tasks involve a broader understanding of overall image features for label assignment. 
Validating AI models on datasets distinct from the trained set is crucial to assess the generalizability of the model. Although all of the included studies conducted internal validation, utilizing methods such as k-fold cross-validation or random data splitting into training/test sets, only three studies performed external validation.28,29,35 CNNs are susceptible to overfitting if not appropriately regularized, leading to drastic accuracy changes when used with data different from the trained dataset.66 Interestingly, Bitton et al.32 attempted to validate a prior deep learning pipeline using an independent out-of-sample dataset and reported comparable satisfactory performance in line with the original result.31 This pipeline initially employed a computer vision approach to quantify corneal EF from AS-OCT images and then utilized this value to classify FECD severity. Despite the encouraging classification performance, it became evident that the same EF was not universally applicable, requiring recalibration for different datasets. Studies in other fields have demonstrated that even models exhibiting adequate performance on internal validation may exhibit significant decreases in sensitivity and specificity during external validation,67 and this decline in performance was evident in the algorithm by Foo et al.28 Beyond a computer vision model for the anterior segment, external validation of AI models in ophthalmology presents several challenges. Data heterogeneity remains a key concern, as models trained on homogeneous populations often underperform on diverse ethnicities or imaging devices.68 Image variability, including field width, quality, and magnification, also significantly impacts performance across datasets.69 Standardization is another issue, given inconsistencies in data collection, imaging protocols, and annotation methods.70 Ideally, addressing these issues requires diverse training datasets and rigorous validation on larger external datasets that represent the general population.69 
Although computer vision models demonstrate promising performance in various clinical contexts of FECD, their applicability is not always evident across all scenarios. Clinical recognition of advanced FECD with endothelial decompensation is relatively straightforward, graft detachment is visibly discernible on AS-OCT, and assessment of endothelial cells is efficiently performed using manufacture-built software in specular microscopy devices. The development of AI in these areas may offer limited additional utility. Further, it is crucial to consider real-world clinical practices when evaluating AI applications, including the choice of clinical pathway and imaging modalities. For example, a highly accurate model for diagnosing multiple corneal endothelial diseases may be valuable,29 but if it relies on a less commonly available and time-consuming imaging modality such as IVCM, its practical utility may be constrained. Therefore, research endeavors should address existing clinical care gaps, such as evaluating and quantifying subclinical corneal edema, predicting disease progression, and prognosticating post-corneal transplantation outcomes. Focusing on these aspects is likely to maximize the potential benefits of AI by complementing human limitations, enabling timely interventions to mitigate disease progression and visual loss. 
Most of the included studies utilized retrospective data for model development and did not involve masked observers for model performance evaluation, potentially introducing detection bias. The reported studies generally had small sample sizes, with only two studies encompassing more than 500 eyes (including controls), and none of the studies conducted a priori power calculations to estimate the required cohort size. Given that sample size significantly impacts AI algorithm development and the accuracy of the model, further studies with larger sample sizes are warranted. 
This systematic review has several limitations. Articles that lacked relevant key terms or were presented solely as abstracts might have been overlooked; however, this rigorous process ensured the inclusion of high-quality articles. Furthermore, the developed pipelines were not always directly comparable. As we intended to be as comprehensive as possible, this review encompassed studies with diverse study designs, sample sizes, input dataset sources, case definitions, imaging modalities, and validation approaches. The heterogeneity among these models posed challenges for meaningful comparisons. Consequently, statistical synthesis and meta-analyses were unfeasible. Additionally, we excluded studies lacking any form of validation and those not conducted on humans, typically proof-of-concept articles outlining experimental steps in model development. This exclusion inevitably resulted in information loss; nevertheless, our intention was to summarize only the approaches with the highest potential for clinical applicability and generalizability. 
Considering ongoing challenges in model generalizability, future investigations could leverage transfer learning, wherein large-scale pretrained architectures (derived from extensive ophthalmic or general medical imaging corpora) are fine tuned using smaller FECD-specific datasets.71 This strategy mitigates data scarcity and enhances overall performance stability. Federated learning frameworks further bolster external validity by facilitating collaborative model development across multiple centers without the need for centralized data pooling, preserving patient privacy.72 Beyond conventional methods, self-supervised and semi-supervised methods can exploit vast repositories of unlabeled corneal images to learn robust feature representations, later refined for FECD-specific applications. Likewise, transformer-based architectures (e.g., Vision Transformers) merit attention for their capacity to capture long-range dependencies within high-resolution corneal scans.73 To overcome the pervasive challenge of limited or imbalanced training data, synthetic data generation through generative adversarial networks74 or diffusion models can expand datasets and facilitate domain adaptation across disparate imaging devices and clinical environments. Finally, integrating explainable AI (XAI) such as attention-based visualization or activation mapping can elucidate the salient image features that drive model outputs, enhancing clinician trust and interpretability.75 Coupled with real-time AI-assisted image acquisition, which can identify optimal imagine angles or highlight regions of subtle pathology, these advancements may streamline corneal imaging workflows and ensure higher-fidelity data for downstream analytical tasks. 
Although AI cannot replace human expertise or existing diagnostic tools, after the necessary regulatory approvals its integration as a decision support system can accelerate clinical workflows in FECD management in busy ophthalmic clinics. Cataract, a highly prevalent age-related cause of visual impairment, frequently coexists with FECD and is the most performed ophthalmic surgery.76 However, cataract surgery can accelerate endothelial cell loss and increase the risk of corneal decompensation in eyes with pre-existing FECD. AI-enabled diagnostic tools in primary care or high-volume cataract preoperative clinics could automate FECD screening, ensuring early identification of at-risk patients and allowing surgeons to adjust surgical techniques to reduce corneal failure risk. The ability of AI to predict rapid disease progression can guide discussions on surgical timing and postoperative expectations. Additionally, AI-driven prognostication of post-corneal transplantation outcomes (e.g., graft detachment, rejection risk) can improve patient counseling, surgical decision-making, and postoperative care planning. AI could also enable remote post-transplant monitoring, detecting early signs of graft rejection or endothelial failure and reducing unnecessary in-person visits while ensuring timely intervention. The true value of AI lies in eventually reducing reliance on expert corneal specialists for image interpretation, democratizing health care, and improving access to quality care, particularly in resource-limited settings. 
Conclusions
We have conducted the most comprehensive review to date, to our knowledge, on AI algorithms for various clinical contexts of the FECD patient journey. Given the prevalence of FECD as a sight-impairing disease, early detection and prognostication are paramount for effective treatment and to prevent vision loss, making them public health priorities. AI algorithms present promising avenues for achieving these goals and enhancing accessibility. Our summary of relevant publications focuses on input data, algorithm selection, and validation approaches. Future research could explore the potential of AI algorithms based on derived topographic, keratometric, and aberrometric parameters rather than only the raw images, which may be less computationally intensive. The Holistic AI in Medicine (HAIM) framework proposed by Soenksen et al.77, demonstrates how leveraging multimodal data sources—including tabular, time-series, text, and image inputs—can significantly improve the performance of AI models in health care, providing more accurate and robust predictions across a range of clinical tasks. Considering known risk factors such as older age, female sex, White ethnicity, and the genetic heterogeneity of FECD,1,2 such multimodal approaches that merge information from corneal images with demographic and genetic data should be explored. This review offers critical insights for researchers, clinicians, and policymakers, aiding their understanding of existing research and guiding future health service strategies. 
Acknowledgments
Supported by a Medical Research Council/Fight for Sight Clinical Research Training Fellowship (MR/X006271/1 to SL); Moorfields Eye Charity PhD studentship (GR001147 to LK); an Amazon Web Services Scholarship (to LK); National Institute for Health Research AI Award (AI_AWARD02488 to NP); and Czech Health Research Agency Program to support applied health research for the years 2024–2030 (NW25-07-00303 to PL); Technology Agency of the Czech Republic TREND (FW10010347 to PL). 
Disclosure: S. Liu, None; L. Kandakji, None; A. Stupnicki, None; D. Sumodhee, None; M.T. Leucci, None; S. Hau, None; S. Balal, None; A. Okonkwo, None; I. Moghul, None; S.P. Kanda, None; B.D. Allan, None; D.M. Gore, None; K. Muthusamy, None; A.J. Hardcastle, None; A.E. Davidson, None; P. Liskova, None; N. Pontikos, None 
References
Fautsch MP, Wieben ED, Baratz KH, et al. TCF4-mediated Fuchs endothelial corneal dystrophy: insights into a common trinucleotide repeat-associated disease. Prog Retin Eye Res. 2021; 81: 100883. [CrossRef] [PubMed]
Liu S, Sadan AN, Bhattacharyya N, et al. Genetic and demographic determinants of Fuchs endothelial corneal dystrophy risk and severity. JAMA Ophthalmol. 2025; 143(4): 338–347. [CrossRef] [PubMed]
Zoega GM, Fujisawa A, Sasaki H, et al. Prevalence and risk factors for cornea guttata in the Reykjavik Eye Study. Ophthalmology. 2006; 113(4): 565–569. [CrossRef] [PubMed]
Kitagawa K, Kojima M, Sasaki H, et al. Prevalence of primary cornea guttata and morphology of corneal endothelium in aging Japanese and Singaporean subjects. Ophthalmic Res. 2002; 34(3): 135–138. [CrossRef] [PubMed]
Higa A, Sakai H, Sawaguchi S, et al. Prevalence of and risk factors for cornea guttata in a population-based study in a southwestern island of Japan: the Kumejima study. Arch Ophthalmol. 2011; 129(3): 332–336. [CrossRef] [PubMed]
Aiello F, Gallo Afflitto G, Ceccarelli F, Cesareo M, Nucci C. Global prevalence of Fuchs endothelial corneal dystrophy (FECD) in adult population: a systematic review and meta-analysis. J Ophthalmol. 2022; 2022: 1–7. [CrossRef]
Reinprayoon U, Jermjutitham M, Kasetsuwan N. Rate of cornea endothelial cell loss and biomechanical properties in Fuchs’ endothelial corneal dystrophy. Front Med (Lausanne). 2021; 8: 757959. [CrossRef] [PubMed]
NHS Blood and Transplant. Cornea activity. Available at: https://nhsbtdbe.blob.core.windows.net/umbraco-assets-corp/19197/section-10-cornea-activity.pdf. Accessed May 15, 2025.
Krachmer JH, Purcell JJ, Young CW, Bucher KD. Corneal endothelial dystrophy. A study of 64 families. Arch Ophthalmol. 1978; 96(11): 2036–2039. [CrossRef] [PubMed]
Wacker K, McLaren JW, Amin SR, Baratz KH, Patel SV. Corneal high-order aberrations and backscatter in Fuchs’ endothelial corneal dystrophy. Ophthalmology. 2015; 122(8): 1645–1652. [CrossRef] [PubMed]
Sun SY, Wacker K, Baratz KH, Patel SV. Determining subclinical edema in Fuchs endothelial corneal dystrophy: revised classification using Scheimpflug tomography for preoperative assessment. Ophthalmology. 2019; 126(2): 195–204. [CrossRef] [PubMed]
Yasukura Y, Oie Y, Kawasaki R, Maeda N, Jhanji V, Nishida K. New severity grading system for Fuchs endothelial corneal dystrophy using anterior segment optical coherence tomography. Acta Ophthalmol. 2021; 99(6): e914–e921. [CrossRef] [PubMed]
Oie Y, Yamaguchi T, Nishida N, et al. Systematic review of the diagnostic criteria and severity classification for Fuchs endothelial corneal dystrophy. Cornea. 2023; 42(12): 1590–1600. [CrossRef] [PubMed]
Gulshan V, Peng L, Coram M, et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA. 2016; 316(22): 2402–2410. [CrossRef] [PubMed]
Paul SK, Pan I, Sobol WM. A systematic review of deep learning applications for optical coherence tomography in age-related macular degeneration. Retina. 2022; 42(8): 1417–1424. [CrossRef] [PubMed]
Daich Varela M, Sen S, De Guimaraes TAC, et al. Artificial intelligence in retinal disease: clinical application, challenges, and future directions. Graefes Arch Clin Exp Ophthalmol. 2023; 261(11): 3283–3297. [CrossRef] [PubMed]
Nguyen Q, Woof W, Kabiri N, et al. Can artificial intelligence accelerate the diagnosis of inherited retinal diseases? Protocol for a data-only retrospective cohort study (Eye2Gene). BMJ Open. 2023; 13(3): e071043. [CrossRef] [PubMed]
Maile H, Li JPO, Gore D, et al. Machine learning algorithms to detect subclinical keratoconus: systematic review. JMIR Med Inform. 2021; 9(12): e27363. [CrossRef] [PubMed]
Lopes BT, Ramos IC, Salomão MQ, et al. Enhanced tomographic assessment to detect corneal ectasia based on artificial intelligence. Am J Ophthalmol. 2018; 195: 223–232. [CrossRef] [PubMed]
Page MJ, McKenzie JE, Bossuyt PM, et al. The PRISMA 2020 statement: an updated guideline for reporting systematic reviews. BMJ. 2021; 372: n71. [PubMed]
Whiting PF, Rutjes AWS, Westwood ME, et al. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Ann Intern Med. 2011; 155(8): 529–536. [CrossRef] [PubMed]
Karmakar R, Nooshabadi SV, Eghrari AO. Mobile-CellNet: automatic segmentation of corneal endothelium using an efficient hybrid deep learning model. Cornea. 2023; 42(4): 456–463. [CrossRef] [PubMed]
Vigueras-Guillén JP, van Rooij J, van Dooren BTH, et al. DenseUNets with feedback non-local attention for the segmentation of specular microscopy images of the corneal endothelium with guttae. Sci Rep. 2022; 12(1): 14035. [CrossRef] [PubMed]
Vigueras-Guillén JP, van Rooij J, Engel A, Lemij HG, van Vliet LJ, Vermeer KA. Deep learning for assessing the corneal endothelium from specular microscopy images up to 1 year after ultrathin-DSAEK surgery. Transl Vis Sci Technol. 2020; 9(2): 49. [CrossRef] [PubMed]
Shilpashree PS, Suresh KV, Sudhir RR, Srinivas SP. Automated image segmentation of the corneal endothelium in patients with Fuchs dystrophy. Transl Vis Sci Technol. 2021; 10(13): 27. [CrossRef] [PubMed]
Sierra JS, Pineda J, Rueda D, et al. Corneal endothelium assessment in specular microscopy images with Fuchs’ dystrophy via deep regression of signed distance maps. Biomed Opt Express. 2023; 14(1): 335–351. [CrossRef] [PubMed]
Qu JH, Qin XR, Peng RM, et al. A fully automated segmentation and morphometric parameter estimation system for assessing corneal endothelial cell images. Am J Ophthalmol. 2022; 239: 142–153. [CrossRef] [PubMed]
Foo VHX, Lim GYS, Liu YC, et al. Deep learning for detection of Fuchs endothelial dystrophy from widefield specular microscopy imaging: a pilot study. Eye Vis (Lond). 2024; 11(1): 11. [CrossRef] [PubMed]
Qu JH, Qin XR, Xie ZJ, et al. Establishment of an automatic diagnosis system for corneal endothelium diseases using artificial intelligence. J Big Data. 2024; 11(1): 67. [CrossRef]
Eleiwa T, Elsawy A, Özcan E, Abou Shousha M. Automated diagnosis and staging of Fuchs’ endothelial cell corneal dystrophy using deep learning. Eye Vis (Lond). 2020; 7(1): 44. [CrossRef] [PubMed]
Zéboulon P, Ghazal W, Bitton K, Gatinel D. Separate detection of stromal and epithelial corneal edema on optical coherence tomography using a deep learning pipeline and transfer learning. Photonics. 2021; 8(11): 483. [CrossRef]
Bitton K, Zéboulon P, Ghazal W, Rizk M, Elahi S, Gatinel D. Deep learning model for the detection of corneal edema before Descemet membrane endothelial keratoplasty on optical coherence tomography images. Transl Vis Sci Technol. 2022; 11(12): 19. [CrossRef] [PubMed]
Treder M, Lauermann JL, Alnawaiseh M, Eter N. Using deep learning in automated detection of graft detachment in Descemet membrane endothelial keratoplasty: a pilot study. Cornea. 2019; 38(2): 157–161. [CrossRef] [PubMed]
Hayashi T, Tabuchi H, Masumoto H, et al. A deep learning approach in rebubbling after Descemet's membrane endothelial keratoplasty. Eye Contact Lens. 2020; 46(2): 121–126. [CrossRef] [PubMed]
Hayashi T, Iliasian RM, Matthaei M, et al. Transferability of an artificial intelligence algorithm predicting rebubblings after Descemet membrane endothelial keratoplasty. Cornea. 2023; 42(5): 544–548. [CrossRef] [PubMed]
Patefield A, Meng Y, Airaldi M, et al. Deep learning using preoperative AS-OCT predicts graft detachment in DMEK. Transl Vis Sci Technol. 2023; 12(5): 14. [CrossRef] [PubMed]
Heslinga FG, Alberti M, Pluim JPW, Cabrerizo J, Veta M. Quantifying graft detachment after Descemet's membrane endothelial keratoplasty with deep convolutional neural networks. Transl Vis Sci Technol. 2020; 9(2): 48. [CrossRef] [PubMed]
Glatz A, Böhringer D, Zander DB, et al. Three-dimensional map of Descemet membrane endothelial keratoplasty detachment: development and application of a deep learning model. Ophthalmol Sci. 2021; 1(4): 100067. [CrossRef] [PubMed]
Muijzer MB, Hoven CMW, Frank LE, Vink G, Wisse RPL. A machine learning approach to explore predictors of graft detachment following posterior lamellar keratoplasty: a nationwide registry study. Sci Rep. 2022; 12(1): 17705. [CrossRef] [PubMed]
Joseph N, Benetz BA, Chirra P, et al. Machine learning analysis of postkeratoplasty endothelial cell images for the prediction of future graft rejection. Transl Vis Sci Technol. 2023; 12(2): 22. [CrossRef] [PubMed]
Geirhos R, Jacobsen JH, Michaelis C, et al. Shortcut learning in deep neural networks. arXiv [csCV]. Published online April 16, 2020, http://arxiv.org/abs/2004.07780.
Ong Tone S, Jurkunas U. Imaging the corneal endothelium in Fuchs corneal endothelial dystrophy. Semin Ophthalmol. 2019; 34(4): 340–346. [CrossRef] [PubMed]
Al-Fahdawi S, Qahwaji R, Al-Waisy AS, et al. A fully automated cell segmentation and morphometric parameter system for quantifying corneal endothelial cell morphology. Comput Methods Programs Biomed. 2018; 160: 11–23. [CrossRef] [PubMed]
Selig B, Vermeer KA, Rieger B, Hillenaar T, Luengo Hendriks CL. Fully automatic evaluation of the corneal endothelium from in vivo confocal microscopy. BMC Med Imaging. 2015; 15(1): 13. [CrossRef] [PubMed]
Tey KY, Cheong EZK, Ang M. Potential applications of artificial intelligence in image analysis in cornea diseases: a review. Eye Vis (Lond). 2024; 11(1): 10. [CrossRef] [PubMed]
Ishikawa S, Kato N, Takeuchi M. Quantitative evaluation of corneal epithelial edema after cataract surgery using corneal densitometry: a prospective study. BMC Ophthalmol. 2018; 18(1): 334. [CrossRef] [PubMed]
Wertheimer CM, Elhardt C, Wartak A, et al. Corneal optical density in Fuchs endothelial dystrophy determined by anterior segment optical coherence tomography. Eur J Ophthalmol. 2021; 31(4): 1771–1778. [CrossRef] [PubMed]
Zéboulon P, Ghazal W, Gatinel D. Corneal edema visualization with optical coherence tomography using deep learning: proof of concept. Cornea. 2021; 40(10): 1267. [CrossRef] [PubMed]
Li Z, Liu F, Yang W, Peng S, Zhou J. A survey of convolutional neural networks: analysis, applications, and prospects. IEEE Trans Neural Netw Learn Syst. 2022; 33(12): 6999–7019. [CrossRef] [PubMed]
Ronneberger O, Fischer P, Brox T. U-Net: convolutional networks for biomedical image segmentation. arXiv. 2015, https://doi.org/10.48550/arXiv.1505.04597.
Liu S, Wong YL, Walkden A. Current perspectives on corneal transplantation. Clin Ophthalmol. 2022; 16: 631–646. [CrossRef] [PubMed]
Ang M, Wilkins MR, Mehta JS, Tan D. Descemet membrane endothelial keratoplasty. Br J Ophthalmol. 2016; 100(1): 15–21. [CrossRef] [PubMed]
Deng SX, Lee WB, Hammersmith KM, et al. Descemet membrane endothelial keratoplasty: safety and outcomes: a report by the American Academy of Ophthalmology. Ophthalmology. 2018; 125(2): 295–310. [CrossRef] [PubMed]
Parekh M, Leon P, Ruzza A, et al. Graft detachment and rebubbling rate in Descemet membrane endothelial keratoplasty. Surv Ophthalmol. 2018; 63(2): 245–250. [CrossRef] [PubMed]
Zhang H, Meng Y, Zhao Y, et al. DTFD-MIL: double-tier feature distillation multiple instance learning for histopathology whole slide image classification. arXiv. 2022, https://doi.org/10.48550/arXiv.2203.12081.
He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Piscataway, NJ: Institute of Electrical and Electronics Engineers; 2016: 770–778.
Rodríguez-Calvo-de-Mora M, Quilendrino R, Ham L, et al. Clinical outcome of 500 consecutive cases undergoing Descemet's membrane endothelial keratoplasty. Ophthalmology. 2015; 122(3): 464–470. [CrossRef] [PubMed]
Baydoun L, Tong CM, Tse WW, et al. Endothelial cell density after Descemet membrane endothelial keratoplasty: 1 to 5-year follow-up. Am J Ophthalmol. 2012; 154(4): 762–763. [CrossRef] [PubMed]
Schlögl A, Tourtas T, Kruse FE, Weller JM. Long-term clinical outcome after Descemet membrane endothelial keratoplasty. Am J Ophthalmol. 2016; 169: 218–226. [CrossRef] [PubMed]
Joseph N, Kolluru C, Benetz BAM, Menegay HJ, Lass JH, Wilson DL. Quantitative and qualitative evaluation of deep learning automatic segmentations of corneal endothelial cell images of reduced image quality obtained following cornea transplant. J Med Imaging (Bellingham). 2020; 7(1): 014503. [PubMed]
Kandakji L, Maile H, Li O, et al. Keratoconus classification and segmentation using deep learning on raw anterior-segment optical coherence tomography imaging. Invest Ophthalmol Vis Sci. 2023; 64(8): 1099–1099.
Maile HP, Li JPO, Fortune MD, et al. Personalized model to predict keratoconus progression from demographic, topographic, and genetic data. Am J Ophthalmol. 2022; 240: 321–329. [CrossRef] [PubMed]
Patel SV. Imaging Fuchs endothelial corneal dystrophy in clinical practice and clinical trials. Cornea. 2021; 40(12): 1505–1511. [CrossRef] [PubMed]
Kuerten D, Plange N, Koch EC, Koutsonas A, Walter P, Fuest M. Central corneal thickness determination in corneal edema using ultrasound pachymetry, a Scheimpflug camera, and anterior segment OCT. Graefes Arch Clin Exp Ophthalmol. 2015; 253(7): 1105–1109. [CrossRef] [PubMed]
Lee CS, Tyring AJ, Deruyter NP, Wu Y, Rokem A, Lee AY. Deep-learning based, automated segmentation of macular edema in optical coherence tomography. Biomed Opt Express. 2017; 8(7): 3440–3448. [CrossRef] [PubMed]
Yamashita R, Nishio M, Do RKG, Togashi K. Convolutional neural networks: an overview and application in radiology. Insights Imaging. 2018; 9(4): 611–629. [CrossRef] [PubMed]
Rashidisabet H, Sethi A, Jindarak P, et al. Validating the generalizability of ophthalmic artificial intelligence models on real-world clinical data. Transl Vis Sci Technol. 2023; 12(11): 8. [CrossRef] [PubMed]
Oshika T. Artificial intelligence applications in ophthalmology. JMA J. 2025; 8(1): 66–75. [PubMed]
Ting DSW, Pasquale LR, Peng L, et al. Artificial intelligence and deep learning in ophthalmology. Br J Ophthalmol. 2019; 103(2): 167–175. [CrossRef] [PubMed]
Chen D, Geevarghese A, Lee S, et al. Transparency in artificial intelligence reporting in ophthalmology – a scoping review. Ophthalmol Sci. 2024; 4(4): 100471. [CrossRef] [PubMed]
Hutchinson ML, Antono E, Gibbons BM, Paradiso S, Ling J, Meredig B. Overcoming data scarcity with transfer learning. arXiv. 2017, https://doi.org/10.48550/arXiv.1711.05099.
Loftus TJ, Ruppert MM, Shickel B, et al. Federated learning for preserving data privacy in collaborative healthcare research. Digit Health. 2022; 8:20552076221134456.
Azad R, Kazerouni A, Heidari M, et al. Advances in medical image analysis with vision Transformers: a comprehensive review. Med Image Anal. 2024; 91: 103000. [CrossRef] [PubMed]
Arora A, Arora A. Generative adversarial networks and synthetic patient data: current challenges and future perspectives. Future Healthc J. 2022; 9(2): 190–193. [CrossRef] [PubMed]
van der Velden BHM, Kuijf HJ, Gilhuijs KGA, Viergever MA. Explainable artificial intelligence (XAI) in deep learning-based medical image analysis. Med Image Anal. 2022; 79: 102470. [CrossRef] [PubMed]
Romano V, Passaro ML, Bachmann B, et al. Combined or sequential DMEK in cases of cataract and Fuchs endothelial corneal dystrophy – a systematic review and meta-analysis. Acta Ophthalmol. 2024; 102(1): e22–e30. [CrossRef] [PubMed]
Soenksen LR, Ma Y, Zeng C, et al. Integrated multimodal artificial intelligence framework for healthcare applications. NPJ Digit Med. 2022; 5(1): 149. [CrossRef] [PubMed]
Figure 1.
 
PRISMA flow diagram outlining the study selection process for the systematic review. This diagram illustrates the steps taken to include and exclude studies based on predefined eligibility criteria, highlighting the number of studies at each stage of the review process, including identification, screening, eligibility, and inclusion.
Figure 1.
 
PRISMA flow diagram outlining the study selection process for the systematic review. This diagram illustrates the steps taken to include and exclude studies based on predefined eligibility criteria, highlighting the number of studies at each stage of the review process, including identification, screening, eligibility, and inclusion.
Figure 2.
 
Heatmap of performance metrics across studies. Rows represent individual studies, and columns indicate accuracy, Dice coefficient, sensitivity, and specificity. Color intensity reflects performance, with darker shades indicating higher values. Metrics are categorized as ≤0.50 (lightest blue), 0.50 to 0.70 (light blue), 0.70 to 0.85 (medium blue), and ≥0.85 (darkest blue). Gray cells indicate that the corresponding study did not report this metric. High-performing models and variability in evaluation criteria can be readily observed in this figure. Comprehensive details on datasets, AI pipelines, and additional evaluation metrics are provided in the Table, which contextualizes these results within the broader systematic review.
Figure 2.
 
Heatmap of performance metrics across studies. Rows represent individual studies, and columns indicate accuracy, Dice coefficient, sensitivity, and specificity. Color intensity reflects performance, with darker shades indicating higher values. Metrics are categorized as ≤0.50 (lightest blue), 0.50 to 0.70 (light blue), 0.70 to 0.85 (medium blue), and ≥0.85 (darkest blue). Gray cells indicate that the corresponding study did not report this metric. High-performing models and variability in evaluation criteria can be readily observed in this figure. Comprehensive details on datasets, AI pipelines, and additional evaluation metrics are provided in the Table, which contextualizes these results within the broader systematic review.
Figure 3.
 
Summary diagram presenting the various AI approaches used in the management of FECD. The diagram categorizes the AI algorithms employed in different clinical contexts, including the assessment of corneal endothelium, evaluation of corneal edema, and prognostication of post-corneal transplantation graft detachment and rejection. This overview highlights the application of AI methods within clinical workflow for managing FECD.
Figure 3.
 
Summary diagram presenting the various AI approaches used in the management of FECD. The diagram categorizes the AI algorithms employed in different clinical contexts, including the assessment of corneal endothelium, evaluation of corneal edema, and prognostication of post-corneal transplantation graft detachment and rejection. This overview highlights the application of AI methods within clinical workflow for managing FECD.
Figure 4.
 
Slit-scanning in vivo confocal microscopy images of the corneal endothelium. (A) Healthy control cornea. Image of a normal corneal endothelium shows a uniform hexagonal cell pattern with consistent cell density and no visible guttae or abnormalities (white solid arrowhead). (B) Early-stage FECD. Image shows isolated guttae (white dashed arrowhead), characterized by dark, round bodies with central white hyperreflectivity, indicating early endothelial cell distress. (C) Late-stage FECD. Increased confluency of guttae (white dashed arrowhead), illustrating a more extensive presence of these structures (dashed-line highlighted area) with evidence of endothelial cell loss.
Figure 4.
 
Slit-scanning in vivo confocal microscopy images of the corneal endothelium. (A) Healthy control cornea. Image of a normal corneal endothelium shows a uniform hexagonal cell pattern with consistent cell density and no visible guttae or abnormalities (white solid arrowhead). (B) Early-stage FECD. Image shows isolated guttae (white dashed arrowhead), characterized by dark, round bodies with central white hyperreflectivity, indicating early endothelial cell distress. (C) Late-stage FECD. Increased confluency of guttae (white dashed arrowhead), illustrating a more extensive presence of these structures (dashed-line highlighted area) with evidence of endothelial cell loss.
Figure 5.
 
Cross-sectional AS-OCT images of the cornea in FECD. (A) Early-stage FECD. AS-OCT image of the cornea shows no clinical signs of corneal edema. The corneal thickness remains within the normal range, and no significant epithelial or stromal changes are visible. (B) Late-stage FECD. AS-OCT image shows increased central corneal thickness, epithelial bullae (fluid-filled cystic spaces), and increased pixel intensity (backscatter) in the stroma, indicating corneal decompensation and edema.
Figure 5.
 
Cross-sectional AS-OCT images of the cornea in FECD. (A) Early-stage FECD. AS-OCT image of the cornea shows no clinical signs of corneal edema. The corneal thickness remains within the normal range, and no significant epithelial or stromal changes are visible. (B) Late-stage FECD. AS-OCT image shows increased central corneal thickness, epithelial bullae (fluid-filled cystic spaces), and increased pixel intensity (backscatter) in the stroma, indicating corneal decompensation and edema.
Figure 6.
 
(A, B) Cross-sectional AS-OCT images of the cornea showing a small partial graft detachment after DMEK (A) and a more extensive graft detachment (B).
Figure 6.
 
(A, B) Cross-sectional AS-OCT images of the cornea showing a small partial graft detachment after DMEK (A) and a more extensive graft detachment (B).
Table.
 
Summary of the Published Studies That Included the Use of AI for the Assessment, Diagnosis, and Prognostication of FECD
Table.
 
Summary of the Published Studies That Included the Use of AI for the Assessment, Diagnosis, and Prognostication of FECD
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×