Translational Vision Science & Technology Cover Image for Volume 14, Issue 6
June 2025
Volume 14, Issue 6
Open Access
Artificial Intelligence  |   June 2025
AI Quantification of Vascular Lesions in Mouse Fundus Fluorescein Angiography
Author Affiliations & Notes
  • Vinodhini Jayananthan
    Department of Ophthalmology and Visual Sciences, College of Medicine, The Ohio State University Wexner Medical Center, Columbus, OH, USA
  • Tyler Heisler Taylor
    Department of Ophthalmology and Visual Sciences, College of Medicine, The Ohio State University Wexner Medical Center, Columbus, OH, USA
  • David Henry Greentree
    Department of Ophthalmology and Visual Sciences, College of Medicine, The Ohio State University Wexner Medical Center, Columbus, OH, USA
  • Bryce Collison
    Department of Neuroscience, College of Medicine, The Ohio State University Wexner Medical Center, Columbus, OH, USA
  • Nagaraj Kerur
    Department of Ophthalmology and Visual Sciences, College of Medicine, The Ohio State University Wexner Medical Center, Columbus, OH, USA
    Department of Microbial Infection and Immunity, College of Medicine, The Ohio State University Wexner Medical Center, Columbus, OH, USA
  • Correspondence: Nagaraj Kerur, Department of Ophthalmology and Visual Sciences, College of Medicine, The Ohio State University Wexner Medical Center, 410 West 10th Avenue, Columbus, OH 43210, USA. e-mail: [email protected] 
Translational Vision Science & Technology June 2025, Vol.14, 4. doi:https://doi.org/10.1167/tvst.14.6.4
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Vinodhini Jayananthan, Tyler Heisler Taylor, David Henry Greentree, Bryce Collison, Nagaraj Kerur; AI Quantification of Vascular Lesions in Mouse Fundus Fluorescein Angiography. Trans. Vis. Sci. Tech. 2025;14(6):4. https://doi.org/10.1167/tvst.14.6.4.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose: Quantifying vascular leakage in fundus fluorescein angiography (FFA) is a critical endpoint in preclinical models of diseases such as neovascular age-related macular degeneration, retinopathy of prematurity, and diabetic retinopathy. Traditional manual methods are labor intensive and prone to variability. We developed an artificial intelligence (AI)-assisted method to improve efficiency and accuracy in quantifying vascular lesions in FFA images.

Methods: Nikon NIS-Elements software with AI functionality was used to create an automated FFA analysis method. FFA images were acquired using the Phoenix MICRON IV imaging system in two mouse models of ocular angiogenesis: (1) very low-density lipoprotein receptor (Vldlr) knockout mice exhibiting spontaneous pathological chorioretinal neovascularization, and (2) a laser-induced choroidal neovascularization model. The AI model was trained on manually segmented FFA images to delineate lesions and quantify lesion area and fluorescence intensity.

Results: The AI model demonstrated high accuracy in quantifying vascular lesions in FFA images, achieving 99.7% agreement with manual counts. It attained a precision, recall, and F1 score of 0.94, with an intraclass correlation coefficient (ICC) of 0.991. The model showed strong spatial agreement with manual segmentations and consistent lesion area measurements. On validation images, it maintained expert-level performance (ICC = 0.998) with high sensitivity and precision. Additionally, it effectively captured temporal changes in vascular leakage by measuring lesion area and fluorescence intensity, demonstrating robustness in real-world experiments.

Conclusions: Our AI model quantifies vascular lesions in FFA images with high accuracy, outperforming manual analysis.

Translational Relevance: AI-based quantification provides a scalable, consistent alternative to manual methods, enhancing research efficiency.

Introduction
Abnormal growth and leakage of blood vessels are characteristic features of several sight-threatening ocular conditions, including neovascular age-related macular degeneration (nvAMD), retinopathy of prematurity, and diabetic retinopathy.1 Fundus fluorescein angiography (FFA) highlights the retinal blood vessels, allowing for detection of abnormal growth and leakage. The traditional methods for quantifying vascular lesions involve manual delineation and measurement of the lesional area in each image. This approach of lesion quantification in FFA images is not only labor intensive and time consuming but also prone to variability and user bias. Additionally, this manual approach limits the feasibility of large-scale studies and introduces inconsistencies. 
Recent advancements in artificial intelligence (AI) and machine learning have opened new possibilities for automating image analysis. AI-based tools have shown impressive performance in analyzing and quantifying pathological features across various imaging modalities, including radiology, pathology, and ophthalmology.2 For FFA assessment in preclinical animal models, AI-based workflows can potentially overcome limitations associated with manual quantification, providing a faster, repeatable, and standardized approach to analyzing vascular lesions. 
In this study, we developed an AI-based analysis method designed for the automated quantification of vascular lesions in mouse FFA images. Specifically, we focused on two preclinical mouse models relevant to nvAMD: the very low-density lipoprotein receptor knockout (Vldlr−/−) mouse, and a laser-induced choroidal neovascularization (CNV) model.35 Although the neovascularization in the laser CNV model mimics the pathology of type 2 nvAMD, where vessels penetrate the retinal pigment epithelium (RPE) and leak into the subretinal space, the lesions in the Vldlr−/− model resemble type 3 nvAMD, characterized by neovascularization originating from the deep retinal capillary plexus and extending downward toward the RPE. Our AI-based method accurately detects and quantifies vascular abnormalities in these models, allowing automated analysis of large number of images. 
Methods
Animals
Male and female C57BL/6J wild-type (WT) mice 8 to 12 weeks old and male and female Vldlr−/− mice 3 to 11 weeks old were used in this study. Both C57BL/6J and Vldlr−/− mice were obtained from The Jackson Laboratory (Bar Harbor, ME). All experimental procedures were approved by the Institutional Animal Care and Use Committee of The Ohio State University. Studies were conducted in accordance with the principles outlined in the ARVO Statement for the Use of Animals in Ophthalmic and Vision Research. The mice were housed in a standard sterile facility with ventilated cage systems. The environment maintained a controlled temperature and humidity, with a 12-hour light/12-hour dark cycle. The animals had access to sterilized, irradiated rodent diet (a standard laboratory mouse diet) ad libitum and received sterile water through automatic water systems. 
Laser-Induced Mouse Model of Choroidal Neovascularization
Laser photocoagulation using the MICRON IV Image-Guided Laser System (Phoenix-Micron, Bend, OR) was performed following a previously established procedure.6 This system utilizes a Merilas 532 green laser photocoagulator (Meridian Medical, Thun, Switzerland). Laser photocoagulation lesions were induced with spot diameter of 50 µm, delivering a 70-ms pulse of 220-mW power. Prior to each experiment, the laser system was calibrated to ensure accurate power output. The mice were anesthetizedwith Avertin (500 µg/g body weight) via intraperitoneal injection. Their eyes were then dilated with tropicamide and phenylephrine drops, and a general eye lubricant (GenTeal Tears) was applied to the cornea. When the pupils were dilated, the mice were positioned to visualize fundus on a computer screen connected to the MICRON IV system, and the laser beam was directed onto the fundus. Four laser burns were created around the optic disc at the 3, 6, 9, and 12 o'clock positions, approximately 2 to 3 optic disc diameters away from the optic nerve. Confirmation of the Bruch's membrane rupture was observed by the formation of a bubble at the center of the laser spot. Following the laser procedure, mice were placed under a heat lamp until they recovered from anesthesia and were returned to the vivarium. 
Vldlr−/− Mouse Model of Spontaneous Retinal Neovascularization
In Vldlr−/− mice, abnormal blood vessels from the deep retinal vasculature within the outer plexiform layer continue to proliferate, forming choroidal anastomosis typically by 2 months of age, mimicking the features of nvAMD and retinal angiomatous proliferation.4,7,8 
FFA Imaging
FFA images were captured from both the Vldlr−/− and laser-induced CNV mouse models using the MICRON IV imaging system. For the laser-induced CNV model, FFA images were taken on day 3 or day 7 post-laser. For the Vldlr−/− mice, images were taken at 3 to 11 weeks of age . Mice were anesthetized and pupils were dilated with topical drops of tropicamide and phenylephrine as described above. Mice were then intraperitoneally administered with fluorescein (30 µg/gm of body weight, Fluorescite 10%), and FFA images were captured at 5 and 9 minutes for the right eye, and at 6 and 10 minutes for the left eye following fluorescein injection. To ensure a diverse dataset, images used for AI training and validation were acquired with a gain setting range of 0 to 3 and frame rate of 30 fps. For temporal leakage analysis, images were consistently captured at a fixed gain setting of 0 and a frame rate of 30 fps. 
Image Calibration
All FFA images used in this study were captured using the MICRON IV fundus camera. The original TIFF files were imported into the NIS-Elements software (Nikon, Tokyo, Japan) and subsequently converted to ND2 format for further analysis. Image calibration was performed within the software, using a pixel size of 2.1 µm per pixel to ensure accurate spatial measurements during downstream processing. 
Training Dataset and Manual Segmentation
FFA images from laser CNV in WT and Vldrl−/− mouse models were manually segmented to delineate vascular lesions using the NIS-Elements manual segmentation tool. The NIS.ai suite within the NIS-Elements software was trained on a dataset of manually segmented FFA images identifying vascular lesions, which served as the ground truth. To account for interoperator variability, seven of these images were manually segmented independently by three different operators. Each of these operator-marked segmentations was treated as a unique training image, resulting in 21 images for training from these seven original images. Additionally, two other images were manually segmented independently by two operators, yielding four images for training from this subset. Finally, 55 unique images, each manually segmented by a single operator, were included in the training dataset. Overall, this segmentation strategy resulted in a final training set of 70 images, incorporating interoperator variability into the ground truth to enhance the ability of the model to generalize. Objects smaller than 60 µm2 were not included to eliminate artifacts. Across these 70 images, 1796 vascular lesions were identified and manually segmented. In each FFA image, the number of vascular lesions, lesion area, and mean fluorescence intensity were computed. 
Images for Validation
Nine images not included in the training set were manually segmented by a single operator to validate the AI method. These images contained a total of 127 FFA lesions, all manually annotated by the same operator. 
Performance Evaluation Metrics
A comprehensive set of metrics was employed to evaluate the performance of the AI method against the manual approach. These metrics included (1) metrics of sensitivity and specificity, capturing the ability of the model to correctly identify true positive and true negative pixels; (2) metrics of spatial agreement between manual and AI methods; and (3) metrics of lesion area measurement accuracy. 
Metrics of Lesion Detection Accuracy and Sensitivity
The object (i.e., vascular lesion) prediction accuracy of the model was assessed using precision, recall, and F1 score metrics. Precision measures the proportion of correctly predicted lesions out of all predicted lesions, and recall measures the proportion of correctly predicted lesions out of all actual lesions. Both metrics are based on the concept of true positives (TPs), false positives (FPs), and false negatives (FNs). To balance both precision and recall, we also calculated the F1 score, a harmonic mean of precision and recall. 
Precision
Precision measures the accuracy of the AI method in predicting the presence of an object (i.e., vascular lesion in this context) in the image. It is the ratio of number of correctly predicted objects (TPs) to all objects predicted by the AI (TP + FP):  
\begin{eqnarray*}{\rm{Precision}} = \frac{{{\rm{TP}}}}{{{\rm{TP}} + {\rm{FP}}}}\end{eqnarray*}
 
The precision value ranges from 0 to 1, where 
  • If precision = 1, there were no FPs (AI correctly identified all objects without extra detections).
  • If precision = 0, none of the predicted positives was correct (all FPs).
Recall (Sensitivity)
Recall measures the ability of the AI method to correctly identify true positives. It is the ratio of correctly predicted objects (TPs) to all actual objects (TP + FN):  
\begin{eqnarray*}{\rm{Recall}} = \frac{{{\rm{TP}}}}{{{\rm{TP}} + {\rm{FN}}}}\end{eqnarray*}
 
The recall value ranges from 0 to 1, where 
  • If recall = 1, all actual positives were correctly predicted (no FNs).
  • If recall = 0, none of the actual positives was predicted (all FNs).
F1 Score
A harmonic mean of precision and recall, the F1 score provides a single metric that balances both precision and recall, giving an overall assessment of the performance of the model.  
\begin{eqnarray*}{\rm{F}}1{\rm{\ Score}} = 2{\rm{\ }} \times {\rm{\ }}\frac{{{\rm{Precision\ }} \times {\rm{\ Recall}}}}{{{\rm{Precision}} + {\rm{Recall}}}}\end{eqnarray*}
 
The F1 score ranges from 0 to 1, where 
  • F1 = 1 indicates perfect precision and recall, meaning the model made no FPs or FNs (all positive predictions were correct and all actual positives were detected).
  • F1 = 0 indicates that either precision or recall was 0, meaning either the model had no correct positive predictions (precision = 0) or it failed to detect any TPs (recall = 0).
Metric of Spatial Agreement Between Manual and AI Methods
To quantitatively assess the spatial overlap between manual and AI-predicted segmentations, we calculated the intersection over union (IoU). IoU is a precise metric for spatial agreement through measuring the ratio of the area of overlap between the two segmentations to their combined area:  
\begin{eqnarray*} {\rm{IoU}} = \frac{{{\rm{Area\ of\ intersection}}}} {\begin{array}{l}{\rm Area\ of\ manual\ segmentation} \\\quad + \, {\rm Area\, of\ AI\ prediction - Area\ of\ intersection}\end{array}} \end{eqnarray*}
 
The IoU value ranges from 0 to 1, wheres 
  • If IoU = 1, the segmentations are identical.
  • If IoU = 0, there is no overlap between the two segmentations.
Metrics of Quantitative Difference in the Lesion Area Measurement Between Manual and AI Methods
The differences in lesion area measurements between the manual and AI-based segmentation methods were quantified using percentage error (PE). PE expresses the discrepancy between the ground truth and AI method lesion areas as a percentage of the ground truth, offering a clear and standardized measure of the magnitude of deviation between AI predictions and manual segmentations:  
\begin{eqnarray*}{\rm{PE}} = \frac{{{\rm{GT\ area}} - {\rm{AI\ area}}}}{{{\rm{GT\ area}}}} \times {\rm{\ }}100\end{eqnarray*}
where GT area is the ground truth area of the lesion, and AI area is the AI-predicted area of the lesion. 
Measurement of Leakage Intensity
Vascular leakage intensity in FFA lesions was quantified using pixel-based mean fluorescence intensity (MFI). Regions of interest (ROIs), delineating individual lesion boundaries, were established through manual or AI-based segmentation. The MFI was then calculated for each ROI, providing a measure of leakage intensity per lesion, as well as an average MFI across all lesions within each image.  
\begin{eqnarray*}{\rm{MFI}} = \ \frac{{\Sigma \left( {{\rm{Pixel\ intensities}}} \right)}}{{{\rm{Number\ of\ pixels}}}}\end{eqnarray*}
 
Intraclass Correlation Coefficient
To evaluate the agreement between AI-predicted ground truth and manual measurements, the intraclass correlation coefficient (ICC) was computed for both lesion prediction and lesion area measurement. The ICC is a statistical measure that assesses both the consistency and absolute agreement between the two methods, providing insight into how well the AI-predicted lesion count and areas align with the manual method. Together these metrics offer a comprehensive evaluation of the accuracy of the AI model in area prediction and the reliability of its performance compared to the ground truth. An ICC of 1 indicates perfect agreement (identical counts for all images), whereas an ICC of 0 signifies no agreement beyond random chance. 
Instructions for Implementing the AI-Based Analysis Method
To facilitate the seamless implementation of our AI-based analysis, we provide the following resources and instructions: (1) Segment Objects AI (.oai) file, which contains the trained AI model optimized for delineating FFA lesions Supplementary Files AI model 2.0.zip; (2) General Analysis 3 (.ga3) file, which defines the analysis workflow, including automated lesion counting, area measurement, intensity analysis, and data tabulation Supplementary Files AI Analysis FFA Lesion area X Intensity with CSV export FINAL.zip and fundus calibration.zip; and (3) a step-by-step guide that provides detailed instructions covering software setup, image preparation, analysis execution, and result tabulation. 
Results
AI Performance in Lesion Detection in Training Images
Seventy mouse FFA images containing 1796 manually demarcated vascular lesions were used to train the segmentation model via the Train Segment Objects.ai feature in the NIS-Elements software. To ensure accurate AI-predicted lesions, an AI-predicted lesion was considered true positive only if it overlapped with at least half of the combined area delineated by both manual and AI methods. This overlap was measured using an IoU threshold of 0.5. 
Performance analysis demonstrated a strong agreement between AI predictions and ground truth lesion counts. The AI method detected 1801 lesions across 70 training images compared to ground truth, which identified 1796 lesions, indicating 99.7% agreement in lesion count overall (Fig. 1A). Of the 1801 AI-predicted lesions, 137 lesions were not identified by the manual method (Fig. 1B). Furthermore, AI failed to detect 132 of the manually marked lesions (Fig. 1B). A detailed comparison of lesion counts across all 70 images demonstrated excellent agreement, with an ICC of 0.991, calculated using a two-way mixed-effects model for absolute agreement (Fig. 1C). 
Figure 1.
 
Evaluation of the AI method performance in lesion detection on training images. (A) Heatmap showing the lesion count in each of the 70 training images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Distribution of percentage deviation in AI-predicted lesion counts per image compared to the ground truth, represented using a box-and-whisker plot and a pie chart. (E) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (F) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method.
Figure 1.
 
Evaluation of the AI method performance in lesion detection on training images. (A) Heatmap showing the lesion count in each of the 70 training images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Distribution of percentage deviation in AI-predicted lesion counts per image compared to the ground truth, represented using a box-and-whisker plot and a pie chart. (E) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (F) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method.
Table.
 
Performance Metrics for the AI Method, Including Results on Training and Validation Images and Interoperator Comparisons
Table.
 
Performance Metrics for the AI Method, Including Results on Training and Validation Images and Interoperator Comparisons
Analysis at the individual image level revealed that exact matches in object counts were observed in 37.1% of the images (26 out of 70). Additionally, 58.6% of the images (41 out of 70) showed a minor discrepancy (±5 lesions), resulting in over 95% of images demonstrating minimal differences. In contrast, two images (2.9%) were classified as underestimated by the AI, with more than five fewer lesions predicted compared to the ground truth. Conversely, one image (1.4%) was classified as overestimated, with the AI predicting more than five additional lesions compared to the ground truth (Fig. 1D). The Bland–Altman analysis revealed that the AI method slightly underestimated lesion counts compared to the manual method, with a mean bias of −0.45% (Fig. 1E). 
Further examination of images with count discrepancies revealed that these differences largely stemmed from the merging of closely located lesions, where accurately defining boundaries posed challenges. Our training dataset included several images independently segmented by three operators to account for interoperator variability. These images, which included merged lesions, also displayed variation in lesion counts across operators, suggesting that separating some merged lesions into distinct entities is inherently subjective (Supplementary Figs. S1, S2). Notably, despite count variations among the three human operators, AI-predicted lesion counts remained consistent, even when the same image was presented multiple times (each instance corresponding to a different operator's annotation) in the training dataset (Supplementary Fig. S1). This consistency underscores the robustness of the AI model in lesion counting, demonstrating that it provides a deterministic, repeatable method of quantification, ensuring reliable results even in the presence of complex lesion morphologies and interoperator differences. 
The lesion detection performance of the AI model was further evaluated using key accuracy and sensitivity metrics: precision, recall, and F1 score. Across the 70 analyzed images, the model achieved an average precision of 0.94, a recall (sensitivity) of 0.94, and an F1 score of 0.94 (Fig. 1F), underscoring its high accuracy and consistency in accurately identifying vascular lesions. A precision of 0.94 indicates that, when the model predicts a lesion, it is correct 94% of the time. Similarly, a recall of 0.94 means the model successfully detects 94% of all actual lesions. The high F1 score, as a harmonic mean of precision and recall, confirms the balanced performance of the model in both identifying true lesions and avoiding incorrect detections. These metrics demonstrate the robustness and reliability of the model in detecting lesions with minimal FPs and missed detections. 
Quantifying Spatial Agreement Between Manual and AI-Predicted Lesions With IoU
The spatial agreement between manual and AI-predicted lesion segmentations was assessed using IoU. IoU was selected over the dice similarity coefficient because of its more stringent requirements for spatial overlap and its extensive use in the literature for evaluating segmentation accuracy in object boundaries in medical imaging.913 To account for potential variations in lesion delineation approaches (example lesions are identified in Supplementary Fig. S2), such as cases where one method merged multiple lesions into a single region but the other separated them, an adapted IoU calculation was applied. For instances of merged lesions, the aggregated area of individual lesions nested within the bigger marking was used to compare with the merged region. This approach allowed for meaningful evaluation of spatial agreement between manual and AI-derived lesion boundaries, even when segmentation styles differed. 
As described above, the AI segmentation model was trained on 70 images containing 1796 manually demarcated lesions. At the individual lesion level, AI predictions demonstrated high fidelity, with an average IoU of 0.86 (95% confidence interval [CI], 0.85–0.86). Notably, 80% of the observations recorded an IoU of 0.8 or higher, highlighting the strong spatial agreement of the model with manual annotations (Fig. 2A). At the individual image level, the overall spatial agreement of the lesion areas further improved, achieving an average IoU of 0.89 (95% CI, 0.87–0.90), suggesting a strong agreement in lesion boundaries across entire FFA images (Fig. 2B). 
Figure 2.
 
Evaluation of the AI method performance in lesion spatial agreement and area on training images. (A) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (B) Spatial agreement at the level of individual image between ground truth and AI method quantified by IoU. (C) Correlation analysis of lesion area per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion area per image between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted lesion area per image compared to the ground truth, represented using a box-and-whisker plot and frequency table.
Figure 2.
 
Evaluation of the AI method performance in lesion spatial agreement and area on training images. (A) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (B) Spatial agreement at the level of individual image between ground truth and AI method quantified by IoU. (C) Correlation analysis of lesion area per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion area per image between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted lesion area per image compared to the ground truth, represented using a box-and-whisker plot and frequency table.
Quantitative Assessment of Lesion Area Measurement: AI Versus Manual Methods
The AI segmentation model, as described above, demonstrated high precision in detecting vascular lesions, achieving an F1 score of 0.94 (Fig. 1F) and a strong spatial agreement with manual segmentations, reflected by an IoU of 0.86 (Figs. 2A, 2B). By determining ICC, bias (Bland–Altman analysis), and percentage error, we have quantitatively assessed the accuracy of lesion area measurements by the AI model. This was performed both at the overall image level (cumulative lesion area for each FFA image) and at the individual lesion level (individual lesions across all 70 images). PE represents the difference between the ground truth lesion area and the AI-estimated area, expressed as a percentage of the ground truth value. A positive PE value indicates that the AI underestimated the lesion area, and a negative PE shows overestimation. This metric helps to quickly assess the accuracy and reliability of the AI model relative to manual measurements across different lesion sizes. 
Image-Level Analysis (Cumulative Lesion Area Per Image)
Cumulative lesion area for each FFA image was calculated by summation of areas of individual lesions. Our AI-based method demonstrated an excellent agreement in the measurement of cumulative lesion area per image across all 70 training images, with an ICC of 0.994, calculated using a two-way mixed-effects model for absolute agreement (Fig. 2C). Next, a Bland–Altman analysis was performed to evaluate the agreement between the AI-based and manual methods for measuring cumulative lesion areas. The analysis revealed minimal bias of 0.09%, suggesting excellent overall agreement between the two methods (Fig. 2D). The 95% limits of agreement were calculated as −10.04% to 10.22%, meaning that, for most images, the AI method underestimated the lesion area by up to 10.04% or overestimated it by up to 10.22% compared to the manual method. The narrow range of the limits of agreement further supports the consistency of the AI measurements relative to the manual method. 
We also assessed the deviation of the AI model from manual measurements by calculating PE. The model demonstrated strong accuracy, with 90% of images showing a deviation of no more than ±10% from the ground truth (Fig. 2E, Supplementary Fig. S3). Additionally, 55.7% of the images demonstrated a deviation within ±2%, emphasizing the reliability of the AI method for lesion area estimation (Fig. 2E, Supplementary Fig. S3). 
Lesion-Level Analysis (Individual Lesion Area)
The AI method detected 1801 lesions across 70 images analyzed compared to ground truth, which identified 1796 lesions. In cases of merged lesions, the combined area of individual lesions nested within the bigger markings (example lesions identified in Supplementary Fig. S2) was used to compare with the merged region, ensuring accurate and fair comparison of lesion areas. The AI method demonstrated strong agreement in lesion area measurements, with an ICC of 0.988, calculated using a two-way mixed-effects model for absolute agreement (Fig. 3A). Bland–Altman analysis indicated a bias of 0.47%, with 95% limits of agreement ranging from −32.64% to 33.57% (Fig. 3B). 
Figure 3.
 
Evaluation of the AI method performance in individual lesion area measurement on training images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (C) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot and frequency table.
Figure 3.
 
Evaluation of the AI method performance in individual lesion area measurement on training images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (C) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot and frequency table.
We examined PE across the 1664 lesions that were detected by both manual and AI methods (Fig. 3C). Nearly half of the lesions (51.4%) had predictions falling within a narrow deviation range of −5% to +5% relative to ground truth (negative values indicate over estimation, and positive values indicate underestimation) (Fig. 3C, Supplementary Fig. S4). Furthermore, 63% of lesions had predictions falling within a ±10% margin of error, and 80% of lesions had predictions within a ±15% range. Together, these findings reflect the consistency and reliability of the model in accurately estimating individual lesion areas (Fig. 3C, Supplementary Fig. S4A). For a broader ±25% range, 90.3% of lesion predictions remained within this boundary, highlighting robust accuracy across most cases, albeit with some variability. High-deviation outliers, where predictions deviated by more than +50%, represented a modest 2.3% of cases, indicating occasional overestimation or underestimation in lesion size (Fig. 3C, Supplementary Fig. S4A). Overall, the AI model performed with high accuracy at the image level. At the individual lesion level, however, variability was higher than at the image level, largely due to challenges in defining boundaries for closely clustered lesions. 
Comparative Evaluation of AI Performance on FFA Images From Vldlr−/− and Laser CNV Models
To assess potential differences in AI performance across the two models, we compared its ability to analyze FFA images from the Vldlr−/− and laser CNV models. The precision, recall, and F1 scores for both models were comparable, with the AI model performing slightly better on images from the laser CNV model (Supplementary Fig. S4B). Similarly, the IoU values for images from both models were comparable (Supplementary Figs. S4C, S4D), as were the PE value distribution ranges (Supplementary Figs. S4E, S4F). These results indicate that the performance of the AI model remained largely consistent across the Vldlr−/− and laser CNV models. 
AI Performance on Validation Images
Following the high accuracy observed during training, we next evaluated the performance of the AI model on a new set of validation images that were not included in the training dataset. The model was tested on nine validation images containing 127 manually annotated FFA lesions (Fig. 4A). The AI model identified 127 lesions, 113 of which corresponded to the manual annotations; however, the model missed 14 manually marked lesions and identified 14 additional lesions that were not part of the manual annotations (Fig. 4B). 
Figure 4.
 
Evaluation of the AI method performance in lesion detection and spatial agreement on validation images. (A) Heatmap showing the lesion count in each of the nine validation images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (E) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method. (F) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (G) Spatial agreement at the level of individual image between ground truth and the AI method quantified by IoU.
Figure 4.
 
Evaluation of the AI method performance in lesion detection and spatial agreement on validation images. (A) Heatmap showing the lesion count in each of the nine validation images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (E) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method. (F) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (G) Spatial agreement at the level of individual image between ground truth and the AI method quantified by IoU.
The AI model demonstrated strong agreement with manual lesion segmentation, achieving an ICC of 0.998 based on a two-way mixed-effects model for absolute agreement (Fig. 4C). Bland–Altman analysis revealed a bias of −5.62%, with 95% limits of agreement ranging from −27.46% to 16.22% (Fig. 4D). The model achieved an average precision, recall (sensitivity), and F1 scores of 0.90, 0.95, and 0.93, respectively, across the nine validation images (Fig. 4E), indicating high accuracy and reliability in detecting vascular lesions. Furthermore, spatial agreement of lesions with manual annotations in validation images revealed IoU values of 0.73 (95% CI, 0.71–0.75) and 0.76 (95% CI, 0.72–0.79), respectively, at individual lesion and image levels (Figs. 4F, 4G). The AI method also demonstrated strong agreement in lesion area measurements, as indicated by ICC (0.976) (Figs. 5A, 5B), Bland–Altman analysis (Figs. 5C, 5D), and PE metrics (Figs. 5E, 5F). 
Figure 5.
 
Evaluation of the AI method performance in lesion area measurement on validation images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Correlation analysis of image-wise cumulative lesion area measurement between the ground truth and AI methods, quantified by the ICC. (C) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (D) Bland–Altman plot comparing image-wise cumulative lesion area measurement between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot. (F) Distribution of percentage error in AI-predicted image-wise cumulative lesion area measurements compared to the ground truth, represented using a box-and-whisker plot.
Figure 5.
 
Evaluation of the AI method performance in lesion area measurement on validation images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Correlation analysis of image-wise cumulative lesion area measurement between the ground truth and AI methods, quantified by the ICC. (C) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (D) Bland–Altman plot comparing image-wise cumulative lesion area measurement between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot. (F) Distribution of percentage error in AI-predicted image-wise cumulative lesion area measurements compared to the ground truth, represented using a box-and-whisker plot.
Notably, these performance metrics on the validation dataset closely matched those observed during training (Figs. 13). Furthermore, the performance of the AI model was comparable to, and in some cases exceeded, the interrater agreement among three independent human operators in all of these metrics (Supplementary Figs. S5S8, Supplementary Table S1, Table). This result underscores the capability of the AI model to achieve expert-level accuracy, making it a potentially valuable tool in preclinical studies of retinal neovascular diseases. 
AI Performance on Measuring Vascular Leakage
We next evaluated the performance of the AI model in a real-world scenario by quantifying vascular leakage in Vldlr−/− mice. FFA images were captured at 5 and 9 minutes post-fluorescein administration, and lesion areas were measured and compared. The AI model successfully quantified that lesion sizes in the 9-minute cohort were significantly larger than those in the 5-minute cohort both at the level of individual lesion and total lesion area per image (Figs. 6A, 6B), consistent with results obtained through manual segmentation of the same image set. Because our AI method accurately identified and delineated the lesions, we next examined vascular leakage intensity by measuring the MFI for individual lesions, as well as the average MFI across all lesions within each image. As expected, the AI method performed comparably to the manual method in quantifying leakage at both the individual lesion level (Fig. 6C) and the whole-image level (Fig. 6D). 
Figure 6.
 
Evaluation of the AI method performance in measuring vascular leakage. (A) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (B) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (C) Manual and AI-assisted assessment of vascular leakage by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (D) Manual and AI-assisted assessment of vascular leakage. Mean fluorescence intensities across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Figure 6.
 
Evaluation of the AI method performance in measuring vascular leakage. (A) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (B) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (C) Manual and AI-assisted assessment of vascular leakage by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (D) Manual and AI-assisted assessment of vascular leakage. Mean fluorescence intensities across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Evaluation of AI Model Performance on Peripheral Lesion Detection and Quantification
All of the images used for AI training and validation were acquired with the optic nerve head centered in the field of view. To assess the versatility of the model, we tested its ability to quantify vascular lesions in peripheral regions. Despite not being specifically trained on such images, the AI model effectively detected and delineated the peripheral lesion boundaries (Fig. 7A). Additionally, our AI method also demonstrated temporal expansion of the lesion area (Figs. 7B, 7C) and leakage intensity (Figs. 7D, 7E). These findings demonstrate the versatility and robustness of our AI model in accurately quantifying vascular leakage across varying conditions, including differences in imaging time points and lesion locations. The ability of the model to generalize to peripheral lesions, despite no training data from non-central regions, highlights its potential for broader application in diverse scenarios. 
Figure 7.
 
Evaluation of the AI model performance on peripheral lesion detection and quantification. (A) Representative FFA images acquired at 5 and 9 minutes, highlighting the ability of the AI model to accurately segment lesions in peripheral regions where the optic nerve head is not centered are presented. (B) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (C) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (D) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (E) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Figure 7.
 
Evaluation of the AI model performance on peripheral lesion detection and quantification. (A) Representative FFA images acquired at 5 and 9 minutes, highlighting the ability of the AI model to accurately segment lesions in peripheral regions where the optic nerve head is not centered are presented. (B) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (C) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (D) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (E) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Discussion
Our study demonstrates the effectiveness of an AI-based analysis pipeline for quantifying vascular lesions in FFA images of mouse models. Our findings highlight the critical advantages of this approach over traditional manual methods, which are time consuming, susceptible to interobserver variability, and difficult to scale for large datasets. Compared to the methodology by Comin et al.,1 which employed log-Gabor quadrature filters and morphological operations for vascular leakage detection, our study introduces an AI-based approach trained on manually segmented FFA images. Although both methods aim for automated quantification of vascular leakage, they differ significantly in scalability, robustness, and manual intervention requirements. 
The method by Comin et al.1 relies on intensity thresholding, making it sensitive to background noise and imaging artifacts, which can compromise accuracy under diverse imaging conditions, and it necessitates frequent adjustment of thresholding parameters. In contrast, our AI model leverages learned feature recognition to adapt to complex variations in lesion morphology and fluorescence intensity, reducing susceptibility to noise and artifacts while enabling autonomous high-throughput analysis across large datasets. Additionally, although the method of Comin et al.1 was validated exclusively on the Vldlr−/− mouse model, our approach demonstrated broader applicability by being validated on multiple preclinical models, including both the Vldlr−/− and laser-induced CNV models. A key advantage of our pretrained AI model is its ability to be directly applied to analyze new FFA images without requiring further training or large annotated datasets. This capability significantly reduces the time and effort required for analysis, allowing researchers to quickly and accurately analyze their images without needing coding expertise. 
To facilitate widespread adoption of our AI method, we have provided the Segment Objects AI (.oai) file and General Analysis 3 (.ga3) file, along with step-by-step user instructions (Supplementary Files AI model 2.0.zip, Supplementary Files AI Analysis FFA Lesion area X Intensity with CSV export FINAL.zip, fundus calibration.zip, Supplementary Methods). The Segment Objects AI (.oai) file contains a pretrained AI model optimized for identifying mouse FFA vascular lesions, and the GA3 file enables automated lesion counting, area measurement, intensity analysis, and data tabulation. Researchers can easily import these files into NIS-Elements and analyze their images in a plug-and-play fashion, making advanced image analysis accessible to a wider research community. 
The AI model achieved high levels of agreement with manual annotations across key measurements such as lesion count, spatial overlap (IoU), area, and leakage intensity. Notably, the performance of the model often surpassed the interobserver variability among human annotators, indicating a high degree of reliability and reproducibility. The model also demonstrated robust generalization capabilities, maintaining consistent performance on both training and unseen validation datasets. Importantly, the AI tool effectively captured the expected temporal changes in vascular leakage in Vldlr−/− mice between 5- and 9-minute FFA captures, suggesting its suitability for real-world experimental conditions. Importantly, its performance remained consistent across both mouse models (laser CNV and Vldlr−/−), highlighting its adaptability and robustness in diverse pathological contexts. 
For a small subset of images, lesion count and area varied significantly between the AI and manual methods. These discrepancies primarily arose from inconsistencies in user-defined lesion boundaries, particularly when two or more closely located lesions had touching or partially overlapping boundaries (Supplementary Fig. S2). However, whereas individual lesion areas could show considerable variation (Fig. 3C), the total lesion area per image (Fig. 2E) remained largely consistent between the AI and manual methods, suggesting that the observed differences were primarily due to the merging of adjacent lesions rather than systematic measurement errors. 
In defining the “true” leakage area in FFA images, a challenge arises due to the diffuse and gradual nature of leakage margins. To address this, the ground truth was established based on manual annotations, where trained graders delineated the leakage boundaries by considering intensity thresholds relative to the background and anatomical context. Given the inherent ambiguity in identifying leakage margins, the AI model was trained on a dataset that accounted for human-annotated variability. Specifically, seven images were manually segmented by three independent operators, yielding 21 unique training images, as detailed in the Methods section. Furthermore, the performance of the model was evaluated by comparing its outputs to both individual grader annotations and interobserver variability, demonstrating strong alignment with human assessments. This approach underscores the robustness of the AI model in handling the challenges associated with leakage margin delineation. 
With the ability to accurately analyze hundreds of images within minutes, our model eliminates the need for time-consuming manual annotation. This enables large-scale studies that have previously been impractical. In conclusion, by streamlining FFA lesion analysis to be faster, more reliable, and easily scalable while maintaining accuracy and precision, this approach holds the potential to greatly improve research efficiency in studies of vascular pathophysiology, particularly in preclinical models of neovascular conditions such as AMD. 
Limitations and Future Directions
Our AI model was trained exclusively on images from the MICRON IV fundus imaging system. Consequently, its performance on images from other platforms remains uncertain due to potential domain shift. Variations in image characteristics (resolution, contrast, illumination, field of view) and preprocessing requirements across different cameras could impact accuracy. However, the successful quantification by the model of peripheral vascular lesions in FFA images, a task outside its training domain, suggests a degree of robust feature recognition. This indicates potential for generalization to other imaging platforms, provided that the essential vascular lesion features are preserved. Future studies incorporating diverse imaging systems, along with potential fine-tuning or domain adaptation, could help enhance the generalizability of our model across diverse imaging platforms, further strengthening its applicability to preclinical and translational research. 
Acknowledgments
The authors thank Eileen Sanz and Miriam Gibbs (The Ohio State University) for technical assistance, and we thank Anthony Vetter (Nikon Instruments) for technical assistance and insightful discussions. ChatGPT and Gemini were utilized to enhance grammar, clarity, and readability. The authors carefully reviewed and revised the text afterward and take full responsibility for the final content of the publication. 
Supported by grants from the National Institutes of Health (P30EY032857 to Ohio State University Vision Sciences Research Core Program), Ohio Lions Eye Research Foundation, Research to Prevent Blindness New Chair Challenge Grant, and a medical student research scholarship to DHG from The Ohio State University College of Medicine. 
Disclosure: V. Jayananthan, None; T.H. Taylor, None; D.H. Greentree, None; B. Collison, None; N. Kerur, None 
References
Comin CH, Tsirukis DI, Sun Y, Xu X. Quantification of retinal blood leakage in fundus fluorescein angiography in a retinal angiogenesis model. Sci Rep. 2021; 11: 19903. [CrossRef] [PubMed]
Bajwa J, Munir U, Nori A, Williams B. Artificial intelligence in healthcare: transforming the practice of medicine. Future Healthc J. 2021; 8: e188–e194. [CrossRef] [PubMed]
Joyal J-S, Sun Y, Gantner ML, et al. Retinal lipid and glucose metabolism dictates angiogenesis through the lipid sensor Ffar1. Nat Med. 2016; 22: 439–445. [CrossRef] [PubMed]
Heckenlively JR, Hawes NL, Friedlander M, et al. Mouse model of subretinal neovascularization with choroidal anastomosis. Retina. 2003; 23: 518–522. [CrossRef] [PubMed]
Lambert V, Lecomte J, Hansen S, et al. Laser-induced choroidal neovascularization model to study age-related macular degeneration in mice. Nat Protoc. 2013; 8: 2197–2211. [CrossRef] [PubMed]
Kumar A, Nagasaka Y, Jayananthan V, et al. Therapeutic targeting of telomerase ameliorates experimental choroidal neovascularization. Biochim Biophys Acta Mol Basis Dis. 2024; 1870: 167156. [CrossRef] [PubMed]
Hua J, Guerin KI, Chen J, et al. Resveratrol inhibits pathologic retinal neovascularization in Vldlr−/− mice. Invest Ophthalmol Vis Sci. 2011; 52: 2809–2816. [CrossRef] [PubMed]
Liu CH, Wang Z, Sun Y, Chen J. Animal models of ocular angiogenesis: from development to pathologies. FASEB J. 2017; 31: 4665–4681. [CrossRef] [PubMed]
Luo S, Jiang H, Wang M. C2BA-UNet: a context-coordination multi-atlas boundary-aware UNet-like method for PET/CT images based tumor segmentation. Comput Med Imaging Graph. 2023; 103: 102159. [CrossRef] [PubMed]
Pan Y, Gou F, Xiao C, Liu J, Zhou J. Semi-supervised recognition for artificial intelligence assisted pathology image diagnosis. Sci Rep. 2024; 14: 21984. [CrossRef] [PubMed]
Hosseini F, Asadi F, Rabiei R, Kiani F, Harari RE. Applications of artificial intelligence in diagnosis of uncommon cystoid macular edema using optical coherence tomography imaging: a systematic review. Surv Ophthalmol. 2024; 69: 937–944. [CrossRef] [PubMed]
Toto L, Romano A, Pavan M, et al. A deep learning approach to hard exudates detection and disorganization of retinal inner layers identification on OCT images. Sci Rep. 2024; 14: 16652. [CrossRef] [PubMed]
Do QV, Hoang HT, Vu NV, et al. Segmentation of hard exudate lesions in color fundus image using two-stage CNN-based methods. Expert Syst Appl. 2024; 241: 122742. [CrossRef]
Figure 1.
 
Evaluation of the AI method performance in lesion detection on training images. (A) Heatmap showing the lesion count in each of the 70 training images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Distribution of percentage deviation in AI-predicted lesion counts per image compared to the ground truth, represented using a box-and-whisker plot and a pie chart. (E) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (F) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method.
Figure 1.
 
Evaluation of the AI method performance in lesion detection on training images. (A) Heatmap showing the lesion count in each of the 70 training images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Distribution of percentage deviation in AI-predicted lesion counts per image compared to the ground truth, represented using a box-and-whisker plot and a pie chart. (E) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (F) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method.
Figure 2.
 
Evaluation of the AI method performance in lesion spatial agreement and area on training images. (A) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (B) Spatial agreement at the level of individual image between ground truth and AI method quantified by IoU. (C) Correlation analysis of lesion area per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion area per image between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted lesion area per image compared to the ground truth, represented using a box-and-whisker plot and frequency table.
Figure 2.
 
Evaluation of the AI method performance in lesion spatial agreement and area on training images. (A) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (B) Spatial agreement at the level of individual image between ground truth and AI method quantified by IoU. (C) Correlation analysis of lesion area per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion area per image between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted lesion area per image compared to the ground truth, represented using a box-and-whisker plot and frequency table.
Figure 3.
 
Evaluation of the AI method performance in individual lesion area measurement on training images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (C) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot and frequency table.
Figure 3.
 
Evaluation of the AI method performance in individual lesion area measurement on training images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (C) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot and frequency table.
Figure 4.
 
Evaluation of the AI method performance in lesion detection and spatial agreement on validation images. (A) Heatmap showing the lesion count in each of the nine validation images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (E) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method. (F) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (G) Spatial agreement at the level of individual image between ground truth and the AI method quantified by IoU.
Figure 4.
 
Evaluation of the AI method performance in lesion detection and spatial agreement on validation images. (A) Heatmap showing the lesion count in each of the nine validation images. (B) Venn diagram illustrating lesion detection overlap between AI and manual methods. (C) Correlation analysis of lesion counts per image between the ground truth and AI methods, quantified by the ICC. (D) Bland–Altman plot comparing lesion counts per image between the ground truth and AI methods. (E) Performance metrics, including precision, recall, and F1 score, for lesion detection by the AI method. (F) Spatial agreement at the level of individual lesion between ground truth and AI method quantified by IoU. (G) Spatial agreement at the level of individual image between ground truth and the AI method quantified by IoU.
Figure 5.
 
Evaluation of the AI method performance in lesion area measurement on validation images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Correlation analysis of image-wise cumulative lesion area measurement between the ground truth and AI methods, quantified by the ICC. (C) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (D) Bland–Altman plot comparing image-wise cumulative lesion area measurement between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot. (F) Distribution of percentage error in AI-predicted image-wise cumulative lesion area measurements compared to the ground truth, represented using a box-and-whisker plot.
Figure 5.
 
Evaluation of the AI method performance in lesion area measurement on validation images. (A) Correlation analysis of individual lesion area measurement between the ground truth and AI methods, quantified by the ICC. (B) Correlation analysis of image-wise cumulative lesion area measurement between the ground truth and AI methods, quantified by the ICC. (C) Bland–Altman plot comparing individual lesion area measurement between the ground truth and AI methods. (D) Bland–Altman plot comparing image-wise cumulative lesion area measurement between the ground truth and AI methods. (E) Distribution of percentage error in AI-predicted individual lesion area measurements compared to the ground truth, represented using a box-and-whisker plot. (F) Distribution of percentage error in AI-predicted image-wise cumulative lesion area measurements compared to the ground truth, represented using a box-and-whisker plot.
Figure 6.
 
Evaluation of the AI method performance in measuring vascular leakage. (A) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (B) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (C) Manual and AI-assisted assessment of vascular leakage by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (D) Manual and AI-assisted assessment of vascular leakage. Mean fluorescence intensities across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Figure 6.
 
Evaluation of the AI method performance in measuring vascular leakage. (A) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (B) Manual and AI-assisted assessment of vascular leakage as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (C) Manual and AI-assisted assessment of vascular leakage by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (D) Manual and AI-assisted assessment of vascular leakage. Mean fluorescence intensities across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Figure 7.
 
Evaluation of the AI model performance on peripheral lesion detection and quantification. (A) Representative FFA images acquired at 5 and 9 minutes, highlighting the ability of the AI model to accurately segment lesions in peripheral regions where the optic nerve head is not centered are presented. (B) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (C) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (D) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (E) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Figure 7.
 
Evaluation of the AI model performance on peripheral lesion detection and quantification. (A) Representative FFA images acquired at 5 and 9 minutes, highlighting the ability of the AI model to accurately segment lesions in peripheral regions where the optic nerve head is not centered are presented. (B) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Areas of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (C) AI-assisted assessment of vascular leakage in peripheral lesions as measured by expansion of lesion area. Cumulative lesion areas of individual images are plotted. P < 0.05 (two-tailed paired t-test) indicates significance. (D) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity of individual lesions. The fluorescence intensities of individual lesions are plotted. P < 0.05 (two-tailed unpaired t-test) indicates significance. (E) AI-assisted assessment of vascular leakage in peripheral lesions, by measuring the mean fluorescence intensity across all lesions within each image are plotted. P < 0.05 (two-tailed paired t-test) indicates significance.
Table.
 
Performance Metrics for the AI Method, Including Results on Training and Validation Images and Interoperator Comparisons
Table.
 
Performance Metrics for the AI Method, Including Results on Training and Validation Images and Interoperator Comparisons
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×