Open Access
Review  |   February 2024
Deep Learning and Machine Learning Algorithms for Retinal Image Analysis in Neurodegenerative Disease: Systematic Review of Datasets and Models
Author Affiliations & Notes
  • Tyler Bahr
    Mayo Clinic, Department of Ophthalmology, Rochester, MN, USA
  • Truong A. Vu
    University of the Incarnate Word, School of Osteopathic Medicine, San Antonio, TX, USA
  • Jared J. Tuttle
    University of Texas Health Science Center at San Antonio, Joe R. and Teresa Lozano Long School of Medicine, San Antonio, TX, USA
  • Raymond Iezzi
    Mayo Clinic, Department of Ophthalmology, Rochester, MN, USA
  • Correspondence: Raymond Iezzi, Mayo Clinic, Department of Ophthalmology, 200 1st St. SW, Rochester, MN 55901, USA. e-mail: iezzi.raymond@mayo.edu 
Translational Vision Science & Technology February 2024, Vol.13, 16. doi:https://doi.org/10.1167/tvst.13.2.16
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Tyler Bahr, Truong A. Vu, Jared J. Tuttle, Raymond Iezzi; Deep Learning and Machine Learning Algorithms for Retinal Image Analysis in Neurodegenerative Disease: Systematic Review of Datasets and Models. Trans. Vis. Sci. Tech. 2024;13(2):16. https://doi.org/10.1167/tvst.13.2.16.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose: Retinal images contain rich biomarker information for neurodegenerative disease. Recently, deep learning models have been used for automated neurodegenerative disease diagnosis and risk prediction using retinal images with good results.

Methods: In this review, we systematically report studies with datasets of retinal images from patients with neurodegenerative diseases, including Alzheimer's disease, Huntington's disease, Parkinson's disease, amyotrophic lateral sclerosis, and others. We also review and characterize the models in the current literature which have been used for classification, regression, or segmentation problems using retinal images in patients with neurodegenerative diseases.

Results: Our review found several existing datasets and models with various imaging modalities primarily in patients with Alzheimer's disease, with most datasets on the order of tens to a few hundred images. We found limited data available for the other neurodegenerative diseases. Although cross-sectional imaging data for Alzheimer's disease is becoming more abundant, datasets with longitudinal imaging of any disease are lacking.

Conclusions: The use of bilateral and multimodal imaging together with metadata seems to improve model performance, thus multimodal bilateral image datasets with patient metadata are needed. We identified several deep learning tools that have been useful in this context including feature extraction algorithms specifically for retinal images, retinal image preprocessing techniques, transfer learning, feature fusion, and attention mapping. Importantly, we also consider the limitations common to these models in real-world clinical applications.

Translational Relevance: This systematic review evaluates the deep learning models and retinal features relevant in the evaluation of retinal images of patients with neurodegenerative disease.

Introduction
The retina is the only neural tissue in the human body that can be directly visualized noninvasively. Findings from retinal imaging can be informative regarding the health of the brain; many abnormalities in retinal imaging have been linked with cerebral pathology.1 
In the past decade, image analysis has been revolutionized by convolutional neural networks and other promising emerging technologies for automated image analysis. These technologies undergoing further development may possibly augment the diagnostic capabilities of several ophthalmologic imaging modalities. Machine learning algorithms may assist us in detecting information in retinal images that may not be readily apparent without computational algorithms.2 Several groups have been using machine learning algorithms to determine if systemic patient health information can be gleaned from retinal images. Such algorithms have demonstrated good accuracy at predicting quantitative variables, such as coronary artery calcium or serum creatinine, and also qualitative variables, such as smoking status or biological sex using retinal fundus images alone.3 It is difficult to identify which features in the retinal images are used by the machine learning algorithms to glean this information. However, there may be more information contained in retinal images than was previously known, and it may be necessary to apply computational algorithms, such as machine learning to establish this.4 
The retina is known to exhibit many of the classic pathologic features of neurodegenerative disease. Multiple postmortem pathology studies have identified beta-amyloid plaques and neurofibrillary tangles in the retina of patients with varying stages of Alzheimer's disease (AD).58 In patients with Parkinson's disease (PD), pathology studies have found lower levels of dopamine in the retina.9,10 The neurodegeneration that accompanies Huntington's disease (HD) and amyotrophic lateral sclerosis (ALS) may appear in the retina as well.11,12 Case-control comparisons suggest that patients with mild cognitive impairment (MCI) or unspecified dementia (D-US) also exhibit retinal thinning.13,14 
Given the known correlation between retinal health and neurodegenerative disease, there is good potential that deep learning algorithms might be able to ascertain information regarding cerebral disease from retinal images.15 Indeed, a growing amount of literature has documented correlations between the progression of neurodegenerative disease and physician-observable retinal findings, such as retinal arteriolar and venular caliber, vessel tortuosity, retinal layer thickness, and optic disc morphology.16 Future research will likely focus on determining what information is contained within optical coherence tomography (OCT), OCT-angiography (OCT-A), and color fundus images.15 Such studies will also need to consider what information is not able to be obtained from retinal imaging. 
OCT, OCT-A, and fundus imaging allow for detailed quantitative and qualitative analysis of retinal features. OCT uses the reflectivity of light to micro-image the anatomy of the retina and optic disk. The peripapillary retinal nerve fiber layer (pRNFL) and macular ganglion cell layer and inner plexiform layer (mGCIPL) are especially implicated in neurodegenerative states, whereas other markers, such as macular volume and choroidal thickness, have also been studied. OCT-A works by comparing retinal layers across time as blood flows through the capillaries. OCT-A captures information regarding retinal vasculature, including microvascular density, branching complexity, and flow density. Fundus imaging allows for the direct visualization of the macula, optic disk, and retinal vasculature. Vessel tortuosity and branching complexity have been identified as helpful biomarkers and other retinal features can be directly visualized through the use of fluorescence imaging. Each imaging modality provides a host of information that has revealed retinal manifestations of neurodegenerative disease. 
Presently, it seems that current research using retinal imaging has only scratched the surface of the information which deep learning algorithms might provide, but there are also significant limitations that have yet to be addressed. Many non-neurologic diseases can have retinal manifestations indistinguishable from the features reportedly used by current machine learning models to distinguish between healthy eyes and ones with neurodegenerative disease. It will be important for future machine learning models to use more diverse datasets, based upon longitudinal data to evaluate whether machine learning models can identify specific features that differentiate true neurodegenerative disease from other diseases with neuroretinal implications. 
Methods
We conducted a systematic literature review utilizing two searching tools to identify datasets; Google Scholar and Ovid MEDLINE. Our search criteria included studies between January 1, 2012, and February 15, 2023, that contain image datasets with identifiable neurodegenerative disease and/ or studies that utilized deep learning analysis models. 
Ovid MEDLINE was used for refined searches, utilizing multi-level Boolean operators (and, or) and specific terminology (exp – explode, .mp – multi-purpose) as described below. As the final compilation of search parameters, step 11 represents final search protocol. The conceptual function of these parameters was to identify all articles with ophthalmic imaging in the context of neurodegenerative disease (steps 1–7), which also reported use of a dataset, database, or image analysis algorithm (steps 8–10). The conceptual search design is displayed in Figure 1
  • 1. optic*.mp or ophthalm*.mp or exp Retina/
  • 2. image.mp or imaging.mp or Biometry/ or Photography/
  • 3. exp Tomography, Optical Coherence/ or exp ophthalmoscopy/ or retinal photography.mp or fundoscopy.mp
  • 4. exp neurodegenerative diseases/ or exp Huntington disease/ or exp Alzheimer disease/ or exp amyotrophic lateral sclerosis/ or exp Parkinsons disease/
  • 5. 1 and 2
  • 6. 5 or 3
  • 7. 6 and 4
  • 8. exp Dataset or exp Databases, Factual/
  • 9. exp Neural Networks, Computer/ or exp Artificial Intelligence/ or exp Deep Learning/ or exp Image Interpretation, Computer-Assisted/ or exp deep learning/ or exp Diagnosis, Computer Assisted/
  • 10. 8 or 9
  • 11. 7 and 10.
Figure 1.
 
Conceptual search design as structured within Ovid MEDLINE search tool. exp, exploded, represents that the search was expanded to include similar related MeSH terms; .mp, multipurpose, indicates that all areas of the publication (title, abstract, body, etc.) were searched.
Figure 1.
 
Conceptual search design as structured within Ovid MEDLINE search tool. exp, exploded, represents that the search was expanded to include similar related MeSH terms; .mp, multipurpose, indicates that all areas of the publication (title, abstract, body, etc.) were searched.
Google search engine and Google Scholar were used for broad searches. The searched key terms were as follows: “retinal photography,” “neurodegenerative disease OCT,” “neurodegenerative disease fundoscopy,” “Parkinson's retina,” “neurodegenerative deep learning,” and “neurodegenerative image dataset.” 
Our OVID broad search resulted in 245 studies and Google resulted in 10 studies. Additionally, we referenced prior meta-analysis studies from Chrysou et al.,17 Zhou et al.,18 Jin et al.,19 Chan et al.,20 Noah et al.,13 Khan et al.,21 Nepal et al.,12 and Katsimpris et al.22 Following our inclusion criteria, each result and dataset were independently reviewed and recorded. In the setting of incomplete information within the article, datasets were assumed to be available upon request (AoR), containing two eyes per case and one image per eye. 
For the review of deep learning models, articles identified in the primary search described above were further screened according to whether they involved predictive models using multivariate data obtained from imaging, predictive models using raw images as input, or predictive models using data extracted from image feature detectors. Articles outside of this scope were excluded from the review of models presented in Section 4: “Strategies used for retinal image analysis in patients with neurodegenerative disease.” 
Results
Dataset Summary
In total, our search yielded 154 datasets containing approximately 70,481 images of 25,053 patients with neurodegenerative disease and 10,115 healthy controls. A summary of each dataset type can be found in Table 1. See Supplementary Table S1 for the comprehensive list of datasets. AD was the most represented disease (47% of datasets, 76% of patients, 66% of controls, and 74% of images), followed by PD (34% of datasets, 10% of patients, 26% of controls, and 14% of images) and MCI (20% of datasets, 4.1% of patients, 15% of controls, and 7.0% of images), whereas D-US was the least represented (2.6% of datasets, 7.2% of patients, 0.7% of controls, and 5.6% of images). No fundus image datasets were found for PD, ALS, and HD. Two datasets were accessible within the article, one was open-access, one required an account, one was unfinished, and the rest (n = 149) were classified as AoR. Most datasets utilized Heidelberg (54/154) and Zeiss (56/154) devices, whereas nine other manufacturers were also represented. The majority of datasets were from publications authored in Europe (66/154), whereas Asia (35/154), the Middle East (25/154), and North America (24/154) were also well represented, and a few datasets were generated in Oceania (2/154) and South America (3/154). 
Table 1.
 
Summary of Retinal Image Datasets in Patients With Neurodegenerative Disease
Table 1.
 
Summary of Retinal Image Datasets in Patients With Neurodegenerative Disease
Retinal Findings Common Among One or More Neurodegenerative Diseases
OCT studies have revealed similarities and differences in the retinas of patients with various neurodegenerative diseases. Table 2 contains a summary of the retinal findings organized by disease. Due to variation in results between individual studies, meta-analyses were referenced when available. In patients with AD, PD, MCI, and D-US, thinning of the pRNFL has been seen in all four quadrants, whereas the temporal and superior quadrants have been the only affected quadrants in patients with HD. Reduced macular volume, mGCIPL loss have been demonstrated in patients with AD, PD, and MCI, but not ALS or HD. Macular thinning has been identified in patients with HD, AD, PD, and MCI, but not ALS or D-US. Although the inner nuclear layer appears to be spared in patients with PD, it is reduced in patients with ALS. Choroidal thinning has been found in patients with AD, HD, and MCI, but not PD. As discussed later, these and other retinal biomarkers have been shown to correlate with disease severity and duration. 
Table 2.
 
Summary of Retinal Findings in Neurodegenerative Disease
Table 2.
 
Summary of Retinal Findings in Neurodegenerative Disease
OCT-A reveals additional information regarding the retinal vasculature of patients with neurodegenerative disease. Decreased microvascular density and an enlarged foveal avascular zone have been demonstrated in patients with AD and PD. Reduced branching complexity has also been associated with AD. Preliminary OCT-A case-control studies of patients with ALS, HD, or MCI have not found conclusive significant retinal findings. Fundus imaging case-control studies have been limited to only AD. 
Alzheimer's Disease
OCT studies have revealed significant retinal neurodegeneration in patients with AD. A 2018 meta-analysis20 found thinning in the pRNFL, mGCIPL, ganglion cell complex, and choroidal layers, as well as reduced overall macular volume and macular thinning in the inner and outer sectors. The mGCIPL thinning has also been shown to correlate with disease severity.14 
Various fluorescent fundus imaging modalities have been used to visualize and quantify retinal AD pathology. Intravenous administration of curcumin, a beta-amyloid-binding fluorophore, revealed that the retinal beta-amyloid burden is doubled in patients with AD. Retinal beta-amyloid levels were linked to cortical beta-amyloid burden and reduced hippocampal volume.23,24 Alternatively, blue autofluorescence has been used to quantify the surface area of retinal inclusion bodies which correlates with preclinical cortical beta-amyloid burden.25 Finally, analysis of fluorescence lifetime imaging ophthalmoscopy revealed differences in patients with phakic AD compared to matched controls.26 
Changes in retinal vasculature have been identified in fundus images of patients with AD. Fractal dimension (FD), a quantitative representation of vascular branching complexity, can be determined by commercially available software or expert analysis. A systematic review in 2019 found that vascular FD was decreased in four case-control fundus imaging studies of AD.27 In addition to reduced FD, one study found increased vessel tortuosity and narrowed venular caliber in patients with AD, although a separate study yielded contradictory findings.28,29 
Advancements in OCT-A have revealed further information regarding vascular changes in patients with AD. A 2021 meta-analysis demonstrated an enlarged foveal avascular zone and reduced macular whole enface superficial and deep vessel densities in patients with AD.19 Notably, features of FD, vessel caliber, and vessel tortuosity that have been found on fundus imaging can also be evaluated using OCT-A, although systematic differences between modalities have been noted.30 
Parkinson's Disease
In addition to reduced dopamine levels, the retina of patients with PD also exhibits neurodegeneration. Meta-analyses of OCT studies in 201917 and 202018 revealed reduced pRNFL, mGCIPL, and macular thickness, as well as decreased macular volume in patients with PD. In addition, disease severity and duration have been shown to be linked with pRNFL thinning and decreased foveal thickness.31,32 Contradictory findings of both increased and decreased choroidal thickness have been reported, likely due to differences in image analysis.33,34 Fundus imaging studies have revealed that patients with retinal thinning compared to age-matched controls have an increased risk of developing PD.35 
PD has also demonstrated an impact on retinal vasculature. A 2023 meta-analysis22 found that patients with PD had reduced microvascular density in the whole superficial vascular plexus (SVP), foveal SVP, parafoveal SVP, and foveal avascular zone (FAZ), and reduced branching complexity has also been implicated.36 Microvascular density, FD, and retinal flow density have each been used to successfully differentiate patients with and without PD.3739 
Amyotrophic Lateral Sclerosis
Studies on ALS have produced conflicting results regarding retinal neurodegeneration. A 2022 meta-analysis12 found inner nuclear layer thickness to be the only statistically significant finding in patients with ALS. However, 6 of the 11 studies within the analysis also demonstrated significant pRNFL thinning. Future studies with differentiation between ALS subtypes (bulbar-onset versus spinal-onset) may help clarify the conflicting results.12 In the only found OCT-A study assessing patients with ALS, retinal microvessel density was not significantly decreased in patients with ALS compared to controls.40 
Huntington's Disease
OCT case-control studies of patients with HD have produced conflicting results. Temporal pRNFL deterioration and reduced macular retinal and choroidal thickness have been demonstrated in patients with HD, but other studies found no significant difference.4148 One study found that temporal pRNFL thinning appeared in preclinical HD, whereas another could not replicate the finding.44,48 Disease duration and severity may be correlated with temporal pRNFL thinning and reduced macular volume.42 Two studies using OCT-A found no difference in vessel density.41,46 A total of 115 patients with HD have been studied, suggesting that larger sample sizes may assist in ascertaining the retinal characteristics of HD. 
Unspecified Dementia and Mild Cognitive Impairment
Patients with MCI have demonstrated retinal changes. A 2020 meta-analysis13 discovered pRNFL thinning across 17 studies. OCT case-control comparisons have also demonstrated decreased mGCIPL, ganglion cell complex, macular, foveal, and choroidal thickness, and reduced macular volume, although all studies did not share the same findings. GCIPL thickness,49 choroidal thickness,50 pRNFL thickness,5153 and the pRNFL granular membrane area54 have been shown to be inversely correlated with cognitive performance, whereas other studies show no correlation.55 Decreased mGCIPL thickness has also been correlated with a reduction of white matter in the fornix of patients with MCI.56 An OCT-A study discovered reduced microvascular density in the superior-nasal region of patients with MCI, but another study did not find any differences.57,58 
Few studies have investigated the relationship between the retina and unspecified dementia. Ferrari et al.14 found pRNFL thinning and GCIPL loss, with GCIPL loss being connected with AD severity. On the other hand, Pillai et al.59 found no statistical significance in RNFL thickness, GCIPL thickness, or macular volume between patients with unspecified dementia and healthy controls. Further studies will help explore the impact of unspecified dementia on the retina. 
Strategies Used for Retinal Image Analysis in Patients With Neurodegenerative Disease
As discussed above, recent studies have elucidated various features found in retinal imaging that are associated with neurodegenerative diseases. On the other hand, several studies have used machine learning models to detect or otherwise learn more about these diseases using retinal images alone, without any a priori knowledge. Table 3 provides an overview of studies using deep learning algorithms and other predictive models for retinal image analysis in neurodegenerative disease, whereas Supplementary Table S2 contains additional details for each model. Deep learning models can detect features from unstructured data to make predictions using that data, with no guidance apart from data examples. The most prevalent deep learning algorithm for image analysis is known as the convolution neural network (CNN). In retinal images of patients with neurodegenerative disease, some of the features learned by CNN models might be features already described in the scientific literature. Some may be observable features that have not yet been described, and others may be features that are even too subtle for a human observer to detect. CNNs have shown great promise for automating decision-making tasks using retinal images in patients with neurodegenerative disease. 
Table 3.
 
Deep Learning Algorithms and Other Predictive Models for Retinal Image Analysis in Neurodegenerative Disease
Table 3.
 
Deep Learning Algorithms and Other Predictive Models for Retinal Image Analysis in Neurodegenerative Disease
An excellent review recently detailed the computational strategies for using CNNs for retinal image analysis in general, which are applicable to the current discussion. The authors outlined the following stepwise elements in the overall framework for implementing a deep learning model for retinal image analysis. These include: (1) image acquisition and annotation, (2) retinal image preprocessing, (3) model architecture and design, (4) model training, (5) generating model predictions, and (6) evaluating model performance.60 For a more detailed explanation of these stages, we refer the reader to this other review.60 In the current work, we will discuss specific architectural elements as they pertain to the problem of retinal image analysis for neurodegenerative disease, and their past and future applications. 
CNN Models: Basic Architectures and Applications
The architecture of a neural network refers to the arrangement of computational steps, which are known as layers. CNNs share a basic unit, the convolution layer, which is used to extract information from images. Computations in convolutional layers are fed forward in series to other types of computational layers including pooling and fully connected (FC) layers, and also other downstream convolution layers. The most common types of computational tasks performed by CNNs are classification, regression, and segmentation. The architecture of VGGNet, a common CNN used for retinal image analysis, is shown in Figure 2
Figure 2.
 
Architecture of VGGNet. Used with permission from Goutam et al. 2022.60 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Figure 2.
 
Architecture of VGGNet. Used with permission from Goutam et al. 2022.60 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Classification
The structure of a basic CNN for the task of image classification usually involves several repeated units which consist of one or more convolution layers followed by a pooling layer. With each iteration of units, the resolution of the output decreases. Feed-forward skip connections are used throughout to preserve important spatial data from earlier steps. Finally, an FC layer is used to transform the spatial image data into a structured classification output. Classification models have been the most common type of CNN among publications on retinal image analysis in the context of neurodegeneration. Accuracy and area under the receiver operator curve (AUC) are the most popular methods for measuring the performance of these models, but binary accuracy, sensitivity, and specificity are also used. 
Disease detection is the most basic classification task; many different models have been constructed to take images of a fundus as input and provide as output a binary label distinguishing whether the image is from a patient with a neurodegenerative disease or from a healthy control. Attempts at detecting AD, PD, and general cognitive impairment using fundus photographs have been described previously, with varied results. CNN models trained on large fundus datasets have demonstrated improved performance compared to models trained on small datasets. For example, a model trained on nearly 13,000 images had an AUC of 0.91,61 whereas a model trained on less than 300 images had an AUC of 0.83.62 Ongoing work is being done to engineer these models to improve their diagnostic accuracy. Multi-class disease detection models have also been constructed. These aim to detect one of several potential disease states; for example, one study used support vector machines (a different type of machine learning model) to classify patients as having AD, PD, or neither.63 Classification is not limited to the diagnosis of disease and may involve multiple classes representing any variable of interest. For example, one study devised a method for classifying the localization of age-related white matter changes to one of six overall brain regions using bilateral fundus photos fed to a CNN, followed by a regression and decision tree for brain region classification.64 This approach was able to classify the location of white matter lesions into one of 6 potential regions: left and right frontal lobes, parietal–occipital lobes, or basal ganglia using fundus photographs, with an AUC of 0.955 based on 10-fold cross-validation. 
Risk stratification is another classification task which involves labeling imaging as belonging to one or two groups of varying risk. Many such studies have attempted to predict the presence of a known risk marker. For example, one model was trained to detect the ApoE4 genotype using fundus photographs. However, this model was unsuccessful with an AUC of 0.47, which may have been due to the low representation of ApoE4 individuals in the dataset, the model structure itself (although the model had good performance on age and sex predictions), or alternatively the inability of fundus photography as a modality to capture information about retinal amyloid deposits.65 
Regression
A very similar type of CNN model can perform a regression task using image inputs. Regression models have the same general framework as classification models, except for the final computational layer. In a regression model, the final layer outputs a numeric variable (rather than categorical variables used in classification). The correlation value R is most commonly used to describe what percent of the variability in the dependent variable can be explained by the model. 
For instance, one model attempted to predict cognitive scores in a cohort of aging adults, although the model was only able to explain 22% of cognitive scores (R = 0.22).65 Another study with a similar objective of predicting the Cardiovascular Risk Factors, Aging, and Incidence of Dementia (CAIDE) yielded a more useful model, with R = 0.76 on an external validation set and an AUC of 0.926 for the detection of high dementia risk defined as a score >10.66 One of the more striking successful examples of regression using CNNs is the estimation of a person's age using a fundus photograph, a biomarker that is known as retinal age, with R = 0.81 in one study.67,68 Other potential applications of regression are hazard and time-risk models. For instance, one model used the retinal age gap (the difference between retinal age and true age) and other demographic data as input into a Cox proportional hazards model to estimate the 5-year incidence of PD.35 
Segmentation and Object Detection
Image segmentation is a different kind of task that uses a very different model architecture. The purpose of a segmentation algorithm is to use an input image to create a segmentation map, which is an image highlighting every pixel from the input image that is associated with a structure of interest. Similar to other CNN models, segmentation models use blocks of convolutional and pooling layers. In a segmentation model, intermediate convolution layers may scale down the resolution of the image data, but up-convolution layers are used to increase the output resolution (back to the original resolution in many cases). Feed-forward skip connections are used to preserve important spatial data from the higher-resolution steps to create the segmentation map. 
Segmentation of the retinal vessels is one of the most important examples of this type of algorithm, as the vascular anatomy contains valuable information regarding the neurodegenerative disease, as noted above. However, segmentation of retinal vessels has also been described as one of the most challenging tasks in retinal image processing.69 Several segmentation algorithms have been developed over the past few years with the objective of generating images of the retinal vascular tree from fundus images, with all other details removed from the image. A vascular segmentation algorithm for OCT-A images has also been developed recently.70 In practice, these models can be used to extract from the original image an image of white vessels only on a black background (or vice-versa). Many studies have used pre-existing vessel segmentation algorithms to process input images and use the resulting vessel map alongside the original image as a second input to a classification or regression model. A standardized method has recently been described for evaluating and comparing the accuracy of vessel maps. Using this method, a per-pixel (vessel or no vessel present in this pixel) AUC can be calculated by averaging over every pixel in an entire image set.70 
Object detection shares some similarities with segmentation but also has some elements of a classification model. These algorithms can detect and quantify one or more features in a fundus photograph. For example, the automated retinal image analysis (ARIA) algorithm has been used to detect and count arteriole-venous nicking, arteriole occlusions, hemorrhages, and exudates,64 and these results can be used as input features in a model. 
Architectural Modifications and Other Tools for Increasing Model Performance
Presently, most studies using retinal images of patients with neurodegenerative disease have been relatively small in size (in the tens to hundreds of data examples) compared with datasets more commonly used for commercial deep learning applications (often in the thousands to over a million data examples). Ample large-scale datasets exist for diseases such as diabetic retinopathy or glaucoma, but there is a scarcity of available retinal image data for patients with neurodegenerative disease. There are some moderately sized datasets for AD and several small datasets that could potentially be pooled. However, sizable retinal image datasets from patients with other common neurodegenerative diseases with lower prevalence like ALS, HD, or frontotemporal dementia are greatly lacking. 
The term “Hand-engineering” refers to the use of a priori knowledge to intentionally construct a model with specific elements to increase its likelihood of recognizing features that are already known to be important in the task to be automated. Hand engineering a model by incorporating various deep learning tools into the model architecture can increase the predictive power of CNN models, especially for small datasets, although these methods can also increase the accuracy of models trained on large datasets as well. 
In this section, we discuss a few of the most important deep learning tools currently available to allow for the harnessing of a priori knowledge, or otherwise increase the power of models used to analyze retinal images from patients with neurodegenerative disease. 
Image Preprocessing and Feature Extraction
One of the simplest means of hand engineering can be preprocessing data to enhance specific features. Different combinations of image color, contrast, noise, or sharpness modifications have proven helpful for improving the visualization of darker vessels, the brighter optic nerve head, or retinal background lesions, for example.60 In small datasets in which images from only one eye are input at a time, it can be helpful to horizontally flip all images of left eyes to match the right eyes to increase dataset homogeneity. 
Increasing the size of a dataset, known as data augmentation, can also improve model performance. For image datasets, this is commonly done using image transformations. The addition of randomly flipped, rotated, cropped, zoomed, or blurred copies of existing images in the dataset can improve model performance. Non-random cropping can also be useful, to focus on certain regions of interest, such as the optic nerve head or macula.60 
Pre-existing deep learning algorithms can also be used to transform an image during preprocessing. Several powerful retinal vessel segmentation algorithms have been developed recently. These tools have the ability to generate segmentation maps isolating the arteries, veins, or both.69 Other segmentation algorithms can detect and create labeled maps of pathologic features including hemorrhages, microaneurysms, exudates, or retinal neovascularization.71 These algorithms were specifically developed with the small-dataset problem in mind. Even a small-scale model can obtain key lesion information reliably if the inputted images are preprocessed by a pre-existing model that can detect these lesions accurately. 
Segmentation algorithms can also be used in the preprocessing of OCT images to identify the various retinal layers or create maps of their spatial thickness distribution. For example, segmentation of OCT layers can be used to generate several 2D maps of the thickness of each retinal layer, which can be used as input into a CNN. This is important feature information, given the known association between neurodegenerative disease and retinal layer thinning. 
Feature extraction is the practice of using computations on original input data, which may include pre-existing neural network models, but also simpler calculations, to extract desired data features. This can be done with pretrained CNNs specialized for extracting a specific type of feature of interest (i.e. vascular information). For example, retina-based microvascular health assessment system (RMHAS) is an algorithm that can extract the following features from a retinal image: average vessel diameter, average vessel length, fractal dimension, branching angles, tortuosity, branching coefficient, asymmetry ratio, junctional exponent deviation, and angular asymmetry.72 Several other models are available for extracting various quantitative features from retinal images. The extracted feature data can then be used as input to a model. Occasionally, when features of interest are unknown, feature extraction can be done using the pretrained weights from early layers of a general image classifier (such as ImageNet), although the outputs from this form of feature extraction are difficult to correlate with any physiological characteristics. 
One potential pitfall of both preprocessing and feature extraction is the loss of data due to the transformation of the original image data. For example, cropping to focus on the optic disk would result in a loss of data for the entire rest of the fundus. Likewise, inputting only segmented images of the vascular tree would neglect the rest of the retinal background. In addition, by the same token, inputting only structured data obtained from feature extraction would neglect any other features potentially contained in the image. For larger datasets, it is usually best to use the original images as parallel inputs together with the preprocessed images to avoid the loss of features within the original image, although for smaller datasets this may not negatively impact model performance. The methods for merging parallel inputs, known as feature fusion, will be discussed in another section below. 
Transfer Learning
In the previous section, we discussed how the use of pre-existing models can be a powerful tool when hand-engineering a model. This is also true for training the model itself. The task of basic recognition of curves, outlines, and shapes is not trivial and must be learned by a naive model. Rather than training a naive model, the more common practice is to begin with a pretrained network, load its learned weights, and then fine-tune the weights of a few final layers during the specialized training for the specific task at hand. Importantly, the use of transfer learning was shown to yield better results compared with naive model training for predicting systemic information using retinal images, even for naive models trained on very large datasets.73 Examples of common pretrained CNN models include ResNet (Microsoft), VGGNet (Oxford), ImageNet (ImageNet), EfficientNet (Google), and GoogLeNet/ Inception (Google). 
Using pretrained networks can be problematic, however. Pretrained CNNs tend to favor features that were important for classifying images from their original training datasets. For example, many pretrained CNNs are known to favor texture features over shape features.74 Although texture, or local patterns distributed over a small area, seems to be very important for neurodegenerative disease classification,75 it may be a pitfall to neglect key shape information, which is associated with objects across a larger area in the image. One solution to this problem is the use of pretrained networks that prioritize shape, such as Stylized-ImageNet, which used different training data to intentionally emphasize shape features over texture. 
Feature Fusion
Feature fusion is the merging of features from different branches or different layers in a model so that the model will consider them together in one or more computational steps. Commonly, this is accomplished by simple concatenation (appending) or summation (adding). Feature fusion can be used to combine parallel branches (i.e. fuse patient metadata with a fundus image) or within elements (i.e. between convolution layers downstream of a single input image). 
One important application of feature fusion to the current discussion is the simultaneous consideration of bilateral fundus images. It seems particularly important to consider data from both eyes when using fundus images to predict extraocular disease states, such as neurodegenerative disease. Several recent studies have described various strategies of feature fusion for the analysis of bilateral fundus images. Recently, a powerful model using feature fusion of bilateral fundus images achieved >99% precision and >99% recall classifying 8 different disease classes, including systemic diseases like diabetes and hypertension using retinal images from the ODIR dataset (n = 5000).76 
Another possibility with feature fusion is the parallel use of metadata which may include demographic information, like sex, age, ApoE genotype, cardiovascular laboratory markers, or potentially any information from the medical history. These data seem especially important for giving systemic contextual information and has proven useful in several models. For example, the use of metadata fused with imaging by simple concatenation improved the accuracy of skin cancer classification using dermoscopy images (see Figure 3).77 Similarly, the detection of anemia using fundus imaging together with metadata was significantly more accurate compared to either alone.78 
Figure 3.
 
Simple approach for combining fundoscopic image processing and metadata processing. Adapted with permission from Gessert et Al. 2020.77 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Figure 3.
 
Simple approach for combining fundoscopic image processing and metadata processing. Adapted with permission from Gessert et Al. 2020.77 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Attention Mechanism for CNNs
The attention mechanism is a deep learning tool that was first developed in language models, giving them the ability to maintain attention to key parts of a sentence that were closely linked either grammatically or semantically despite being distant from one another (many words apart within a sentence or paragraph). For CNNs, the mechanism of spatial attention is the ability to give more weight to certain regions of the input image and less to others, despite being distant from one another, and the attention function is learnable by the model during training rather than hard-coded. 
A new method for incorporating attention into feature fusion has been described recently.79 Attentional feature fusion allows a model the ability to preferentially give attention to different features rather than giving equal attention to different features, which is the result of simple methods like concatenation and summation. 
Class Activation Mapping: Shedding Light on Influential Features
A major drawback to the use of deep neural networks for image analysis has been their black-box nature, or, in other words, their inability to explain their decision-making processes. Class activation mapping (CAM) is a method of highlighting the key features either in the original image or in any of the convolutional layers that are most influential in a model's decision and can be used for classification and segmentation tasks. This is a method of inspecting a convolution layer's implicit attention by displaying a map of the relative weights for each pixel. Self-Attention class activation maps SA-CAM80 have recently been described as a new method by which spatial attention maps can be visualized, which is useful for displaying explicit rather than implicit model attention. 
The benefit of using CAM is that it may help find meaningful clinical correlations of model behavior. In diabetic retinopathy models, attention maps have shown model attention to lesions known to be important for Early Treatment Diabetic Retinopathy Study (ETDRS) classification of diabetic retinopathy, such as microaneurysms and hemorrhages (Figure 4).81 There is currently no known classification system for neurodegenerative disease based on retinal image findings. However, CAM in a model for the detection of symptomatic AD showed that the model gave attention to areas of decreased vessel density in the fovea and temporal macula in images classified as positive for AD.62 Making sense of the features used by CNNs in their decision making will be important for linking model predictions with physiologic features. 
Figure 4.
 
Example of attention maps used in the classification of diabetic retinopathy. Pixels in the image that are of higher relevance to model decision making are highlighted in yellow and red. Adapted with permission from Zhang et al. 2022.81
Figure 4.
 
Example of attention maps used in the classification of diabetic retinopathy. Pixels in the image that are of higher relevance to model decision making are highlighted in yellow and red. Adapted with permission from Zhang et al. 2022.81
Discussion
In the recent literature, a variety of studies have demonstrated that retinal image analysis using deep learning can provide highly useful predictions regarding neurodegenerative disease diagnosis and risk assessment with good accuracy. Using retinal fundus photographs and/or OCT images, convolutional neural network algorithms have performed impressive tasks such as automated diagnosis of AD and PD, calculation of risk of future PD or dementia, and localization of cerebral white matter disease. 
Recent advances in deep learning should be considered in the development of future models. For retinal image analysis in general, transfer learning, or the use of a convolutional neural network with pretrained weights, seems to be superior to training a model with naive weights, even when a large fundus image dataset is available for training. New advances, such as attentional feature fusion, will also allow for improved ability for researchers to understand model decision making. Ongoing development of feature extraction algorithms is enabling the generation of improved maps of lesions and vasculature, as well as the generation of structured feature information such as retinal age or whole-fundus vessel caliber. However, the use of extracted feature information together with original images yields better results than extracted features alone. 
Because the performance and generalizability of convolutional neural network models tend to improve as more images are used during training, the predictive power of future models is expected to increase as more retinal image data becomes available in patients with neurodegenerative disease. Several existing datasets are available and contain various imaging modalities primarily in patients with AD, although dataset size for most of these is on the order of tens to a few hundred. There is very little image data available for PD, HD, ALS, or other neurodegenerative diseases. At present, although cross-sectional imaging data is becoming more abundant, datasets with longitudinal imaging are altogether lacking. Furthermore, the use of bilateral and multimodal imaging together with metadata seems to improve model performance, thus larger multimodal datasets with patient metadata and bilateral images are also needed. 
Thus far, a broad range of device manufacturers have been represented in datasets. It could certainly be a pitfall to deploy a machine learning model to interpret images acquired using devices different from those used for model training. This could lead to unpredictable model behavior. However, using images from different devices and manufacturers together in the same training set would not be expected to detract from the rigor of a model as long as the images in the test dataset were acquired using one of the devices used for training. In fact, a model trained using images from several different devices or manufacturing systems could have better generalizability than a model trained on images from a single device. 
Although deep learning can potentially be a powerful tool for neurodegenerative disease, virtually all models in the current literature share a similar limitation. Features such as retinal thinning are highly nonspecific and could represent a variety of pathologies, such as glaucoma, diabetes, or other inflammatory retinopathies. In addition, whereas many reports have claimed to detect neurodegenerative diseases with high specificity and sensitivity, most of these datasets are poorly representative of a realistic clinical population. It will be important for future models to use more diverse datasets that do not exclude disease with ocular manifestations and re-evaluate whether these models can identify features differentiating true neurodegenerative disease from other diseases with retinal implications. Given these limitations, it remains uncertain whether automated retinal image analysis using machine learning algorithms will be useful for the diagnosis of neurodegenerative disease in clinical practice. 
Acknowledgments
Disclosure: T. Bahr, None; T.A. Vu, None; J.J. Tuttle, None; R. Iezzi, None 
References
Richardson A, Robbins CB, Wisely CE, Henao R, Grewal DS, Fekrat S. Artificial intelligence in dementia. Curr Opin Ophthalmol. 2022; 33(5): 425–431, doi:10.1097/ICU.0000000000000881. [CrossRef] [PubMed]
Ranchod TM. Systemic retinal biomarkers. Curr Opin Ophthalmol. 2021; 32(5): 439–444, doi:10.1097/ICU.0000000000000784. [CrossRef] [PubMed]
Wagner SK, Fu DJ, Faes L, et al. Insights into systemic disease through retinal imaging-based oculomics. Transl Vis Sci Technol. 9(2): 6, doi:10.1167/tvst.9.2.6. [CrossRef] [PubMed]
Patil AD, Biousse V, Newman NJ. Artificial intelligence in ophthalmology: an insight into neurodegenerative disease. Curr Opin Ophthalmol. 2022; 33(5): 432–439, doi:10.1097/ICU.0000000000000877. [CrossRef] [PubMed]
Koronyo Y, Biggs D, Barron E, et al. Retinal amyloid pathology and proof-of-concept imaging trial in Alzheimer's disease. JCI Insight. 2(16): e93621, doi:10.1172/jci.insight.93621. [CrossRef] [PubMed]
Koronyo-Hamaoui M, Koronyo Y, Ljubimov AV, et al. Identification of amyloid plaques in retinas from Alzheimer's patients and noninvasive in vivo optical imaging of retinal plaques in a mouse model. NeuroImage. 2011; 54(Suppl 1): S204–S217, doi:10.1016/j.neuroimage.2010.06.020. [PubMed]
Shi H, Koronyo Y, Rentsendorj A, et al. Identification of early pericyte loss and vascular amyloidosis in Alzheimer's disease retina. Acta Neuropathol (Berl). 2020; 139(5): 813–836, doi:10.1007/s00401-020-02134-w. [CrossRef]
La Morgia C, Ross-Cisneros FN, Koronyo Y, et al. Melanopsin retinal ganglion cell loss in Alzheimer disease. Ann Neurol. 2016; 79(1): 90–109, doi:10.1002/ana.24548. [CrossRef] [PubMed]
Harnois C, Di Paolo T. Decreased dopamine in the retinas of patients with Parkinson's disease. Invest Ophthalmol Vis Sci. 1990; 31(11): 2473–2475. [PubMed]
Djamgoz MB, Hankins MW, Hirano J, Archer SN. Neurobiology of retinal dopamine in relation to degenerative states of the tissue. Vision Res. 1997; 37(24): 3509–3529, doi:10.1016/S0042-6989(97)00129-6. [CrossRef] [PubMed]
Kersten HM, Danesh-meyer HV, Kilfoyle DH, Roxburgh RH. Optical coherence tomography findings in Huntington's disease: a potential biomarker of disease progression. J Neurol. 2015; 262(11): 2457–2465, doi:10.1007/s00415-015-7869-2. [CrossRef] [PubMed]
Nepal G, Kharel S, Coghlan MA, et al. Amyotrophic lateral sclerosis and retinal changes in optical coherence tomography: a systematic review and meta-analysis. Brain Behav. 2022; 12(9): e2741, doi:10.1002/brb3.2741. [CrossRef] [PubMed]
Noah AM, Almghairbi D, Moppett IK. Optical coherence tomography in mild cognitive impairment – systematic review and meta-analysis. Clin Neurol Neurosurg. 2020; 196: 106036, doi:10.1016/j.clineuro.2020.106036. [CrossRef] [PubMed]
Ferrari L, Huang SC, Magnani G, Ambrosi A, Comi G, Leocani L. Optical coherence tomography reveals retinal neuroaxonal thinning in frontotemporal dementia as in Alzheimer's disease. J Alzheimers JAD. 2017; 56(3): 1101–1107, doi:10.3233/JAD-160886. [CrossRef] [PubMed]
Yuan A, Lee CS. Retinal biomarkers for Alzheimer disease: the facts and the future. Asia-Pac J Ophthalmol. 2022; 11(2): 140–148, doi:10.1097/APO.0000000000000505. [CrossRef]
Vujosevic S, Parra MM, Hartnett ME, et al. Optical coherence tomography as retinal imaging biomarker of neuroinflammation/neurodegeneration in systemic disorders in adults and children. Eye. 2023; 37: 203–219, doi:10.1038/s41433-022-02056-9. [CrossRef] [PubMed]
Chrysou A, Jansonius NM, van Laar T. Retinal layers in Parkinson's disease: a meta-analysis of spectral-domain optical coherence tomography studies. Parkinsonism Relat Disord. 2019; 64: 40–49, doi:10.1016/j.parkreldis.2019.04.023. [CrossRef] [PubMed]
Zhou WC, Tao JX, Li J. Optical coherence tomography measurements as potential imaging biomarkers for Parkinson's disease: a systematic review and meta-analysis. Eur J Neurol. 2021; 28(3): 763–774, doi:10.1111/ene.14613. [CrossRef] [PubMed]
Jin Q, Lei Y, Wang R, Wu H, Ji K, Ling L. A systematic review and meta-analysis of retinal microvascular features in Alzheimer's disease. Front Aging Neurosci. 2021; 13: 683824, doi:10.3389/fnagi.2021.683824. [CrossRef] [PubMed]
Chan VTT, Sun Z, Tang S, et al. Spectral domain-optical coherence tomography measurements in Alzheimer's disease: a systematic review and meta-analysis. Ophthalmology. 2019; 126(4): 497–510, doi:10.1016/j.ophtha.2018.08.009. [CrossRef] [PubMed]
Khan SM, Liu X, Nath S, et al. A global review of publicly available datasets for ophthalmological imaging: barriers to access, usability, and generalisability. Lancet Digit Health. 2021; 3(1): e51–e66, doi:10.1016/S2589-7500(20)30240-5. [CrossRef] [PubMed]
Katsimpris A, Papadopoulos I, Voulgari N, et al. Optical coherence tomography angiography in Parkinson's disease: a systematic review and meta-analysis. Eye. 2023; 37: 2847–2854, doi:10.1038/s41433-023-02438-7. [CrossRef] [PubMed]
Koronyo Y, Biggs D, Barron E, et al. Retinal amyloid pathology and proof-of-concept imaging trial in Alzheimer's disease. JCI Insight. 2017; 2(16): e93621, 93621, doi:10.1172/jci.insight.93621. [CrossRef] [PubMed]
Dumitrascu OM, Lyden PD, Torbati T, et al. Sectoral segmentation of retinal amyloid imaging in subjects with cognitive decline. Alzheimers Dement Amst Neth. 2020; 12(1): e12109, doi:10.1002/dad2.12109.
Snyder PJ, Johnson LN, Lim YY, et al. Nonvascular retinal imaging markers of preclinical Alzheimer's disease. Alzheimers Dement Diagn Assess Dis Monit. 2016; 4: 169–178, doi:10.1016/j.dadm.2016.09.001.
Sadda SR, Borrelli E, Fan W, et al. A pilot study of fluorescence lifetime imaging ophthalmoscopy in preclinical Alzheimer's disease. Eye Lond Engl. 2019; 33(8): 1271–1279, doi:10.1038/s41433-019-0406-2.
Lemmens S, Devulder A, Van Keer K, Bierkens J, De Boever P, Stalmans I. Systematic review on fractal dimension of the retinal vasculature in neurodegeneration and stroke: assessment of a potential biomarker. Front Neurosci. 2020; 14: 16. Accessed January 31, 2023, doi:10.3389/fnins.2020.00016. [CrossRef] [PubMed]
Cheung CYL, Ong YT, Ikram MK, et al. Microvascular network alterations in the retina of patients with Alzheimer's disease. Alzheimers Dement J Alzheimers Assoc. 2014; 10(2): 135–142, doi:10.1016/j.jalz.2013.06.009. [CrossRef]
Williams MA, McGowan AJ, Cardwell CR, et al. Retinal microvascular network attenuation in Alzheimer's disease. Alzheimers Dement Amst Neth. 2015; 1(2): 229–235, doi:10.1016/j.dadm.2015.04.001. [CrossRef]
Falavarjani KG, Al-Sheikh M, Akil H, Sadda SR. Image artefacts in swept-source optical coherence tomography angiography. Br J Ophthalmol. 2017; 101(5): 564–568, doi:10.1136/bjophthalmol-2016-309104. [CrossRef] [PubMed]
Jiménez B, Ascaso FJ, Cristóbal JA, López del Val J. Development of a prediction formula of Parkinson disease severity by optical coherence tomography. Mov Disord. 2014; 29(1): 68–74, doi:10.1002/mds.25747. [CrossRef] [PubMed]
Garcia-Martin E, Rodriguez-Mena D, Satue M, et al. Electrophysiology and optical coherence tomography to evaluate Parkinson disease severity. Invest Ophthalmol Vis Sci. 2014; 55(2): 696–705, doi:10.1167/iovs.13-13062. [CrossRef] [PubMed]
Satue M, Obis J, Alarcia R, et al. Retinal and choroidal changes in patients with Parkinson's disease detected by swept-source optical coherence tomography. Curr Eye Res. 2018; 43(1): 109–115, doi:10.1080/02713683.2017.1370116. [CrossRef] [PubMed]
Eraslan M, Cerman E, Yildiz Balci S, et al. The choroid and lamina cribrosa is affected in patients with Parkinson's disease: enhanced depth imaging optical coherence tomography study. Acta Ophthalmol (Copenh). 2016; 94(1): e68–75, doi:10.1111/aos.12809. [CrossRef]
Hu W, Wang W, Wang Y, et al. Retinal age gap as a predictive biomarker of future risk of Parkinson's disease. Age Ageing. 2022; 51(3): afac062, doi:10.1093/ageing/afac062. [CrossRef] [PubMed]
Shi C, Chen Y, Kwapong WR, et al. Characterization by fractal dimension analysis of the retinal capillary network in Parkinson disease. Retina Phila Pa. 2020; 40(8): 1483–1491, doi:10.1097/IAE.0000000000002641. [CrossRef]
Xu B, Wang X, Guo J, et al. Retinal microvascular density was associated with the clinical progression of Parkinson's disease. Front Aging Neurosci. 2022; 14: 818597, doi:10.3389/fnagi.2022.818597. [CrossRef] [PubMed]
Zhang Y, Zhang D, Gao Y, et al. Retinal flow density changes in early-stage Parkinson's disease investigated by swept-source optical coherence tomography angiography. Curr Eye Res. 2021; 46(12): 1886–1891, doi:10.1080/02713683.2021.1933054. [CrossRef] [PubMed]
Zhou M, Wu L, Hu Q, et al. Visual impairments are associated with retinal microvascular density in patients with Parkinson's disease. Front Neurosci. 2021; 15: 718820, doi:10.3389/fnins.2021.718820. [CrossRef] [PubMed]
Cennamo G, Montorio D, Ausiello FP, et al. Correlation between retinal vascularization and disease aggressiveness in amyotrophic lateral sclerosis. Biomedicines. 2022; 10(10): 2390, doi:10.3390/biomedicines10102390. [CrossRef] [PubMed]
Amini E, Moghaddasi M, Habibi SAH, et al. Huntington's disease and neurovascular structure of retina. Neurol Sci Off J Ital Neurol Soc Ital Soc Clin Neurophysiol. 2022; 43(10): 5933–5941, doi:10.1007/s10072-022-06232-3.
Kersten HM, HV Danesh-meyer, Kilfoyle DH, Roxburgh RH. Optical coherence tomography findings in Huntington's disease: a potential biomarker of disease progression. J Neurol. 2015; 262(11): 2457–2465, doi:10.1007/s00415-015-7869-2. [CrossRef] [PubMed]
Gatto E, Parisi V, Persi G, et al. Optical coherence tomography (OCT) study in Argentinean Huntington's disease patients. Int J Neurosci. 2018; 128(12): 1157–1162, doi:10.1080/00207454.2018.1489807. [CrossRef] [PubMed]
Mazur-Michałek I, Kowalska K, Zielonka D, et al. Structural abnormalities of the optic nerve and retina in Huntington's disease pre-clinical and clinical settings. Int J Mol Sci. 2022; 23(10): 5450, doi:10.3390/ijms23105450. [CrossRef] [PubMed]
Andrade C, Beato J, Monteiro A, et al. Spectral-domain optical coherence tomography as a potential biomarker in Huntington's disease. Mov Disord Off J Mov Disord Soc. 2016; 31(3): 377–383, doi:10.1002/mds.26486. [CrossRef]
Di Maio LG, Montorio D, Peluso S, et al. Optical coherence tomography angiography findings in Huntington's disease. Neurol Sci Off J Ital Neurol Soc Ital Soc Clin Neurophysiol. 2021; 42(3): 995–1001, doi:10.1007/s10072-020-04611-2.
Schmid RD, Remlinger J, Abegg M, et al. No optical coherence tomography changes in premanifest Huntington's disease mutation carriers far from disease onset. Brain Behav. 2022; 12(6): e2592, doi:10.1002/brb3.2592. [CrossRef] [PubMed]
Gulmez Sevim D, Unlu M, Gultekin M, Karaca C. Retinal single-layer analysis with optical coherence tomography shows inner retinal layer thinning in Huntington's disease as a potential biomarker. Int Ophthalmol. 2019; 39(3): 611–621, doi:10.1007/s10792-018-0857-7. [CrossRef] [PubMed]
Choi SH, Park SJ, Kim NR. Macular ganglion cell -inner plexiform layer thickness is associated with clinical progression in mild cognitive impairment and Alzheimers disease. PLoS One. 2016; 11(9): e0162202, doi:10.1371/journal.pone.0162202. [CrossRef] [PubMed]
Bulut M, Yaman A, Erol MK, et al. Choroidal thickness in patients with mild cognitive impairment and Alzheimer's type dementia. J Ophthalmol. 2016; 2016: 7291257, doi:10.1155/2016/7291257. [CrossRef] [PubMed]
Oktem EO, Derle E, Kibaroglu S, Oktem C, Akkoyun I, Can U. The relationship between the degree of cognitive impairment and retinal nerve fiber layer thickness. Neurol Sci J Ital Neurol Soc Ital Soc Clin Neurophysiol. 2015; 36(7): 1141–1146, doi:10.1007/s10072-014-2055-3.
Zhang L, Xu Y, Zhu Y. Relationships between optic nerve damage and the severity of cognitive impairment in patients with mild cognitive impairment and Alzheimer's disease. Biomed Res. 2017; 28(13): 5708–5712.
Shen Y, Liu L, Cheng Y, et al. Retinal nerve fiber layer thickness is associated with episodic memory deficit in mild cognitive impairment patients. Curr Alzheimer Res. 2014; 11(3): 259–266, doi:10.2174/1567205011666140131114418. [CrossRef] [PubMed]
Zhang YS, Onishi AC, Zhou N. Characterization of inner retinal hyperreflective alterations in early cognitive impairment on adaptive optics scanning laser ophthalmoscopy. Invest Ophthalmol Vis Sci. 2019; 60(10): 3527–3536, doi:10.1167/iovs.19-27135. [CrossRef] [PubMed]
Gao L, Liu Y, Li X, Bai Q, Liu P. Abnormal retinal nerve fiber layer thickness and macula lutea in patients with mild cognitive impairment and Alzheimer's disease. Arch Gerontol Geriatr. 2015; 60(1): 162–167, doi:10.1016/j.archger.2014.10.011. [CrossRef] [PubMed]
Liu S, Ong YT, Hilal S. The association between retinal neuronal layer and brain structure is disrupted in patients with cognitive impairment and Alzheimer's disease. J Alzheimers JAD. 2016; 54(2): 585–595, doi:10.3233/JAD-160067. [CrossRef] [PubMed]
Jiang H, Wei Y, Shi Y. Altered macular microvasculature in mild cognitive impairment and Alzheimer disease. J Neuro-Ophthalmol J North Am Neuro-Ophthalmol Soc. 2018; 38(3): 292–298, doi:10.1097/WNO.0000000000000580. [CrossRef]
Yoon SP, Grewal DS, Thompson AC. Retinal microvascular and neurodegenerative changes in Alzheimer's disease and mild cognitive impairment compared to controls. Ophthalmol Retina. 2019; 3(6): 489–499, doi:10.1016/j.oret.2019.02.002. [CrossRef] [PubMed]
Pillai JA, Bermel R, Bonner-Jackson A, et al. Retinal nerve fiber layer thinning in Alzheimer's disease: a case–control study in comparison to normal aging, Parkinson's disease, and non-Alzheimer's dementia. Am J Alzheimers Dis Dementiasr. 2016; 31(5): 430–436, doi:10.1177/1533317515628053. [CrossRef]
Goutam B, Hashmi MF, Geem ZW, Bokde ND. A comprehensive review of deep learning strategies in retinal disease diagnosis using fundus images. IEEE Access. 2022; 10: 57796–57823, doi:10.1109/ACCESS.2022.3178372. [CrossRef]
Cheung CY, Ran AR, Wang S, et al. A deep learning model for detection of Alzheimer's disease based on retinal photographs: a retrospective, multicentre case-control study. Lancet Digit Health. 2022; 4(11): e806–e815, doi:10.1016/S2589-7500(22)00169-8. [CrossRef] [PubMed]
Wisely CE, Wang D, Henao R, et al. Convolutional neural network to identify symptomatic Alzheimer's disease using multimodal retinal imaging. Br J Ophthalmol. 2022; 106(3): 388–395, doi:10.1136/bjophthalmol-2020-317659. [CrossRef] [PubMed]
Nunes A, Silva G, Duque C, et al. Retinal texture biomarkers may help to discriminate between Alzheimer's, Parkinson's, and healthy controls. PLoS One. 2019; 14(6): e0218826, doi:10.1371/journal.pone.0218826. [CrossRef] [PubMed]
Zee B, Wong Y, Lee J, et al. Machine-learning method for localization of cerebral white matter hyperintensities in healthy adults based on retinal images. Brain Commun. 2021; 3(3): fcab124, doi:10.1093/braincomms/fcab124. [CrossRef] [PubMed]
Corbin D, Lesage F. Assessment of the predictive potential of cognitive scores from retinal images and retinal fundus metadata via deep learning using the CLSA database. Sci Rep. 2022; 12(1): 5767, doi:10.1038/s41598-022-09719-3. [CrossRef] [PubMed]
Hua R, Xiong J, Li G, et al. Development and validation of a deep learning algorithm based on fundus photographs for estimating the CAIDE dementia risk score. Age Ageing. 2022; 51(12): afac282, doi:10.1093/ageing/afac282. [CrossRef] [PubMed]
Poplin R, Varadarajan AV, Blumer K, et al. Prediction of cardiovascular risk factors from retinal fundus photographs via deep learning. Nat Biomed Eng. 2018; 2(3): 158–164, doi:10.1038/s41551-018-0195-0. [CrossRef] [PubMed]
Zhu Z, Shi D, Guankai P, et al. Retinal age gap as a predictive biomarker for mortality risk. Br J Ophthalmol. 2023; 107(4): 547–554, doi:10.1136/bjophthalmol-2021-319807. [CrossRef] [PubMed]
Abdulsahib AA, Mahmoud MA, Aris H, Gunasekaran SS, Mohammed MA. An automated image segmentation and useful feature extraction algorithm for retinal blood vessels in fundus images. Electronics. 2022; 11(9): 1295, doi:10.3390/electronics11091295. [CrossRef]
Ma Y, Hao H, Xie J. ROSE: a retinal OCT-angiography vessel segmentation dataset and new model. IEEE Trans Med Imaging. 2021; 40(3): 928–939, doi:10.1109/TMI.2020.3042802. [CrossRef] [PubMed]
Lam C, Yu C, Huang L, Rubin D. Retinal lesion detection with deep learning using image patches. Invest Ophthalmol Vis Sci. 2018; 59(1): 590–596, doi:10.1167/iovs.17-22721. [CrossRef] [PubMed]
Shi D, Lin Z, Wang W, et al. A deep learning system for fully automated retinal vessel measurement in high throughput image analysis. Front Cardiovasc Med. 2022; 9: 823436. Accessed February 16, 2023, doi:10.3389/fcvm.2022.823436. [CrossRef] [PubMed]
Khan NC, Perera C, Dow ER, et al. Predicting systemic health features from retinal fundus images using transfer-learning-based artificial intelligence models. Diagnostics. 2022; 12(7): 1714, doi:10.3390/diagnostics12071714. [CrossRef] [PubMed]
Geirhos R, Rubisch P, Michaelis C, Bethge M, Wichmann FA, Brendel W. ImageNet-trained CNNs are biased towards texture; increasing shape bias improves accuracy and robustness. ArXiv Prepr 2018. ArXiv:1811.12231, doi:10.48550/arXiv.1811.12231.
Nunes A, Silva G, Duque C, et al. Retinal texture biomarkers may help to discriminate between Alzheimer's, Parkinson's, and healthy controls. PLoS One. 2019; 14(6): e0218826, doi:10.1371/journal.pone.0218826. [CrossRef] [PubMed]
Li Z, Xu M, Yang X, Han Y. Multi-label fundus image classification using attention mechanisms and feature fusion. Micromachines. 2022; 13(6): 947, doi:10.3390/mi13060947. [CrossRef] [PubMed]
Gessert N, Nielsen M, Shaikh M, Werner R, Schlaefer A. Skin lesion classification using ensembles of multi-resolution EfficientNets with meta data. MethodsX. 2020; 7: 100864, doi:10.1016/j.mex.2020.100864. [CrossRef] [PubMed]
Mitani A, Huang A, Venugopalan S, et al. Detection of anaemia from retinal fundus images via deep learning. Nat Biomed Eng. 2020; 4(1): 18–27, doi:10.1038/s41551-019-0487-z. [CrossRef] [PubMed]
Dai Y, Gieseke F, Oehmcke S, Wu Y, Barnard K. Attentional Feature Fusion. In: 2021 IEEE Winter Conference on Applications of Computer Vision (WACV). 2021: 3559–3568, doi:10.1109/WACV48630.2021.00360.
Liang Y, Li M, Jiang C. Generating self-attention activation maps for visual interpretations of convolutional neural networks. Neurocomputing. 2022; 490: 206–216, doi:10.1016/j.neucom.2021.11.084. [CrossRef]
Zhang X, Li F, Li D, et al. Automated detection of severe diabetic retinopathy using deep learning method. Graefes Arch Clin Exp Ophthalmol. 2022; 260(3): 849–856, doi:10.1007/s00417-021-05402-x. [CrossRef] [PubMed]
Tian J, Smith G, Guo H, et al. Modular machine learning for Alzheimer's disease classification from retinal vasculature. Sci Rep. 2021; 11: 238, doi:10.1038/s41598-020-80312-2. [CrossRef] [PubMed]
Marquié M, Valero S, Castilla-Marti M, et al. Association between retinal thickness and β-amyloid brain accumulation in individuals with subjective cognitive decline: fundació ACE healthy brain initiative. Alzheimers Res Ther. 2020; 12(1): 37, doi:10.1186/s13195-020-00602-9. [CrossRef] [PubMed]
Sandeep CS, Sukesh Kumar A, Mahadevan K, Manoj P. Analysis of retinal OCT images for the early diagnosis of Alzheimer's disease. In: Chattopadhyay S, Roy T, Sengupta S, Berger-Vachon C, eds. Modelling and Simulation in Science, Technology and Engineering Mathematics. Advances in Intelligent Systems and Computing. New York, NY: Springer International Publishing; 2019, doi:10.1007/978-3-319-74808-5_43.
Lemmens S, Van Craenendonck T, Van Eijgen J, et al. Combination of snapshot hyperspectral retinal imaging and optical coherence tomography to identify Alzheimer's disease patients. Alzheimers Res Ther. 2020; 12: 144, doi:10.1186/s13195-020-00715-1. [CrossRef] [PubMed]
Larrosa JM, Garcia-Martin E, Bambo MP. Potential new diagnostic tool for Alzheimer's disease using a linear discriminant function for Fourier domain optical coherence tomography. Invest Ophthalmol Vis Sci. 2014; 55(5): 3043–3051, doi:10.1167/iovs.13-13629. [CrossRef] [PubMed]
Zhang Q, Li J, Bian M, et al. Retinal imaging techniques based on machine learning models in recognition and prediction of mild cognitive impairment. Neuropsychiatr Dis Treat. 2021; 17: 3267–3281, doi:10.2147/NDT.S333833. [CrossRef] [PubMed]
Ma Y, Hao H, Xie J. ROSE: a retinal OCT-angiography vessel segmentation dataset and new model. IEEE Trans Med Imaging. 2021; 40(3): 928–939, doi:10.1109/TMI.2020.3042802. [CrossRef] [PubMed]
Figure 1.
 
Conceptual search design as structured within Ovid MEDLINE search tool. exp, exploded, represents that the search was expanded to include similar related MeSH terms; .mp, multipurpose, indicates that all areas of the publication (title, abstract, body, etc.) were searched.
Figure 1.
 
Conceptual search design as structured within Ovid MEDLINE search tool. exp, exploded, represents that the search was expanded to include similar related MeSH terms; .mp, multipurpose, indicates that all areas of the publication (title, abstract, body, etc.) were searched.
Figure 2.
 
Architecture of VGGNet. Used with permission from Goutam et al. 2022.60 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Figure 2.
 
Architecture of VGGNet. Used with permission from Goutam et al. 2022.60 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Figure 3.
 
Simple approach for combining fundoscopic image processing and metadata processing. Adapted with permission from Gessert et Al. 2020.77 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Figure 3.
 
Simple approach for combining fundoscopic image processing and metadata processing. Adapted with permission from Gessert et Al. 2020.77 CONV, a conversion of the convolutional neural network; FC, fully connected layer; ReLU, rectified linear unit.
Figure 4.
 
Example of attention maps used in the classification of diabetic retinopathy. Pixels in the image that are of higher relevance to model decision making are highlighted in yellow and red. Adapted with permission from Zhang et al. 2022.81
Figure 4.
 
Example of attention maps used in the classification of diabetic retinopathy. Pixels in the image that are of higher relevance to model decision making are highlighted in yellow and red. Adapted with permission from Zhang et al. 2022.81
Table 1.
 
Summary of Retinal Image Datasets in Patients With Neurodegenerative Disease
Table 1.
 
Summary of Retinal Image Datasets in Patients With Neurodegenerative Disease
Table 2.
 
Summary of Retinal Findings in Neurodegenerative Disease
Table 2.
 
Summary of Retinal Findings in Neurodegenerative Disease
Table 3.
 
Deep Learning Algorithms and Other Predictive Models for Retinal Image Analysis in Neurodegenerative Disease
Table 3.
 
Deep Learning Algorithms and Other Predictive Models for Retinal Image Analysis in Neurodegenerative Disease
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×