August 2024
Volume 13, Issue 8
Open Access
Artificial Intelligence  |   August 2024
Multimodal Retinal Imaging Classification for Parkinson's Disease Using a Convolutional Neural Network
Author Affiliations & Notes
  • Alexander Richardson
    Duke Eye Center, Department of Ophthalmology, Duke University School of Medicine, Durham, NC, USA
    iMIND Research Group, Duke University School of Medicine, Durham, NC, USA
    Department of Computer Science, Duke University, Durham, NC, USA
  • Anita Kundu
    Duke Eye Center, Department of Ophthalmology, Duke University School of Medicine, Durham, NC, USA
    iMIND Research Group, Duke University School of Medicine, Durham, NC, USA
  • Ricardo Henao
    iMIND Research Group, Duke University School of Medicine, Durham, NC, USA
    Department of Computer Science, Duke University, Durham, NC, USA
  • Terry Lee
    Duke Eye Center, Department of Ophthalmology, Duke University School of Medicine, Durham, NC, USA
    iMIND Research Group, Duke University School of Medicine, Durham, NC, USA
  • Burton L. Scott
    iMIND Research Group, Duke University School of Medicine, Durham, NC, USA
    Department of Neurology, Duke University School of Medicine, Durham, NC, USA
  • Dilraj S. Grewal
    Duke Eye Center, Department of Ophthalmology, Duke University School of Medicine, Durham, NC, USA
    iMIND Research Group, Duke University School of Medicine, Durham, NC, USA
  • Sharon Fekrat
    Duke Eye Center, Department of Ophthalmology, Duke University School of Medicine, Durham, NC, USA
    iMIND Research Group, Duke University School of Medicine, Durham, NC, USA
    Department of Neurology, Duke University School of Medicine, Durham, NC, USA
  • Correspondence: Sharon Fekrat, Duke Eye Center, Department of Ophthalmology, Duke University School of Medicine, Box 3802, Erwin Road, Durham, NC 27710, USA. e-mail: imind@duke.edu 
Translational Vision Science & Technology August 2024, Vol.13, 23. doi:https://doi.org/10.1167/tvst.13.8.23
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Alexander Richardson, Anita Kundu, Ricardo Henao, Terry Lee, Burton L. Scott, Dilraj S. Grewal, Sharon Fekrat; Multimodal Retinal Imaging Classification for Parkinson's Disease Using a Convolutional Neural Network. Trans. Vis. Sci. Tech. 2024;13(8):23. https://doi.org/10.1167/tvst.13.8.23.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose: Changes in retinal structure and microvasculature are connected to parallel changes in the brain. Two recent studies described machine learning algorithms trained on retinal images and quantitative data that identified Alzheimer's dementia and mild cognitive impairment with high accuracy. Prior studies also demonstrated retinal differences in individuals with PD. Herein, we developed a convolutional neural network (CNN) to classify multimodal retinal imaging from either a Parkinson's disease (PD) or control group.

Methods: We trained a CNN to receive retinal image inputs of optical coherence tomography (OCT) ganglion cell–inner plexiform layer (GC-IPL) thickness color maps, OCT angiography 6 × 6-mm en face macular images of the superficial capillary plexus, and ultra-widefield (UWF) fundus color and autofluorescence photographs to classify the retinal imaging as PD or control. The model consists of a shared pretrained VGG19 feature extractor and image-specific feature transformations which converge to a single output. Model results were assessed using receiver operating characteristic (ROC) curves and bootstrapped 95% confidence intervals for area under the ROC curve (AUC) values.

Results: In total, 371 eyes of 249 control subjects and 75 eyes of 52 PD subjects were used for training, validation, and testing. Our best CNN variant achieved an AUC of 0.918. UWF color photographs were the most effective imaging input, and GC-IPL thickness maps were the least contributory.

Conclusions: Using retinal images, our pilot CNN was able to identify individuals with PD and serves as a proof of concept to spur the collection of larger imaging datasets needed for clinical-grade algorithms.

Translational Relevance: Developing machine learning models for automated detection of Parkinson's disease from retinal imaging could lead to earlier and more widespread diagnoses.

Introduction
Parkinson's disease (PD) is a neurodegenerative disorder exhibiting motor and non-motor features affecting 3% of adults over age 65.1 Loss of dopaminergic cells in the substantia nigra, associated with cardinal motor symptoms including bradykinesia, rigidity, and a resting tremor, is the hallmark of PD. When cardinal motor symptoms develop, over 80% of nigrostriatal dopaminergic cells are already affected.2 
Multiple medications, such as levodopa and carbidopa, lessen PD symptoms, yet there is no pharmacologic cure. Medications for PD symptoms are more effective when administered earlier in the disease course; thus, earlier diagnosis and treatment may delay symptom progression and improve quality of life.3 The standard of PD diagnosis remains the postmortem histological presence of α-synuclein–positive inclusions in neuronal cells.4 Various diagnostic genetic, molecular, and biochemical biomarkers of PD have been studied; however, no definitive biomarkers have been identified thus far that provide a high-fidelity model of PD disease identification or any prediction of course.5 
A clinical diagnosis of PD is currently made by clinical assessment in conjunction with scores and tools that aim to standardize evaluation, such as the Unified Parkinson's Disease Rating Scale.6 Even with standardized scales, an effect of interprovider variability in assessment and evaluation is observed that varies across the PD continuum. Studies have shown that PD diagnosis in the earlier stages may have a diagnostic accuracy as low as 26%.7 Given the devastating effects of PD on quality of life, as well as the lack of a cure, it is not surprising that a large unmet need exists for specific, cost-effective, and non-invasive tools to assist in PD screening, diagnosis, and monitoring. 
Changes in retinal structure and vasculature have been associated with neurodegenerative changes in the brain.8,9 Cross-sectional studies in PD using optical coherence tomography (OCT) and OCT angiography (OCTA) imaging have revealed a decrease in retinal perfusion density, vessel density, and choroidal vascularity index in patients with PD compared to cognitively normal controls without PD.9 Another study revealed increased retinal venule tortuosity in PD patients compared to controls using ultra-widefield (UWF) scanning laser ophthalmoscopy (SLO).10 An investigation into OCT features found that ganglion cell–inner plexiform layer (GC-IPL) and inner nuclear layer thinning was significantly associated with a PD diagnosis.11 Further, longitudinal retinal imaging data have suggested that there may be significantly faster neuronal and microvascular loss over time in individuals with PD, suggesting its prognostic potential.12 
The use of machine learning in PD diagnosis is a rapidly growing area of interest. Retinal imaging is a commercially available and broadly accessible imaging type with significant translational potential.1321 Herein, we developed a convolutional neural network (CNN) model with the goal of identifying PD using solely multimodal retinal images comprised of OCT, OCTA, and UWF SLO images. 
Methods
Participants
This study (clinicaltrials.gov identifier NCT03233646) was approved by the Duke Health Institutional Review Board (Pro00082598). The study followed the Health Insurance Portability and Accountability Act of 1996 and adhered to all tenets of the Declaration of Helsinki. Written informed consent was provided by all study subjects or their legally authorized representative prior to study participation. 
Study participant images were collected from 2017 to 2023 from the existing Eye Multimodal Imaging in Neurodegenerative Disease (iMIND) database. Prior to enrollment, all patients with PD were evaluated by an experienced movement disorders specialist (BLS), and the clinical diagnosis of PD was confirmed according to the clinical criteria of the International Parkinson and Movement Disorder Society.22 The control group consisted of cognitively normal individuals from the iMIND database. Subjects with prior retinal surgery, high myopia, or hyperopia (greater than 6 diopters in magnitude), significant head tremor, diabetes, uncontrolled hypertension, glaucoma, history of retinal or optic nerve pathology, or corrected visual acuity less than 20/40 on the day of image acquisition were excluded. Study participants also underwent cognitive function evaluation by trained iMIND study personnel on the day of image acquisition using the Mini-Mental State Examination (MMSE), an 11-question measure that tests five areas of cognitive function.23 
Image Acquisition
Non-mydriatic images were acquired by experienced iMIND study personnel using the ZEISS CIRRUS 5000 Spectral-Domain OCT with AngioPlex software (version 11.0.0.29946; Carl Zeiss Meditec, Dublin, CA), and layer segmentation was done using the integrated ZEISS software. Active retinal motion tracking software was used to reduce motion artifact during acquisition. Images were each reviewed manually by trained study personnel (AK, DSG) for quality assessment. Images with lower than 7/10 signal strength index, motion artifact, segmentation artifact, shadow artifact, or focal signal loss were excluded. These designations were designed to exclude images where analysis would be hindered by the severe artifact. Only subjects that had good-quality images as determined by trained iMIND study staff (AK, SF, DSG) for all four image inputs, GC-IPL color thickness maps, 6 × 6-mm OCTA superficial capillary plexus images, UWF color photographs, and UWF autofluorescence (FAF) photographs, were included in the analysis. 
GC-IPL Thickness Color Maps
One 512 × 128-µm macular cube scan was acquired for each eye, and the GC-IPL thickness color map was exported in tagged image file format (TIFF) for analysis. 
OCTA Scans
For each eye, a 6 × 6-mm OCTA scan centered on the fovea was acquired, and the superficial capillary plexus (SCP) image was exported in bitmap format for analysis. 
UWF Fundus Color and FAF Photographs
UWF fundus color and FAF photographs were obtained by trained study personnel using nonmydriatic SLO (Optos California; Optos, Marlborough, MA). Fundus photos with peripheral distortion or motion and eyelid artifact within the major vascular arcades were excluded. 
Model Development
In total, 296 control eyes (80%) were used for training, 36 (10%) for validation, and 39 (10%) for testing; 59 PD eyes (79%) were used for training, six for validation (8%), and 10 (14%) for testing. Two hundred and four eyes of control subjects and 39 eyes of PD subjects were used for development and testing of the model that considered only right eyes, randomly assigned to an 80% train, 10% validation, and 10% test split. Also, 169 eyes of control subjects and 36 eyes of PD subjects were used for development and testing of the model that considered only left eyes, randomly assigned to an 80% train, 10% validation, and 10% test split. 
Deep learning is a branch of machine learning that uses neural networks with multiple layers to model complex patterns. Three separate deep-learning models were developed, using images from the right eye only, left eye only, or both eyes. The single-eye models were developed because some subjects had only one imageable eye or a poor image quality for one eye. All models were trained to receive inputs of GC-IPL thickness color maps, OCTA SCP 6 × 6-mm en face images, and UWF fundus color and FAF photographs to produce a score suggesting whether a subject carried a clinical diagnosis of PD. 
For dataset preprocessing, all images were resized with Python Imaging Library to 256 × 256 pixels and converted to 224 × 224 PyTorch tensors to be compatible with the VGG19 architecture. For dataset augmentation, all images were subjected to random horizontal and vertical flips with a probability of 0.5. For all models, an 80% train, a 10% validation, and a 10% test split were used. To ensure that there was no overlap among the training, validation, and test groups, random assignment was done by patient ID so that eyes of the same subject were assigned to the same group (i.e., training, validation, or test). The model parameters with the best performance on the validation split were selected for testing. The median area under the receiver operating characteristic (ROC) curve (AUC) of 10 model training runs was selected. 
The model consisted of a shared pretrained convolutional encoder, parallel image modality–specific feature transformations, and prediction heads for each modality that converge into a single output (Fig. 1). Transfer learning was leveraged from a VGG19 model pretrained on ImageNet as the convolutional encoder.24 Pretraining is a machine learning technique where a model is trained on a large general dataset so it can learn basic image feature extraction such as edge detection, and transfer learning is where the pretrained neural network is fine-tuned to a smaller and specific problem, such as Parkinson's disease classification with retinal imaging. The images were then processed through the convolutional encoder, before being fed to image modality–specific feature transformations consisting of three fully connected layers: a dropout layer, a batch normalization function, and a rectified linear unit (ReLU) activation function. The outputs of these parallel feature transformations were averaged and processed through a sigmoid activation function to produce a total model output. The model was trained for 50 epochs using the adaptive moment estimation (ADAM) optimizer25 with a weight decay of 0.01. Two methods were used to counteract the imbalanced training dataset, which contained fewer PD eyes than control eyes: a modified binary cross entropy loss function and a weighted random sampler. The weighted random sampler oversampled PD samples to create a 1:1 class ratio in training batches. 
Figure 1.
 
Illustration of the PD classification model architecture. The model leverages a pretrained VGG19 CNN as a shared-weight image feature extractor (shown in the blue dotted box) to generate feature maps for each imaging modality, then feeds the results to modality-specific functions. The model averages the outputs from these functions and converts the result into a PD probability score in the [0, 1] range to obtain a final prediction result.
Figure 1.
 
Illustration of the PD classification model architecture. The model leverages a pretrained VGG19 CNN as a shared-weight image feature extractor (shown in the blue dotted box) to generate feature maps for each imaging modality, then feeds the results to modality-specific functions. The model averages the outputs from these functions and converts the result into a PD probability score in the [0, 1] range to obtain a final prediction result.
Statistical Analysis
The average age of the cognitively normal control cohort was compared with that of the PD cohort using an independent sample t-test. As a performance metric, we considered the ROC curve, the AUC, sensitivity, and specificity. Bootstrapped ROC computations with 10,000 samples were used to estimate the 95% confidence intervals (CIs) for the AUC values. To convert the model outputs into a clinical classification (PD vs. control), we thresholded the model scores with the label set to optimize the Youden index,26 a classification measure specifically designed for medical contexts. The model results were compared for each imaging modality as the sole input to provide information on their relative efficacy and contrast with the full model. The DeLong test, a statistical method to compare the predictive power of different classification algorithms, was used to compare the performance metrics for different inputs.27 We compared the predictions for a model using each imaging modality as the sole input to the model utilizing all inputs for both the right eye and left eye models and averaged the DeLong test results. A higher DeLong test P value and a higher AUC for a single feature model can be interpreted as a proxy of significant predictive power, as the predictions of the single feature model are not statistically significantly different from the predictions of the model processing all imaging types. 
Results
A total of 371 eyes (204 right and 169 left) of 249 control subjects and 75 eyes (39 right and 36 left) of 52 PD subjects were included for development and testing of the model that considered both right and left eyes. The PD cohort had a mean age of 65.9 ± 9.5 years, and the control cohort had a mean age of 66.4 ± 9.1 years (P = 0.656). The PD cohort had 31 males and 21 females, and the control cohort had 67 males and 184 females (P < 0.05). 
The Table provides the AUC values for the right, left, and both eyes models, as well as the AUC values for the model trained on each modality individually. Each model used the structure described in Figure 1, but with various input combinations as outlined in the Table. The best performing models used all inputs, including GC-IPL, OCTA, UWF color, and UWF FAF. The right-eye-only model had an AUC of 0.910 (95% CI, 0.798–1), and the left-eye-only model had an AUC of 0.918 (95% CI, 0.787–1) when including all inputs. The model using both eyes with all imaging modalities had an AUC of 0.861 (95% CI, 0.750–0.955). The right-eye-only model achieved a sensitivity of 100% and a specificity of 85%. The left-eye-only model achieved a sensitivity of 100% and specificity of 85.7%. The UWF color image was the best single input (right-eye-only AUC = 0.798, left-eye-only AUC = 0.838), followed by OCTA (right-eye-only AUC = 0.711, left-eye-only AUC = 0.811), and UWF FAF (right-eye-only AUC = 0.682, left-eye-only AUC = 0.794). GC-IPL thickness maps (right-eye-only AUC = 0.630, left-eye-only AUC = 0.556) were the worst performing single input. Figure 2 demonstrates model scores for each eye in the PD and control test sets evaluated by the three eye models including all inputs. Figure 3 displays the ROC curves for the right-eye, left-eye, and both eyes models with all inputs. The ROC curves in Figure 3 were also used to calculate the Youden index threshold values shown in Figure 2, which functioned as the decision point to determine sensitivity and specificity.26 Figures 4A and 4B show the confusion matrices, demonstrating how eyes in the test sets were classified by the models using the Youden thresholds. Figure 4C shows the confusion matrix for the model considering both right and left eyes. Figure 5 displays the precision–recall curves, which are often employed for binary classifiers on datasets with class imbalances, for each of the three eye models including all inputs.27 The AUC values for each of the precision–recall curves were 0.405, 0.535, and 0.677 for the models utilizing both eyes, right eyes only, and left eyes only, respectively. 
Table.
 
AUC on the Test Set Describes the Performance of Each Model in Predicting a Symptomatic PD Diagnosis Given the Imaging Inputs Described in the Left Column
Table.
 
AUC on the Test Set Describes the Performance of Each Model in Predicting a Symptomatic PD Diagnosis Given the Imaging Inputs Described in the Left Column
Figure 2.
 
Model outputs for each eye in the PD and control test sets evaluated by the right-eye-only, left-eye-only, and both eyes models. The red dashed line indicates the decision boundary of the model for its output probabilities. Less overlap between the box plots for the control and PD groups indicates a model had better classification performance, and points on the box plot closer to the threshold (red dashed line) represent patients that the model classified with lower confidence.
Figure 2.
 
Model outputs for each eye in the PD and control test sets evaluated by the right-eye-only, left-eye-only, and both eyes models. The red dashed line indicates the decision boundary of the model for its output probabilities. Less overlap between the box plots for the control and PD groups indicates a model had better classification performance, and points on the box plot closer to the threshold (red dashed line) represent patients that the model classified with lower confidence.
Figure 3.
 
ROC curve for the right-eye-only, left-eye-only, and both eyes models using all features on the test set.
Figure 3.
 
ROC curve for the right-eye-only, left-eye-only, and both eyes models using all features on the test set.
Figure 4.
 
Confusion matrices for the right eye only (A), left eye only (B), and both right and left eyes (C).
Figure 4.
 
Confusion matrices for the right eye only (A), left eye only (B), and both right and left eyes (C).
Figure 5.
 
Precision–recall curves for each of the three models incorporating all image types. The AUC value is 0.405 for the model utilizing both eyes, 0.535 for the model utilizing right eyes, and 0.677 for the model utilizing left eyes only.
Figure 5.
 
Precision–recall curves for each of the three models incorporating all image types. The AUC value is 0.405 for the model utilizing both eyes, 0.535 for the model utilizing right eyes, and 0.677 for the model utilizing left eyes only.
Discussion
Our findings demonstrate that using deep learning applied to retinal imaging for identifying Parkinson's disease yields strong predictive performance. The right-eye, left-eye, and both eyes models used GC-IPL thickness color maps, en face OCTA images, and UWF color and FAF images to achieve AUCs of 0.910, 0.918, and 0.861, respectively, when applied to an independent test set. Because the 95% CI for the both eyes model closely overlapped with the 95% CIs for the right-eye and left-eye models, we cannot conclude that the single-eye models are superior, matching the findings of our previous work into the interocular symmetry of vasculature biomarkers in OCTA.28 Similarly, we cannot conclude that there are differences between the right-eye-only and left-eye-only models when each imaging input was isolated due to their closely overlapping CIs. The precision–recall curves for the three models that used all images aligned closely with each other and did not show evidence of extreme class imbalance in prediction accuracy. Our research group has previously published CNN models capable of using GC-IPL, OCTA, and UWF images, along with quantitative data, to predict Alzheimer's disease (AD) and mild cognitive impairment (MCI) diagnosis with high accuracy.20,21 The performance qualities were similar to the performance demonstrated in our previously developed CNN for classification of AD and control subjects using multimodal retinal images, but our PD model uses only images and does not include quantitative data.21 
Previous studies have established a link between PD pathology and findings in the optic nerve and retina. Thinning of the peripapillary retinal nerve fiber layer (RNFL), foveal thinning, and outer plexiform layer thickening have been observed in the eyes of PD subjects.29,30 In Parkinson's disease, α-synuclein has been found in neuronal inclusions in the ganglion cell layer and inner nuclear layer.31,32 The loss of dopaminergic neurons in the substantia nigra has also been associated with RNFL thinning that may correlate with PD severity.33 It is possible that the CNN presented in this paper may be recognizing thinning in the RNFL or ganglion cell layer as a visual indicator of PD.29,30 Current evidence also indicates a chain of vascular events observed in postmortem tissue and preclinical models of PD that might manifest in the capillary network of the retina.34 The loss of dopaminergic neurons in PD widens the close contacts between neurons and retinal vasculature, impairing the blood–retinal barrier and triggering angiogenesis, but angiogenesis fails and vascular regression occurs.34,35 The retinal capillaries of patients with PD compared to controls are fewer in number, shorter in length, and larger in diameter.35 
Machine learning–based diagnostic tools using retinal image inputs also have systemic disease utility. Retinal microvascular analysis has shown promise in identifying cardiovascular disease risk, including predicting risk factors such as blood pressure and coronary artery calcium score.36 Algorithms using retinal photographs to identify diabetic retinopathy can identify symptoms and predict progression in individual patients, serving as a non-invasive tool for diabetes management.3741 Machine learning models that work with widely accessible data, such as retinal photographs obtained in primary care and optometry clinics, have impactful translational potential in medicine, as the pool of patients is significantly larger than in ophthalmology. The efficacy of UWF color images as a single input in our PD algorithm is therefore particularly important, given the existing widespread availability of UWF imaging devices and the potential of rapid deployment for point-of-care testing. 
Our limited dataset size constrained the complexity of our model architecture. Complex neural networks with numerous parameters require larger training datasets to minimize overfitting, which may impair the ability to generalize to new samples. Models such as the CNN created herein would benefit from larger datasets to fully extract features, especially for complex or subtle changes such as retinal vasculature and nerve architecture differences in PD patients. Although we leveraged techniques such as transfer learning from a pretrained VGG19 neural network, weighted random sampling, and a weighted loss function, the performance of our model was still constrained by the dataset size. The manual grading of images required for diagnosis and assessment of image quality is a time-intensive component, but this may be mitigated by using algorithms such as the model developed by iMIND to automate and standardize image quality assessment.42 Another limitation was class imbalance, as there were fewer PD eyes than control eyes. A dataset with an imbalanced class distribution is a hindrance to classification, especially of the minority class, but our model consistently had better performance for PD eyes, with a sensitivity near 100% for most model variations. This can be a useful attribute, as sensitivity is preferred over specificity in the clinical screening setting. 
The current PD model can classify patient retinal imaging as belonging to a PD group or a control group. We hope that our proof-of-concept model will encourage the development of the larger datasets needed to train clinical-quality algorithms and longitudinal datasets with subjects who have developed PD from a neurological normal state. Longitudinal datasets are particularly important, because they could be used to build deep-learning models that identify early-stage PD. 
A similar model could be trained to predict the development of PD sequelae. Another future direction is combining previous classification algorithms for MCI and AD with this PD algorithm to screen for multiple neurodegenerative diseases.20,21 Continued recruitment and imaging of PD and control subjects, using standardized retinal imaging protocols, to build a larger training set will also improve the performance of the current CNN. 
The model described herein has potential for clinical applications in assisting PD risk stratification. It also pushes the frontiers of this emerging field of oculomics, where retinal phenotyping and machine learning converge to allow scientists and clinicians to better understand the pathophysiology and diagnosis of neurodegenerative diseases including PD. Our pilot model herein serves as a proof of concept to spur the growth of larger datasets needed for advancements in predictive PD models and combined MCI, AD, and PD classifiers. 
Acknowledgments
Optos loaned the scanning laser ophthalmoscope used in this study; otherwise, no conflicting relationship exists for any author. 
Presented at the Duke Ophthalmology Trainee Day Scientific Session, June 23, 2023, Durham, NC. 
Disclosure: A. Richardson, None; A. Kundu, None; R. Henao, None; T. Lee, None; B.L. Scott, None; D.S. Grewal, None; S. Fekrat, None 
References
Poewe W, Seppi K, Tanner CM, et al. Parkinson disease. Nat Rev Dis Primers. 2017; 3: 17013. [CrossRef] [PubMed]
Sveinbjornsdottir S. The clinical symptoms of Parkinson's disease. J Neurochem. 2016; 139(suppl 1): 318–324. [PubMed]
Murman DL. Early treatment of Parkinson's disease: opportunities for managed care. Am J Manag Care. 2012; 18(7 suppl): S183–S188. [PubMed]
Rizzo G, Copetti M, Arcuti S, Martino D, Fontana A, Logroscino G. Accuracy of clinical diagnosis of Parkinson disease: a systematic review and meta-analysis. Neurology. 2016; 86(6): 566–576. [CrossRef] [PubMed]
Lotankar S, Prabhavalkar KS, Bhatt LK. Biomarkers for Parkinson's disease: recent advancement. Neurosci Bull. 2017; 33(5): 585–597. [CrossRef] [PubMed]
Jankovic J. Parkinson's disease: clinical features and diagnosis. J Neurol Neurosurg Psychiatry. 2008; 79(4): 368–376. [CrossRef] [PubMed]
Adler CH, Beach TG, Hentz JG, et al. Low clinical diagnostic accuracy of early vs advanced Parkinson disease: clinicopathologic study. Neurology. 2014; 83(5): 406–412. [CrossRef] [PubMed]
Yoon SP, Grewal DS, Thompson AC, et al. Retinal microvascular and neurodegenerative changes in Alzheimer's disease and mild cognitive impairment compared with control participants. Ophthalmol Retina. 2019; 3(6): 489–499. [CrossRef] [PubMed]
Robbins CB, Thompson AC, Bhullar PK, et al. Characterization of retinal microvascular and choroidal structural changes in Parkinson disease. JAMA Ophthalmol. 2021; 139(2): 182–188. [CrossRef] [PubMed]
Ma JP, Robbins C, Pead E, et al. Retinal vascular changes in Parkinson's disease on ultra-widefield retinal imaging. Invest Ophthalmol Vis Sci. 2021; 62(8): 1779.
Wagner SK, Romero-Bascones D, Cortina-Borja M, et al. Retinal optical coherence tomography features associated with incident and prevalent Parkinson disease. Neurology. 2023; 101(16): e1581–e1593. [CrossRef] [PubMed]
Kundu A, Ma JP, Robbins CB, et al. Longitudinal analysis of retinal microvascular and choroidal imaging parameters in Parkinson disease compared to controls. Ophthalmol Sci. 2023; 3(4): 100393. [CrossRef] [PubMed]
Mei J, Desrosiers C, Frasnelli J. Machine learning for the diagnosis of Parkinson's disease: a review of literature. Front Aging Neurosci. 2021; 13: 633752. [CrossRef] [PubMed]
Zhang Y, Zhang D, Gao Y, et al. Retinal flow density changes in early-stage Parkinson's disease investigated by swept-source optical coherence tomography angiography. Curr Eye Res. 2021; 46(12): 1886–1891. [CrossRef] [PubMed]
Ahn S, Shin J, Song SJ, et al. Neurologic dysfunction assessment in Parkinson disease based on fundus photographs using deep learning. JAMA Ophthalmol. 2023; 141(3): 234–240. [CrossRef] [PubMed]
Tran C, Shen K, Liu K, et al. Deep learning predicts prevalent and incident Parkinson's disease from UK Biobank fundus imaging. Sci Rep. 2024; 14(1): 3637. [CrossRef] [PubMed]
Katsimpris A, Papadopoulos I, Voulgari N, et al. Optical coherence tomography angiography in Parkinson's disease: a systematic review and meta-analysis. Eye (Lond). 2023; 37(14): 2847–2854. [CrossRef] [PubMed]
Zhou WC, Tao JX, Li J. Optical coherence tomography measurements as potential imaging biomarkers for Parkinson's disease: a systematic review and meta-analysis. Eur J Neurol. 2021; 28(3): 763–774. [CrossRef] [PubMed]
Gende M, Mallen V, de Moura J, et al. Automatic segmentation of retinal layers in multiple neurodegenerative disorder scenarios. IEEE J Biomed Health Inform. 2023; 27(11): 5483–5494. [CrossRef] [PubMed]
Wisely CE, Richardson A, Henao R, et al. A convolutional neural network using multimodal retinal imaging for differentiation of mild cognitive impairment from normal cognition. Ophthalmol Sci. 2023; 4(1): 100355. [CrossRef] [PubMed]
Wisely CE, Wang D, Henao R, et al. Convolutional neural network to identify symptomatic Alzheimer's disease using multimodal retinal imaging. Br J Ophthalmol. 2022; 106(3): 388–395. [CrossRef] [PubMed]
Postuma RB, Poewe W, Litvan I, et al. Validation of the MDS clinical diagnostic criteria for Parkinson's disease. Mov Disord. 2018; 33(10): 1601–1608. [CrossRef] [PubMed]
Folstein MF, Folstein SE, McHugh PR. “Mini-mental state”. A practical method for grading the cognitive state of patients for the clinician. J Psychiatr Res. 1975; 12(3): 189–198. [CrossRef] [PubMed]
Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. ImageNet: a large-scale hierarchical image database. In: Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition. Piscataway, NJ: Institute of Electrical and Electronics Engineers; 2009: 248–255.
Kingma DP, Ba J. Adam: a method for stochastic optimization. ArXiv, 2017, https://doi.org/10.48550/arXiv.1412.6980.
Youden WJ. Index for rating diagnostic tests. Cancer. 1950; 3(1): 32–35. [CrossRef] [PubMed]
Saito T, Rehmsmeier M. The precision-recall plot is more informative than the ROC plot when evaluating binary classifiers on imbalanced datasets. PLoS One. 2015; 10(3): e0118432. [CrossRef] [PubMed]
Ma JP, Robbins CB, Stinnett SS, et al. Repeatability of peripapillary OCT angiography in neurodegenerative disease. Ophthalmol Sci. 2021; 1(4): 100075. [CrossRef] [PubMed]
Pilat A, McLean RJ, Proudlock FA, et al. In vivo morphology of the optic nerve and retina in patients with Parkinson's disease. Invest Ophthalmol Vis Sci. 2016; 57(10): 4420–4427. [CrossRef] [PubMed]
Ucak T, Alagoz A, Cakir B, Celik E, Bozkurt E, Alagoz G. Analysis of the retinal nerve fiber and ganglion cell–inner plexiform layer by optical coherence tomography in Parkinson's patients. Parkinsonism Relat Disord. 2016; 31: 59–64. [CrossRef] [PubMed]
Bodis-Wollner I, Kozlowski PB, Glazman S, Miri S. α-Synuclein in the inner retina in Parkinson disease. Ann Neurol. 2014; 75(6): 964–966. [CrossRef] [PubMed]
Hart de Ruyter FJ, Morrema THJ, den Haan J, et al. α-Synuclein pathology in post-mortem retina and optic nerve is specific for α-synucleinopathies. NPJ Parkinsons Dis. 2023; 9(1): 124. [CrossRef] [PubMed]
Ahn J, Lee JY, Kim TW, et al. Retinal thinning associates with nigral dopaminergic loss in de novo Parkinson disease. Neurology. 2018; 91(11): e1003–e1012. [CrossRef] [PubMed]
Paul G, Elabi OF. Microvascular changes in Parkinson's disease- focus on the neurovascular unit. Front Aging Neurosci. 2022; 14: 853372. [CrossRef] [PubMed]
Guan J, Pavlovic D, Dalkie N, et al. Vascular degeneration in Parkinson's disease. Brain Pathol. 2013; 23(2): 154–164. [CrossRef] [PubMed]
Wong DYL, Lam MC, Ran A, Cheung CY. Artificial intelligence in retinal imaging for cardiovascular disease prediction: current trends and future directions. Curr Opin Ophthalmol. 2022; 33(5): 440–446. [CrossRef] [PubMed]
Vujosevic S, Aldington SJ, Silva P, et al. Screening for diabetic retinopathy: new perspectives and challenges. Lancet Diabetes Endocrinol. 2020; 8(4): 337–347. [CrossRef] [PubMed]
Ruamviboonsuk P, Tiwari R, Sayres R, et al. Real-time diabetic retinopathy screening by deep learning in a multisite national screening programme: a prospective interventional cohort study. Lancet Digit Health. 2022; 4(4): e235–e244. [CrossRef] [PubMed]
Grzybowski A, Singhanetr P, Nanegrungsunk O, Ruamviboonsuk P. Artificial intelligence for diabetic retinopathy screening using color retinal photographs: from development to deployment. Ophthalmol Ther. 2023; 12(3): 1419–1437. [CrossRef] [PubMed]
Wu JH, Liu TYA, Hsu WT, Ho JH, Lee CC. Performance and limitation of machine learning algorithms for diabetic retinopathy screening: meta-analysis. J Med Internet Res. 2021; 23(7): e23863. [CrossRef] [PubMed]
Arcadu F, Benmansour F, Maunz A, Willis J, Haskova Z, Prunotto M. Deep learning algorithm predicts diabetic retinopathy progression in individual patients. NPJ Digit Med. 2019; 2(1): 92. [CrossRef] [PubMed]
Lee T, Rivera A, Brune M, et al. Convolutional neural network–based automated quality assessment of OCT and OCT angiography image maps in individuals with neurodegenerative disease. Transl Vis Sci Technol. 2023; 12(6): 30. [CrossRef]
Figure 1.
 
Illustration of the PD classification model architecture. The model leverages a pretrained VGG19 CNN as a shared-weight image feature extractor (shown in the blue dotted box) to generate feature maps for each imaging modality, then feeds the results to modality-specific functions. The model averages the outputs from these functions and converts the result into a PD probability score in the [0, 1] range to obtain a final prediction result.
Figure 1.
 
Illustration of the PD classification model architecture. The model leverages a pretrained VGG19 CNN as a shared-weight image feature extractor (shown in the blue dotted box) to generate feature maps for each imaging modality, then feeds the results to modality-specific functions. The model averages the outputs from these functions and converts the result into a PD probability score in the [0, 1] range to obtain a final prediction result.
Figure 2.
 
Model outputs for each eye in the PD and control test sets evaluated by the right-eye-only, left-eye-only, and both eyes models. The red dashed line indicates the decision boundary of the model for its output probabilities. Less overlap between the box plots for the control and PD groups indicates a model had better classification performance, and points on the box plot closer to the threshold (red dashed line) represent patients that the model classified with lower confidence.
Figure 2.
 
Model outputs for each eye in the PD and control test sets evaluated by the right-eye-only, left-eye-only, and both eyes models. The red dashed line indicates the decision boundary of the model for its output probabilities. Less overlap between the box plots for the control and PD groups indicates a model had better classification performance, and points on the box plot closer to the threshold (red dashed line) represent patients that the model classified with lower confidence.
Figure 3.
 
ROC curve for the right-eye-only, left-eye-only, and both eyes models using all features on the test set.
Figure 3.
 
ROC curve for the right-eye-only, left-eye-only, and both eyes models using all features on the test set.
Figure 4.
 
Confusion matrices for the right eye only (A), left eye only (B), and both right and left eyes (C).
Figure 4.
 
Confusion matrices for the right eye only (A), left eye only (B), and both right and left eyes (C).
Figure 5.
 
Precision–recall curves for each of the three models incorporating all image types. The AUC value is 0.405 for the model utilizing both eyes, 0.535 for the model utilizing right eyes, and 0.677 for the model utilizing left eyes only.
Figure 5.
 
Precision–recall curves for each of the three models incorporating all image types. The AUC value is 0.405 for the model utilizing both eyes, 0.535 for the model utilizing right eyes, and 0.677 for the model utilizing left eyes only.
Table.
 
AUC on the Test Set Describes the Performance of Each Model in Predicting a Symptomatic PD Diagnosis Given the Imaging Inputs Described in the Left Column
Table.
 
AUC on the Test Set Describes the Performance of Each Model in Predicting a Symptomatic PD Diagnosis Given the Imaging Inputs Described in the Left Column
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×