September 2022
Volume 11, Issue 9
Open Access
Artificial Intelligence  |   September 2022
Feasibility of Automated Segmentation of Pigmented Choroidal Lesions in OCT Data With Deep Learning
Author Affiliations & Notes
  • Philippe Valmaggia
    Department of Biomedical Engineering, University of Basel, Allschwil, Switzerland
    Institute of Molecular and Clinical Ophthalmology Basel (IOB), Basel, Switzerland
    Department of Ophthalmology, University Hospital Basel, Basel, Switzerland
  • Philipp Friedli
    Supercomputing Systems AG, Zürich, Switzerland
  • Beat Hörmann
    Supercomputing Systems AG, Zürich, Switzerland
  • Pascal Kaiser
    Supercomputing Systems AG, Zürich, Switzerland
  • Hendrik P. N. Scholl
    Institute of Molecular and Clinical Ophthalmology Basel (IOB), Basel, Switzerland
    Department of Ophthalmology, University Hospital Basel, Basel, Switzerland
  • Philippe C. Cattin
    Department of Biomedical Engineering, University of Basel, Allschwil, Switzerland
  • Robin Sandkühler
    Department of Biomedical Engineering, University of Basel, Allschwil, Switzerland
  • Peter M. Maloca
    Institute of Molecular and Clinical Ophthalmology Basel (IOB), Basel, Switzerland
    Department of Ophthalmology, University Hospital Basel, Basel, Switzerland
    Moorfields Eye Hospital NHS Foundation Trust, London, EC1V 2PD, UK
  • Correspondence: Philippe Valmaggia, Department of Biomedical Engineering, University of Basel, 4123 Allschwil, Switzerland. e-mail: philippe.valmaggia@unibas.ch 
  • Footnotes
    *  RS and PMM contributed equally to this work.
Translational Vision Science & Technology September 2022, Vol.11, 25. doi:https://doi.org/10.1167/tvst.11.9.25
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      Philippe Valmaggia, Philipp Friedli, Beat Hörmann, Pascal Kaiser, Hendrik P. N. Scholl, Philippe C. Cattin, Robin Sandkühler, Peter M. Maloca; Feasibility of Automated Segmentation of Pigmented Choroidal Lesions in OCT Data With Deep Learning. Trans. Vis. Sci. Tech. 2022;11(9):25. https://doi.org/10.1167/tvst.11.9.25.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose: To evaluate the feasibility of automated segmentation of pigmented choroidal lesions (PCLs) in optical coherence tomography (OCT) data and compare the performance of different deep neural networks.

Methods: Swept-source OCT image volumes were annotated pixel-wise for PCLs and background. Three deep neural network architectures were applied to the data: the multi-dimensional gated recurrent units (MD-GRU), the V-Net, and the nnU-Net. The nnU-Net was used to compare the performance of two-dimensional (2D) versus three-dimensional (3D) predictions.

Results: A total of 121 OCT volumes were analyzed (100 normal and 21 PCLs). Automated PCL segmentations were successful with all neural networks. The 3D nnU-Net predictions showed the highest recall with a mean of 0.77 ± 0.22 (MD-GRU, 0.60 ± 0.31; V-Net, 0.61 ± 0.25). The 3D nnU-Net predicted PCLs with a Dice coefficient of 0.78 ± 0.13, outperforming MD-GRU (0.62 ± 0.23) and V-Net (0.59 ± 0.24). The smallest distance to the manual annotation was found using 3D nnU-Net with a mean maximum Hausdorff distance of 315 ± 172 µm (MD-GRU, 1542 ± 1169 µm; V-Net, 2408 ± 1060 µm). The 3D nnU-Net showed a superior performance compared with stacked 2D predictions.

Conclusions: The feasibility of automated deep learning segmentation of PCLs was demonstrated in OCT data. The neural network architecture had a relevant impact on PCL predictions.

Translational Relevance: This work serves as proof of concept for segmentations of choroidal pathologies in volumetric OCT data; improvements are conceivable to meet clinical demands for the diagnosis, monitoring, and treatment evaluation of PCLs.

Introduction
The development of novel deep learning methods has shown substantial improvements for many applications in medical image analysis, such as the automated segmentation of various lesions on the pixel level. The aim of image segmentation is to identify target structures at a pixel level and separate them from the remaining image. In the medical field, image segmentation is used for physiological as well as pathological structures such as tumours.13 
Optical coherence tomography (OCT) is based on the reflective pattern of laser beams and allows for the imaging of ocular structures at the micrometer scale.4 OCT has been used increasingly for the automated segmentation of structures, such as the ocular compartments, retinal layers, or fluids.57 Using the vast amount of OCT data, deep learning has facilitated automated segmentation of OCT images at the pixel level for retinal pathologies such as age-related macular degeneration, diabetic retinopathy, and glaucoma.810 
In contrast to the very abundant deep learning literature concerning the retina, there have been few reports on the choroid.1114 Pigmented choroidal lesions (PCLs) are a type of choroidal lesion that can be benign (such as nevi) or malignant (such as melanoma).15 They are often found incidentally during ophthalmic examination and further characterized using ultrasound examination, fluorescein angiography, or OCT.16,17 OCT is currently not the main imaging modality to investigate PCLs but can be used to quantify their volumes. Its micrometer resolution allows very precise measurements, which could be helpful for PCL monitoring, especially in the case of small lesions.16 However, manual annotations for the volumetric quantification of lesions are time-consuming and their quantification could be facilitated by automated segmentation. To the best of our knowledge, there is no report on the automated segmentation of PCLs in OCT data. Therefore, this proof-of-concept study assessed such automated segmentation using three different deep neural networks, named multi-dimensional gated recurrent units (MD-GRU), V-Net, and nnU-Net.1820 
Methods
Data Acquisition
Existing data from a previous study were used to compare the segmentation of PCLs.21 Thus, the current study was performed in accordance with the ethical standards of the Helsinki Declaration and was approved by the Ethics Committee Nordwest- und Zentralschweiz, Switzerland (ID: EKNZ UBE-15/89 and EKNZ UBE-l5/72). Written informed consent was obtained from all subjects. 
All subjects were examined via slit-lamp examination and a clinical fundus inspection before OCT acquisition. Refractive tests without axial length measurements were performed. In the absence of pathological retinal changes, the subject was included in the normal group. The normal eyes were included as a negative control group for the training and testing of the deep learning neural networks. The inclusion criterion for the PCL group was clinical visibility of a melanocytic and relatively well-demarcated lesion that could be recorded entirely with the OCT imaging protocol. There was no exclusion based on the thickness, location, or reflectivity of the PCLs. Exclusion criteria were the absence or inability to consent, presence of trauma, ocular inflammation, poor visibility of the fundus, retinal pathologies such as retinal detachment or subretinal fluid, marked gliosis, presence of a macular hole, vascular occlusion, and inability to steadily fixate. 
The data for this study were acquired with a standard swept-source OCT device without dilated pupils (Triton DRI SS-OCT, Topcon, Tokyo, Japan) and comprised optical volumes, each measuring 6 mm × 6 mm × 2.6 mm (256 B-scans, 512 × 992 pixels). 
Manual Annotation of PCLs
Two independent human graders were responsible for the manual annotations. PM is a trained ophthalmologist with 25 years of clinical experience and 20 years of experience with OCT systems; PV is a medical student with 2 years of work experience with OCT data. To generate the labels for supervised learning, the OCT volumes were manually annotated with the open source FIJI imaging software v2.1.22 The graders visually identified the borders of each PCL using consequential OCT B-scans. The anterior border was delineated at the Bruch's membrane, and the lateral borders were identified in the change of texture between the choroidal tissue and the more homogeneous and hyperreflective PCLs. The posterior border was generated via an interpolation between the choroidal–scleral borders of the adjacent healthy tissue in case the posterior border was not visualizable through the shadow behind the PCL. Finally, each outlined lesion was filled and highlighted with maximum intensity, and the rest of the image was set as a dark background. The manual annotation process is presented in Figure 1
Figure 1.
 
Manual annotation process for PCL segmentation in volumetric OCT data. (a) Outline of the PCL in a single B-scan. (b) Filling of the outline to generate a highlighted PCL. (c) Binary label creation through clearing of the background. (d) Assemblage of the binary labels for each B-scan in the volume produced a 3D label of the PCL.
Figure 1.
 
Manual annotation process for PCL segmentation in volumetric OCT data. (a) Outline of the PCL in a single B-scan. (b) Filling of the outline to generate a highlighted PCL. (c) Binary label creation through clearing of the background. (d) Assemblage of the binary labels for each B-scan in the volume produced a 3D label of the PCL.
To test the reproducibility of the method, a test volume consisting of 30 B-scans containing a PCL was annotated three times by each grader; then, the obtained volumes were compared for repeatability using the Dice coefficient. For intragrader reproducibility, the Dice coefficient was calculated between each of the annotations of a single grader (three comparisons per grader); for intergrader reproducibility, the Dice coefficient was calculated between all volumes of different graders (nine comparisons in total). 
After stacking the generated annotations, the PCL label was smoothed with 3D Slicer v4.11 (The Slicer Community, Boston, MA, USA), preserving the same dimensions as the original volume.23 Hence, for each voxel in the original volume, the corresponding generated label stack indicated the location of the PCL. To optimize the data for neural network training, the volumes were down-sampled to one-half the size in each dimension (128 × 256 × 496 pixels). 
Training of the Neural Networks
Three different neural networks were trained with the data: MD-GRU, V-Net, and nnU-Net. The code for each neural network is publicly available on GitHub.2426 The MD-GRU was developed at the Department of Biomedical Engineering at the University of Basel.27,28 It was developed specifically for volumetric image segmentation and initially evaluated on tumors as well, namely, for magnetic resonance imaging (MRI)-derived brain data. The V-Net was developed at the Technical University of Munich, also specifically for the purpose of volumetric image segmentation.19 The hyperparameters of the MD-GRU and V-Net models were determined empirically on a segmentation dataset containing brain tumours.2 The nnU-Net was developed at the German Cancer Research Center to deal with the vast differences found in medical datasets. It provides a U-Net–based neural network that self-configures according to the data it is presented with based on several heuristics, such as the image size, the dataset size, and the available GPU.20 nnU-Net was used to compare the performance of two-dimensional (2D) versus three-dimensional (3D) predictions. A summary of the specific neural network configurations is shown in Table 1. Training and testing of the neural networks were performed on an NVIDIA Tesla V100 DGXS 16 GB or a NVIDIA Titan RTX 24 GB (Nvidia, Santa Clara, US). 
Table 1.
 
Neural Network Configuration Parameters for the 2D and 3D Neural Networks
Table 1.
 
Neural Network Configuration Parameters for the 2D and 3D Neural Networks
The dataset was split into myopic (spherical equivalent <−1.0 diopter [D]), emmetropic (≥−1.0 D and ≤+1.0 D), hyperopic (>+1.0 D), and PCL-containing OCT stacks. These data were then randomly assigned to 10 different equally sized folds. This was done to guarantee that the single folds contained a balanced amount of data from each group, and, in particular, a similar amount of PCL data. Neural networks were then trained and tested with k-fold cross-validation on all folds. For each fold, k–1 subsets of the data were used for training, and one was used for testing. The training for each fold was restarted from scratch. The results of all the test subsets were then evaluated as shown in Figure 2b. This practice ensured that none of the training data were used to evaluate the neural network models. 
Figure 2.
 
Image processing pipeline for PCL segmentation. (a) OCT data with their corresponding labels were loaded into different deep neural networks (MD-GRU, V-Net, and nnU-Net). (b) Training and testing of the neural networks were performed using k-fold cross-validation with training from scratch for each fold. (c) The resulting lesion predictions are displayed in blue, red, green, and yellow according to each neural network.
Figure 2.
 
Image processing pipeline for PCL segmentation. (a) OCT data with their corresponding labels were loaded into different deep neural networks (MD-GRU, V-Net, and nnU-Net). (b) Training and testing of the neural networks were performed using k-fold cross-validation with training from scratch for each fold. (c) The resulting lesion predictions are displayed in blue, red, green, and yellow according to each neural network.
Data Analyses
The evaluation of the predicted labels was conducted with statistical, similarity, and distance measures using common metrics.29 The statistical methods analyzed the parameters according to the confusion matrix between the prediction and the manual annotations. An overview of the formulas used for the evaluation can be found in Supplementary Material 1
Accuracy was assessed for each neural network. Accuracy is usually a well-established parameter for deep learning; however, in the current study, there was a major class imbalance between the PCL and background voxels. In the present study, a total of 1,966,604,288 matched voxels were present, of which 1,599,136 voxels represented manually annotated PCLs. Hence, if a neural network had predicted all the voxels as background, the accuracy would still have been more than 0.999. This is due to the class imbalance between PCLs and the background data. 
A more appropriate evaluation of the neural networks was achieved with the statistical measures recall and precision. Recall is an indicator of the ratio of the relevant data points predicted compared with the total amount in the manual annotation, and precision indicates the ratio of predicted data points that are relevant.30 The similarity between the manual annotation and the prediction was assessed with the Dice coefficient. The Hausdorff distance was used to calculate the distance between the two sets, where the maximum and the 95th percentile of the Hausdorff distance are presented. The summarized data are presented as mean ± standard deviation over the cross-validation samples. The programming, configuration of the neural networks, and evaluations were done with Unix shell scripts and Python v.3.6 (Python Software Foundation, Wilmington, DE, USA). The data were visualized with Python v.3.6 and 3DSlicer v4.11.31 
Results
In total, 121 labelled OCT volumes from 71 subjects were included in the study. Twenty-one eyes from 21 subjects contributed a PCL, of which 7 were located within the macula. Fifty eyes contributed 100 normal volume OCTs. The normal group consisted of 34 emmetropic eyes, 32 myopic eyes, and 34 hyperopic eyes. The annotations on the test volume showed good reproducibility of the method with a mean intragrader Dice coefficient of 0.92 ± 0.02 and a mean intergrader Dice coefficient of 0.88 ± 0.02. 
Quantitative Analysis
All neural networks generated automated PCL segmentation. The 3D nnU-Net generated the best predictions regarding the mentioned evaluation criteria. The recall of the 3D nnU-Net was 0.77 ± 0.22, and the precision was 0.59 ± 0.38. Regarding the precision, the nnU-Net2D trained on only 2D scans generated the best results at 0.69 ± 0.32. The mean Dice coefficient for the predictions of the 3D nnU-Net was 0.78 ± 0.13, outperforming MD-GRU (0.62 ± 0.23) and V-Net (0.59 ± 0.24). This Dice coefficient indicates that there is a major overlap between the predicted and manually annotated PCLs. Furthermore, the predicted PCLs were of similar size compared with the manual annotations. The 3D version of the nnU-Net performed best in terms of distance metrics, with a maximal Hausdorff distance of 315 ± 172 µm (MD-GRU, 1542 ± 1169 µm; V-Net, 2408 ± 1060 µm). In a comparison of models trained on 2D and 3D, the 3D version of the nnU-Net was superior to the 2D version regarding nearly every metric. A detailed summary of the evaluation parameters for 2D and 3D is presented in Table 2
Table 2.
 
Evaluation Parameters for the 2D and 3D Neural Networks
Table 2.
 
Evaluation Parameters for the 2D and 3D Neural Networks
Qualitative Analysis
This superior performance of the volumetric neural networks is visualized intuitively in Figure 3a and 3b. The 3D PCL predictions were more coherent than the split predictions for the 2D nnU-Net. On single B-scans, the predictions for all the neural networks were often very similar (Figure 3c). Yet when analyzing the entire C-Scan, nnU-Net2D missed PCLs in single B-scans, even though the lesion was identified in adjacent B-scans. An example of missed lesions in B-scans by the 2D neural network in the PCL core is illustrated in Figure 3d. These missing predictions are due to the input into the neural network where the spatial volume context is lost when only training with isolated 2D images. 
Figure 3.
 
Visualisations of the neural network predictions with overlays on the OCT images and the manual annotations. (a) Volume-rendered retinal and choroidal compartments. (b) Three-dimensional manual annotations and model predictions. (c) Two-dimensional OCT images and predictions as overlays. (d) Enlarged 2D manual annotations and predictions as overlays.
Figure 3.
 
Visualisations of the neural network predictions with overlays on the OCT images and the manual annotations. (a) Volume-rendered retinal and choroidal compartments. (b) Three-dimensional manual annotations and model predictions. (c) Two-dimensional OCT images and predictions as overlays. (d) Enlarged 2D manual annotations and predictions as overlays.
The visualizations provide further insight into the segmentation performance of the evaluated neural networks. Overall, the 3D nnU-Net provided the most condensed segmentations of PCLs (Figure 3). Three additional examples are included in Figure 4, where the generated predictions are presented with the individual Dice coefficient and the maximum Hausdorff distance. As can be seen, the predicted shapes are centrally located in the manual annotation; the borders correspond with the manual annotation, slightly surpassing or missing the border. The data show that different shapes and locations of PCLs can be predicted, be it a peripheral longitudinal PCL (Fig. 4a), a tortuous PCL on the posterior pole (Fig. 4b), or a small PCL between the optic nerve head and the macula (Fig. 4c). 
Figure 4.
 
Example automated PCL segmentations generated using 3D nnU-Net. PCL predictions are shown as a green overlay, and manual annotations in are shown in white. (a, d, g) Volume rendered retinal and choroidal compartments. Predicted PCL with corresponding Dice coefficient and maximum Hausdorff distance. (b, e, h) 3D predictions overlaid over the manual annotations. (c, f, i) Enlarged 2D manual annotations with overlaid predictions. Axial stretching was performed for a better visualization according to isometric pixels.
Figure 4.
 
Example automated PCL segmentations generated using 3D nnU-Net. PCL predictions are shown as a green overlay, and manual annotations in are shown in white. (a, d, g) Volume rendered retinal and choroidal compartments. Predicted PCL with corresponding Dice coefficient and maximum Hausdorff distance. (b, e, h) 3D predictions overlaid over the manual annotations. (c, f, i) Enlarged 2D manual annotations with overlaid predictions. Axial stretching was performed for a better visualization according to isometric pixels.
Discussion
A PCL is common and typically identified as a brownish, darkly pigmented lesion in the course of a routine clinical investigation of the fundus.32,33 Although the majority of PCLs are benign choroidal nevi, malignant transformation is still possible and is highly difficult to detect at an early stage, because tissue sampling for diagnosis is not possible without damaging healthy tissue.34,35 In this context, OCT may represent a possible imaging tool to enable early PCL diagnosis at a micrometer level. It could be used to guide the quantitative monitoring of PCLs and support treatment decisions, particularly if it is enhanced with deep learning tumor segmentation.36 
This study presents the first successful automated segmentation of PCLs in OCT data. In addition, it compares the segmentation performance of three different deep learning architectures, MD-GRU, V-Net, and nnU-Net. Overall, the nnU-Net trained on 3D data showed the best performance, outperforming the other models on the key metrics of recall, Dice coefficient, and maximum Hausdorff distance. 
Recall, also known as sensitivity, punishes large numbers of false negatives in the prediction. Recall was chosen as the key statistical metric, because it is crucial not to miss anomalous findings. If the analyzed PCL in future evaluations was malignant, it would be even more important not to miss parts of the PCL. The best-achieved recall score of 0.77 ± 0.22 was achieved with the 3D nnU-Net, which means that the algorithm detected around three-quarters of all lesion voxels. 
Further, the shape and position of the predicted PCL should correspond with the actual underlying tumor. For the similarity evaluation, the Dice coefficient was used, which is a commonly used key metric in image segmentation.20 The resulting values of the 3D nnU-Net model with a mean Dice coefficient 0.78 could be used as an orientation when comparing future PCL segmentations. Although the Dice coefficients for different tasks are difficult to compare, the presented model generated predictions in the same similarity range as neural networks in the RETOUCH challenge. This challenge investigated the segmentation of different fluids in retinal OCT images where the best-performing team achieved a mean Dice coefficient of 0.77, and the average of the participating teams was 0.68.7,37 It is also notable that PCL segmentation is thought to be a harder task than retinal fluid segmentation, because the borders are more diffuse for PCLs than retinal fluids in OCT images. 
The deviation between the manual annotation and the prediction can be visualized and qualitatively evaluated in 3D as shown in Figure 4. The mismatch can best be quantified with the Hausdorff distance. The Hausdorff distance is a good measure for translating the pixel domain into real-world measures. For example, with the relatively large voxel sizes in clinical MRI, a single-pixel deviation means that a deviation of more than 1 mm would be induced. A small distance between predictions and real localization is particularly important for the radiotherapy guidance of malignant melanoma.38 With regard to the OCT pixel domain, a deviation of a single pixel corresponds only to a deviation in the micrometer range. The proposed 3D nnU-Net prediction had an average maximum Hausdorff distance of 315 ± 172 µm. 
The Dice coefficient as a sole criterion can be misleading in ocular lesions. As shown in Figure 4, PCLs 4a and 4c had an almost identical Dice coefficient, but the segmentation quality could be further discriminated by the Hausdorff distance in µm. Conversion from pixel distance to micrometers was performed by readjusting to the original voxel sizes. This measure was relevant, as it is possible that the predicted segmentation was correct for a large part of the lesion and had a good score on similarity but omitted a section of the PCL (such as a spur). 
The implemented neural network configuration was important; major deviations between the different neural network architectures were found. Although MD-GRU and V-Net were trained with parameters that worked well for segmentation tasks, the self-configuring nnU-Net framework performed best. The nnU-Net is based on a heuristic approach, which determines the neural network configuration based on the input data characteristics and the available hardware.20 Because the settings from the heuristic approach might be further optimizable, this shows that using state-of-the-art neural network architectures is relevant for the tasks. Further, in this study, the 3D model outperformed the 2D model, which produced some nonphysiological predictions with gaps on single B-scans. As all corporal structures are volumetric, 3D segmentation models should be selected when feasible. 
This work could have translational relevance for patients with PCL because the obtained segmentations allow for a novel kind of quantification of the PCL volume and surface area. Clinically, it is not always evident if such a lesion has a risk of malignant transformation.15 Using OCT-derived volume analysis, the evolution of potentially malignant lesions could be followed up quantitatively, and the automated segmentation of PCLs could support clinical decisions. The risk for metastasis could be monitored at a micrometer scale, compared with conventional ocular ultrasound imaging that can only characterize lesions at a millimeter scale at a high enough accuracy.3942 Therefore, automated segmentation of these small lesions with OCT could potentially enhance follow-up for patients and support an earlier diagnosis of malignant PCL transformation. 
Automated segmentation could facilitate the radiotherapy of malignant lesions; for this, the correct demarcation of the target tissue is essential. The image modalities used in a common proton beam radiotherapy planning include fundus photographs, ultrasound examination, orthogonal radiographs, and MRI.4345 In the future, the researchers assume that OCT imaging could be integrated in adjunction to other modalities into such therapy planning pipelines, because it provides volumetric information and a much higher resolution than other imaging modalities currently in use. Automatically segmented lesions could be highlighted and controlled by a clinician, which could ease or even eradicate the time-consuming process of manual annotations. By integrating automated deep learning segmentation into the data visualization pipeline, the structural understanding of a PCL could be facilitated for both clinicians and patients. 
In terms of limitations, the current work has a relatively low number of PCL cases that were not part of a screening population. The control group consisted of healthy participants without PCL masquerades. However, the main goal of this study was to investigate feasibility and compare different deep learning algorithms, which was successful. The obtained data were derived only from one OCT device without enhanced-depth imaging. However, the used swept-source OCT technology allowed the visualization of the choroid and ensured data homogeneity for this initial proof-of-concept study. To investigate the generalizability of automated PCL segmentation, future studies should test for domain adaptation with different acquisition protocols, diverse study populations and different OCT scanners. 
The intrasubject evolution of PCLs was not assessed and has yet to be tested to check for the feasibility of a quantitative long-term follow-up study. The quantitative monitoring of the lesions could compare automated segmentation against manual annotations, because, to our knowledge, the volumetric monitoring of PCLs with OCT has not yet been presented. Another limitation was that a histological examination of the PCLs was not possible without harming healthy tissue. Thus, the nature of the lesion could not be determined, and it remained unknown whether the lesions were choroidal nevi or if they were malignant tumors. However, because only relatively small, well-delineated PCLs were included, it can be assumed that the majority were choroidal nevi. Furthermore, human annotation of PCLs may show deviations; future manually annotated data from different graders should be included in studies going beyond the present proof-of-concept study. 
Improving the automated segmentation could be achieved by modifying the neural networks. The three different neural networks used here were trained with the best-performing hyperparameters, but they were not further optimized for this task. In further optimized configurations, the class imbalance of the PCL pixel and the background pixel could be taken into account during training for further improvements.46 As the self-configuring nnU-Net performed the best, optimization for other network architectures seems feasible. Such improvements and further validation are needed to meet clinical demands for the diagnosis, monitoring, and treatment evaluation of PCLs. 
Conclusions
This proof-of-concept study showed that automated segmentation of PCLs in OCT data with deep learning is feasible. The neural network architecture had a relevant impact on PCL predictions. Overall, the 3D nnU-Net showed the best performance among the evaluated models. Automated deep learning segmentation could be integrated into a volumetric visualization pipeline and facilitate structural understanding for both clinicians and patients. 
Acknowledgments
The authors thank the Domarena Foundation, Meggen, Switzerland, and Hirslanden Gruppe, Luzern, Switzerland, for their financial support and Thomas Sammer, Mediconsult, Switzerland, for technical support for this study. The researchers also thank all the participants who volunteered for this study. 
Disclosure: P. Valmaggia, Swiss National Science Foundation (Grant 323530_199395) (F); P. Friedli, None; B. Hörmann, Supercomputing Systems AG, Zurich, Switzerland (E); P. Kaiser, Supercomputing Systems AG, Zurich, Switzerland (E); H.P.N. Scholl, Swiss National Science Foundation (Project funding: “Developing novel outcomes for clinical trials in Stargardt disease using structure/function relationship and deep learning” #310030_201165, and National Center of Competence in Research Molecular Systems Engineering: “NCCR MSE: Molecular Systems Engineering (phase II)” #51NF40-182895) (F), the Wellcome Trust (PINNACLE study) (F), and the Foundation Fighting Blindness Clinical Research Institute (ProgStar study) (F), Astellas Pharma Global Development, Inc./Astellas Institute for Regenerative Medicine (S), Boehringer Ingelheim Pharma GmbH & Co (S), Gyroscope Therapeutics Ltd. (S), Janssen Research & Development, LLC (Johnson & Johnson) (S), Novartis Pharma AG (CORE) (S), Okuvision GmbH (S), and Third Rock Ventures, LLC (S), Gerson Lehrman Group (C), Guidepoint Global, LLC (C), and Tenpoint Therapeutics Limited (C), Data Monitoring and Safety Board/Committee of Belite Bio (CT2019-CTN-04690-1), ReNeuron Group Plc/Ora Inc. (NCT02464436), F. Hoffmann-La Roche Ltd (VELODROME trial, NCT04657289; DIAGRID trial, NCT05126966) and member of the Steering Committee of Novo Nordisk (FOCUS trial; NCT03811561). All arrangements have been reviewed and approved by the University of Basel (Universitätsspital Basel, USB) and the Board of Directors of the Institute of Molecular and Clinical Ophthalmology Basel (IOB) in accordance with their conflict of interest policies. Compensation is being negotiated and administered as grants by USB, which receives them on its proper accounts. Funding organizations had no influence on the design, performance or evaluation of the current study (N); P.C. Cattin, (N); R. Sandkühler, (N); P.M. Maloca, Roche (C), MIMO AG and VisionAI, Switzerland (O) 
References
Shen D, Wu G, Suk H-I. Deep learning in medical image analysis. Annu Rev Biomed Eng. 2017; 19: 221–248, doi:10.1146/annurev-bioeng-071516-044442. [CrossRef] [PubMed]
Ghaffari M, Sowmya A, Oliver R. Automated brain tumor segmentation using multimodal brain scans: A survey based on models submitted to the BraTS 2012-2018 Challenges. IEEE Rev Biomed Eng. 2020; 13: 156–168, doi:10.1109/RBME.2019.2946868. [CrossRef] [PubMed]
Ting DSW, Peng L, Varadarajan AV, et al. Deep learning in ophthalmology: The technical and clinical considerations. Prog Retin Eye Res. 2019; 72: 100759. doi:10.1016/j.preteyeres.2019.04.003. [CrossRef] [PubMed]
Huang D, Swanson E, Lin C, et al. Optical coherence tomography. Science. 1991; 254(5035): 1178–1181, doi:10.1126/science.1957169. [CrossRef] [PubMed]
Maloca PM, Lee AY, De Carvalho ER, et al. Validation of automated artificial intelligence segmentation of optical coherence tomography images. PLoS One. 2019; 14(8): e0220063. doi:10.1371/journal.pone.0220063. [CrossRef] [PubMed]
Montuoro A, Waldstein SM, Gerendas BS, et al. Joint retinal layer and fluid segmentation in OCT scans of eyes with severe macular edema using unsupervised representation and auto-context. Biomed Opt Express. 2017; 8(3): 1874. doi:10.1364/boe.8.001874. [CrossRef] [PubMed]
Bogunovic H, Venhuizen F, Klimscha S, et al. RETOUCH: The retinal OCT fluid detection and segmentation benchmark and challenge. IEEE Trans Med Imaging. 2019; 38(8): 1858–1874, doi:10.1109/TMI.2019.2901398. [CrossRef] [PubMed]
Yan Q, Weeks DE, Xin H, et al. Deep-learning-based prediction of late age-related macular degeneration progression. Nat Mach Intell. 2020; 2(2): 141–150, doi:10.1038/s42256-020-0154-9. [CrossRef] [PubMed]
Ting DSW, Cheung CY-L, Lim G, et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA. 2017; 318(22): 2211–2223, doi:10.1001/jama.2017.18152. [CrossRef] [PubMed]
Ran AR, Tham CC, Chan PP, et al. Deep learning in glaucoma with optical coherence tomography: A review. Eye (Lond). 2021; 35(1): 188–201, doi:10.1038/s41433-020-01191-5. [CrossRef] [PubMed]
Muller J, Alonso-Caneiro D, Read SA, et al. Application of deep learning methods for binarization of the choroid in optical coherence tomography images. Transl Vis Sci Technol. 2022; 11(2): 23. doi:10.1167/tvst.11.2.23. [CrossRef] [PubMed]
Hsia WP, Tse SL, Chang CJ, et al. Automatic segmentation of choroid layer using deep learning on spectral domain optical coherence tomography. Appl Sci. 2021; 11(12): 5488. doi:10.3390/app11125488. [CrossRef]
Masood S, Fang R, Li P, et al. Automatic choroid layer segmentation from optical coherence tomography images using deep learning. Sci Rep. 2019; 9(1): 3058. doi:10.1038/s41598-019-39795-x. [CrossRef] [PubMed]
Kang NY, Ra H, Lee K, et al. Classification of pachychoroid on optical coherence tomography using deep learning. Graefes Arch Klin Exp Ophthalmol. 2021; 259(7): 1803–1809, doi:10.1007/s00417-021-05104-4. [CrossRef]
Shields CL, Manalac J, Das C, et al. Choroidal melanoma: Clinical features, classification, and top 10 pseudomelanomas. Curr Opin Ophthalmol. 2014; 25(3): 177–185, doi:10.1097/ICU.0000000000000041. [CrossRef] [PubMed]
Maloca P, Gyger C, Hasler PW. A pilot study to compartmentalize small melanocytic choroidal tumors and choroidal vessels with speckle-noise free 1050 nm swept source optical coherence tomography (OCT choroidal “tumoropsy”). Graefes Arch Klin Exp Ophthalmol. 2016; 254(6): 1211–1219, doi:10.1007/s00417-016-3270-9. [CrossRef]
Chien JL, Sioufi K, Surakiatchanukul T, et al. Choroidal nevus. Curr Opin Ophthalmol. 2017; 28(3): 228–237, doi:10.1097/icu.0000000000000361. [CrossRef] [PubMed]
Andermatt S, Pezold S, Cattin P. Multi-dimensional gated recurrent units for the segmentation of biomedical 3D-data. Deep Learn Data Labeling Med Appl Lect Notes Comput Sci. 2016: 142–151, doi:10.1007/978-3-319-46976-8_15.
Milletari F, Navab N, Ahmadi SA. V-Net: Fully convolutional neural networks for volumetric medical image segmentation. In: Proceedings of the 2016 Fourth International Conference on 3D Imaging, Modeling, Processing, Visualization and Transmission (3DIMPVT). Stanford, CA, December 19, 2016. Published online 2016: 565–571, doi:10.1109/3DV.2016.79.
Isensee F, Jaeger PF, Kohl SAA, et al. nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nat Methods. 2021; 18(2): 203–211, doi:10.1038/s41592-020-01008-z. [CrossRef] [PubMed]
Maloca P, Gyger C, Schoetzau A, et al. Inter-device size variation of small choroidal nevi measured using stereographic projection ultra-widefield imaging and optical coherence tomography. Graefes Arch Klin Exp Ophthalmol. 2016; 254(4): 797–808, doi:10.1007/s00417-015-3209-6. [CrossRef]
Schindelin J, Arganda-Carreras I, Frise E, et al. Fiji: An open-source platform for biological-image analysis. Nat Methods. 2012; 9(7): 676–682, doi:10.1038/nmeth.2019. [CrossRef] [PubMed]
Fedorov A, Beichel R, Kalpathy-Cramer J, et al. 3D Slicer as an image computing platform for the quantitative imaging network. Magn Reson Imaging. 2012; 30(9): 1323–1341, doi:10.1016/j.mri.2012.05.001. [CrossRef] [PubMed]
Andermatt S . MD-GRU. Published 2021. Available at: https://github.com/zubata88/mdgru. Accessed May 31, 2021.
Long K . V-Net. Published 2021. Available at: https://github.com/jackyko1991/vnet-tensorflow. Accessed June 16, 2021.
Division of Medical Image Computing GCRC (DKFZ). nnUNet. Published 2021. Available at: https://github.com/MIC-DKFZ/nnunet. Accessed July 7, 2021.
Andermatt S, Pezold S, Amann M, et al. Multi-dimensional gated recurrent units for automated anatomical landmark localization. arXiv Prepr arXiv170802766. Published online 2017, https://arxiv.org/abs/1708.02766. Accessed February 10, 2022.
Andermatt S, Pezold S, Cattin PC. Multi-dimensional gated recurrent units for brain tumor segmentation. In: International MICCAI BraTS Challenge. Pre-Conference Proceedings. Quebec City, Canada. Published online 2017:15–19, https://www.cbica.upenn.edu/sbia/Spyridon.Bakas/MICCAI_BraTS/MICCAI_BraTS_2017_proceedings_shortPapers.pdf. Accessed February 10, 2022.
Taha AA, Hanbury A. Metrics for evaluating 3D medical image segmentation: Analysis, selection, and tool. BMC Med Imaging. 2015; 15(1): 29. doi:10.1186/s12880-015-0068-x. [CrossRef] [PubMed]
Buckland MK, Gey FC. The relationship between recall and precision. J Am Soc Inf Sci. 1994; 45: 12–19. [CrossRef]
Fedorov A, Beichel R, Kalpathy-Cramer J, et al. 3D Slicer as an image computing platform for the quantitative imaging network. Magn Reson Imaging. 2012; 30(9): 1323–1341, doi:10.1016/j.mri.2012.05.001. [CrossRef] [PubMed]
Shields CL, Mashayekhi A, Ho T, et al. Solitary congenital hypertrophy of the retinal pigment epithelium: Clinical features and frequency of enlargement in 330 patients. Ophthalmology. 2003; 110(10): 1968–1976, doi:10.1016/S0161-6420(03)00618-3. [CrossRef] [PubMed]
Shields CL, Furuta M, Mashayekhi A, et al. Clinical spectrum of choroidal nevi based on age at presentation in 3422 consecutive eyes. Ophthalmology. 2008; 115(3): 546–552.e2. doi:10.1016/j.ophtha.2007.07.009. [CrossRef] [PubMed]
Augsburger JJ, Corrêa ZM, Trichopoulos N, et al. Size overlap between benign melanocytic choroidal nevi and choroidal malignant melanomas. Invest Ophthalmol Vis Sci. 2008; 49(7): 2823–2828, doi:10.1167/iovs.07-1603. [CrossRef] [PubMed]
Singh AD, Kalyani P, Topham A. Estimating the risk of malignant transformation of a choroidal nevus. Ophthalmology. 2005; 112(10): 1784–1789, doi:10.1016/j.ophtha.2005.06.011. [CrossRef] [PubMed]
Damato EM, Damato BE. Detection and time to treatment of uveal melanoma in the United Kingdom: An evaluation of 2,384 patients. Ophthalmology. 2012; 119(8): 1582–1589, doi:10.1016/j.ophtha.2012.01.048. [CrossRef] [PubMed]
Lu D, Heisler M, Lee S, et al. Deep-learning based multiclass retinal fluid segmentation and detection in optical coherence tomography images using a fully convolutional neural network. Med Image Anal. 2019; 54: 100–110, doi:10.1016/j.media.2019.02.011. [CrossRef] [PubMed]
Stannard C, Sauerwein W, Maree G, et al. Radiotherapy for ocular tumours. Eye (Lond). 2013; 27: 119–127, doi:10.1038/eye.2012.241. [CrossRef] [PubMed]
Shields CL, Furuta M, Thangappan A, et al. Metastasis of uveal melanoma millimeter-by-millimeter in 8033 consecutive eyes. Arch Ophthalmol. 2009; 127(8): 989–998, doi:10.1001/archophthalmol.2009.208. [CrossRef] [PubMed]
Singh P, Singh A. Choroidal melanoma. Oman J Ophthalmol. 2012; 5(1): 3–9, doi:10.4103/0974-620X.94718. [CrossRef] [PubMed]
Cheung A, Scott I, Murray T, et al. Distinguishing a choroidal nevus from a choroidal melanoma. EyeNet Mag. 2012: 39–40.
Sobottka B, Schlote T, Krumpaszky HG, et al. Choroidal metastases and choroidal melanomas: Comparison of ultrasonographic findings. Br J Ophthalmol. 1998; 82(2): 159–161, doi:10.1136/bjo.82.2.159. [CrossRef] [PubMed]
Daftari IK, Mishra KK, O'Brien JM, et al. Fundus image fusion in EYEPLAN software: An evaluation of a novel technique for ocular melanoma radiation treatment planning. Med Phys. 2010; 37(10): 5199–5207, doi:10.1118/1.3488891. [CrossRef] [PubMed]
Hoffmann A, Oborn B, Moteabbed M, et al. MR-guided proton therapy: A review and a preview. Radiat Oncol. 2020; 15(1): 1–13, doi:10.1186/s13014-020-01571-x. [CrossRef]
MacKay RI. Image guidance for proton therapy. Clin Oncol. 2018; 30(5): 293–298, doi:10.1016/j.clon.2018.02.004. [CrossRef]
Eelbode T, Bertels J, Berman M, et al. Optimization for medical image segmentation: Theory and practice when evaluating with dice score or Jaccard index. IEEE Trans Med Imaging. 2020; 39(11): 3679–3690, doi:10.1109/TMI.2020.3002417le. [CrossRef] [PubMed]
Figure 1.
 
Manual annotation process for PCL segmentation in volumetric OCT data. (a) Outline of the PCL in a single B-scan. (b) Filling of the outline to generate a highlighted PCL. (c) Binary label creation through clearing of the background. (d) Assemblage of the binary labels for each B-scan in the volume produced a 3D label of the PCL.
Figure 1.
 
Manual annotation process for PCL segmentation in volumetric OCT data. (a) Outline of the PCL in a single B-scan. (b) Filling of the outline to generate a highlighted PCL. (c) Binary label creation through clearing of the background. (d) Assemblage of the binary labels for each B-scan in the volume produced a 3D label of the PCL.
Figure 2.
 
Image processing pipeline for PCL segmentation. (a) OCT data with their corresponding labels were loaded into different deep neural networks (MD-GRU, V-Net, and nnU-Net). (b) Training and testing of the neural networks were performed using k-fold cross-validation with training from scratch for each fold. (c) The resulting lesion predictions are displayed in blue, red, green, and yellow according to each neural network.
Figure 2.
 
Image processing pipeline for PCL segmentation. (a) OCT data with their corresponding labels were loaded into different deep neural networks (MD-GRU, V-Net, and nnU-Net). (b) Training and testing of the neural networks were performed using k-fold cross-validation with training from scratch for each fold. (c) The resulting lesion predictions are displayed in blue, red, green, and yellow according to each neural network.
Figure 3.
 
Visualisations of the neural network predictions with overlays on the OCT images and the manual annotations. (a) Volume-rendered retinal and choroidal compartments. (b) Three-dimensional manual annotations and model predictions. (c) Two-dimensional OCT images and predictions as overlays. (d) Enlarged 2D manual annotations and predictions as overlays.
Figure 3.
 
Visualisations of the neural network predictions with overlays on the OCT images and the manual annotations. (a) Volume-rendered retinal and choroidal compartments. (b) Three-dimensional manual annotations and model predictions. (c) Two-dimensional OCT images and predictions as overlays. (d) Enlarged 2D manual annotations and predictions as overlays.
Figure 4.
 
Example automated PCL segmentations generated using 3D nnU-Net. PCL predictions are shown as a green overlay, and manual annotations in are shown in white. (a, d, g) Volume rendered retinal and choroidal compartments. Predicted PCL with corresponding Dice coefficient and maximum Hausdorff distance. (b, e, h) 3D predictions overlaid over the manual annotations. (c, f, i) Enlarged 2D manual annotations with overlaid predictions. Axial stretching was performed for a better visualization according to isometric pixels.
Figure 4.
 
Example automated PCL segmentations generated using 3D nnU-Net. PCL predictions are shown as a green overlay, and manual annotations in are shown in white. (a, d, g) Volume rendered retinal and choroidal compartments. Predicted PCL with corresponding Dice coefficient and maximum Hausdorff distance. (b, e, h) 3D predictions overlaid over the manual annotations. (c, f, i) Enlarged 2D manual annotations with overlaid predictions. Axial stretching was performed for a better visualization according to isometric pixels.
Table 1.
 
Neural Network Configuration Parameters for the 2D and 3D Neural Networks
Table 1.
 
Neural Network Configuration Parameters for the 2D and 3D Neural Networks
Table 2.
 
Evaluation Parameters for the 2D and 3D Neural Networks
Table 2.
 
Evaluation Parameters for the 2D and 3D Neural Networks
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×