Figure 4 presents multiple input OCT B-scans (left column) from the testing dataset, the predicted RNFL segmentation results (middle column), and the ground truth segmentation results obtained through manual annotation (right column). The predicted segmentation results are very close to the ground truth despite the speckles in the input OCT images, thereby indicating that our segmentation model is very robust for minimizing the intrinsic noise in the OCT images. Moreover, the signal strength in the fourth OCT image is particularly low—with the central part of the RNFL fading out. Nevertheless, our segmentation model generates an almost-perfect matching result despite such defects that often lead to artifacts.
Table 1 compares the metrics obtained from our OCT segmentation network, ResUNet with transfer learning, against the other approaches, including the original UNet
12, ReLayNet
16, a variant of UNet which was designed specifically for the OCT segmentation task, Panoptic Feature Pyramid Network (FPN)
17, a lightweight, top-performing network for general image segmentation task, as well as our network without pre-trained weights, on the 300 SD-OCT B-scans with the testing data of the OCT imaging dataset. For RNFL thickness estimation, our method achieves the lowest
DIFF of only 0.0009 mm. For RNFL segmentation, although the average
SEN of our model is not the highest, its overall performance is still the best since it achieves the highest average
ACC, SPE, DSC of 0.992, 0.998, 0.929, respectively. To demonstrate that the performance increase of our method is significant, we conduct a series of two-sample
t tests hypothesizing that the segmentation metrics (
ACC, SEN, SPE and
DSC) of our method are higher than the other methods.
Table 2 presents the
p-values of these metrics for different methods.