Abstract
Objective:
This study aims to compare a new prototype for a portable anterior eye segment imaging system with the standard method for ophthalmology examination.
Methods:
The new imaging system consisted of two IMX219 Arducam autofocus sensors (Arducam, China, Nanjing) for Raspberry Pi V2 camera module connected to a Raspberry Pi Zero W (Raspberry Pi Foundation, UK, Cambridge) that clips to a wearable headset. The 2D videos of the anterior eye segment were recorded with the new system and a 720p FaceTime HD camera (Apple, Cupertino, CA). Afterward, ophthalmologists evaluated the videos using a standard clinical eye examination form. These evaluations were compared with the standard slit-lamp clinical assessment performed during the patient's visit.
Results:
Thirty-five eyes were evaluated. The sensitivity and specificity percentages were statistically significant between the two imaging modalities (P ≤ 0.001). The evaluations performed from videos obtained with the new imaging system had better sensitivity and specificity percentages overall. However, statistically significant differences were only observed in cornea, anterior chamber, iris, and lens.
Conclusions:
Specificity percentages were higher than sensitivity percentages in both imaging modalities, indicating that video evaluations are less accurate for pathological screening. Nevertheless, the new system evaluations were significantly better than the webcam evaluations.
Translational Relevance:
This study presented an alternative system to assess eye conditions for telemedicine, one that provides more details than the current standard and uses new wearable headsets technologies.
Telemedicine presents an alternative to traditional face-to-face medical consultations. Teleophthalmology increases access to ophthalmic care in rural areas where specialists are few. In addition, teleophthalmology lowers screening barriers and facilitates diagnosis follow-ups.
1 Moreover, during the coronavirus disease 2019 (COVID-19) pandemic, telemedicine systems were required to continue providing healthcare standards. Leading telehealth platforms reported an increase in virtual patient visits from 257% to almost 700% because of the pandemic.
2 However, the most used method for teleophthalmology constituted video visits and eye evaluations through computer web cameras. Even though web cameras are a widely available technology, the quality of the images, the room lighting, and camera focus limits doctors’ ability to evaluate eye conditions. Therefore, there is a need to develop accurate and reliable teleophthalmic systems and methods that achieve the accuracy of in-person examinations.
3
Anterior and posterior eye segment imaging constitutes a fundamental tool for ophthalmic diagnosis and a necessary component in teleophthalmology. However, clinical imaging devices are complex, not portable, and require trained clinical technicians. The current imaging systems used in teleophthalmology include smartphones, compact mobile cameras with lens adaptors, and computer web cameras.
4–6 Several studies compared the diagnostic capabilities of images from portable cameras with pictures taken with the clinical gold standard devices. Although there is a wide range of sensitivity and specificity values, these studies show the promising application of portable imaging systems to diagnose anterior eye segment conditions.
7,8
On the other hand, portable devices available for teleophthalmology also require another person's assistance to maintain them correctly positioned during imaging and testing. Additionally, several mobile devices target a specific test or imaging technique. Therefore, our research group intends to develop a comprehensive ophthalmic diagnostic tool that allows the autonomous testing of a patient's eye conditions. We propose the application of headset technologies as a potential solution. The users can comfortably wear them, and they have strong computer capabilities and specialized optical elements and cameras, making them an excellent option for teleophthalmology applications.
The computer technology of the latest virtual reality/augmented reality (VR/AR) headsets allows the development of several software applications for teleophthalmology, from visual field testing to vision augmentation (Goldbach, et al. IOVS 2021;62: ARVO E-Abstract 1017; Kashem, et al. IOVS 2021;62: ARVO E-Abstract 33189). However, obtaining images of the eye requires external high-quality cameras. This paper discusses the development of a wearable camera system for anterior eye segment imaging. In addition, this study assesses the accuracy and reliability of spot diagnosis of anterior eye segment pathologies by specialized ophthalmologists using 2D videos recorded with the new wearable camera system and 2D videos recorded with regular web camera. These evaluations are compared with the standard slit lamp biomicroscope assessment performed during the clinical visit. The purpose of this study is to validate the application of this new system as an improved teleophthalmology evaluation tool for anterior segment examinations.
The research represents a clinic-based comparative study. The University of Miami Institutional Review Board reviewed the study protocol to verify it followed the ethical guidelines for research involving humans established on the Declaration of Helsinki and the Health Insurance Portability and Accountability Act (HIPAA). Patients were recruited from Bascom Palmer Eye Institute, Miami, Florida, at their scheduled clinical consults. Patients’ inclusion criteria were not based on the presence or absence of pathology and eye dilation. The eyes evaluated in the study corresponded to the eye examined by the clinician during the visit. Participants were eligible to be included in the study if they were over 18 years old and were willing and able to sign the informed consent form. Twenty patients agreed to participate in the study, and 35 eyes were assessed by the attending doctor during the clinical visit. A total of 70 videos were recorded using the wearable headset camera system and the computer webcam. The patients were recruited from the clinic of 2 anterior eye segment ophthalmologists, and the 2D videos recorded were later divided and sent to 4 other ophthalmologists for evaluation.
The wearable headset camera system for anterior eye segment imaging developed for this study consists of two Arducam autofocus sensors (Arducam, China, Nanjing) for Raspberry Pi V2 camera module connected to a Raspberry Pi Zero W (Raspberry Pi Foundation, UK, Cambridge) attached to a HoloLens 2 (Microsoft, Redmond, WA). The camera mount prototype clips on the glasses, as shown in
Figures 1 and
2. This prototype can be manually adjusted to acquire a centered view of the eye through a system of hinges, a ball-and-socket articulation, and the HoloLens 2 visor mechanism. The latter allows the glasses’ display to be removed from the field of view of the camera when the imaging test is required. The cameras possess a one fourth in 8 MP IMX219 autofocus lens sensor with an effective focal length of 2.8 mm. However, the videos were recorded at 30 frame per seconds with an image resolution of 1080p using the H264 encoder to support Raspberry Pi Zero W video processing memory. These camera specifications allowed a magnification of × 0.069. A secure VNC connection allowed users to control the Raspberry Pi Zero W from a local computer and preview the camera’s feedback to adjust its position and focus. In addition, six diffused white LEDs were added to the camera to improve illumination while imaging the eyes. The code written to control the camera and the LEDs was developed in Python (Python Software Foundation, Wilmington, DE).
On the other hand, to introduce this new imaging system as an alternative option for telemedicine, videos were also recorded with computer webcams, the standard method used during the COVID-19 pandemic to perform the video visits. The computer used to record the videos was a 2012 MacBook Pro with a 720p FaceTime HD camera (Apple, Cupertino, CA). The images acquired had a resolution of 1280 × 720 and a magnification of × 0.0205. Room lighting was used for imaging with the web camera. To standardize illumination conditions the same clinical room was utilized, all the windows were closed, and all the room lights were on when recording started.
First, doctors evaluated the patient during the clinical visit with the traditional slit-lamp biomicroscope. Afterward, a research fellow explained to the patients how to correctly position the wearable headset. Because this study utilized the first prototype for the wearable imaging system, and the autonomous image acquisition and camera positioning modules are still under development, the research fellow had to also assist the patients in acquiring the images and repositioning the system to improve image quality.
The researcher positioned the HoloLens glasses on the patient’s head with the display flipped upward to set the cameras at eye level, as seen in
Figure 3. The research fellow accessed the Raspberry Pi Zero on an external local computer to record and view the live camera feedback. Based on the camera feedback, the headset and camera position were adjusted to center the eye image in the preview frame. The computer program recorded 2D videos one at a time, starting with the right eye, and turned-on corresponding LEDs while recording. Each video lasted about 1 minute. The camera is programmed to focus automatically, but manual focusing with the keyboard is also available in case the camera position changed or if the best focus was not achieved after the autofocus trial. In addition, the researcher could adjust LEDs’ brightness before starting recording based on the patient's mydriasis status or if the patient complained about light sensitivity. The researcher indicated the patient to move the eyes up, down, right, left, and center. Movement of the eyes while recording was standardized for both headset and web camera systems. Because the camera position for both imaging modalities is fixed while recording, moving the eyes allowed a better evaluation, and increased recording time.
Next, the fellow asked the patient to come closer to the computer web camera. The position of the web camera had to be adjusted to get a centered image of the eye. Because a laptop built-in web camera was used, the display angle and the height of the laptop with respect to the patient position were modified if required. On the camera preview, a green rectangle was added to serve as a guide to position the patient’s eyes, the purpose was to keep the eyes within the rectangle while recording. The patient was directed to look straight into the camera’s green light at first for positioning. During recording, the fellow asked the patient to move the eyes up, down, right, left, and center, as previously done with headset system. Another program developed in MATLAB (MathWorks, Natick, MA) recorded 2D web camera videos.
For all the patients, the videos with the wearable headset were recorded first followed by the web camera recordings. This order was determined to save time in case a patient was called for another appointment, because the new imaging prototype required more time to setup. In addition, recording the web camera videos was faster because the patients were more familiar with this technology. Examples of the images acquired with both imaging systems are presented in
Figures 4,
5, and
6.
This study assessed the accuracy and reliability of performing eye evaluations based on the videos obtained from a wearable headset camera system and a computer web camera. These evaluations were compared with the standard clinical assessment performed by an ophthalmologist with a slit-lamp biomicroscope during the clinical visit.
Results showed that the evaluations for both imaging modalities had high specificity but low sensitivity. Overall, eye evaluations from the videos have low accuracy in screening for pathologies. On the other hand, the evaluations performed from the videos recorded with the wearable headset camera system had statistically significantly higher specificity and sensitivity than the evaluations obtained from webcam videos. Similarly, the specificity and sensitivity values were higher in the headset camera system than in the webcam for all ocular features. However, significant differences were only seen in the cornea, anterior chamber, iris, and lens evaluations. Moreover, the number of UT evaluations was highest in the webcam videos, and the greatest percentage was seen when evaluating the cornea, iris, anterior chamber, and lens. Therefore, the wearable headset camera system allowed better evaluation of these eye structures than the current telemedicine eye evaluation standard, but not as significant as the clinical examinations.
In a similar study performed in Perth, Australia, at the Lion Eye Institute, researchers compared the detection of anterior segment disorders from still digital images between a portable slit-lamp camera and the standard slit lamp biomicroscope with the traditional clinical assessment.
7 This research studied 392 eyes, and 33 specific conditions were identified. The measured specificity percentages ranged from 98% to 99% for the portable imaging system. However, according to researchers, a low prevalence in many categories of anterior segment conditions limited the study, and the sensitivity and specificity percentages are not particularly useful when less than five subjects have a specific condition. This is a similar issue encountered in our study, where few of the identified conditions were not present in multiple patients. In addition, in the Lion Eye Institute research, three ophthalmologists evaluated all the patients; therefore, in our future studies, doctors should evaluate all patients to better compare inter subject agreement. In summary, they determined that standardized views of still images did not compare with the clinical examinations, because the cameras did not offer a full view of eyelids and conjunctiva and had a shallow depth of focus. The prototype presented in this study addresses some of these issues by analyzing videos instead of still images, allowing control of the camera focus, and adjusting camera position and lighting conditions. These features can be utilized to improve evaluation of different ocular structures.
Another study compared the photographs captured with an iTouch 5G smartphone camera (Apple, Cupertino, CA) and a portable diffuse light ophthalmic camera (Nidek, Fremont, CA), to the gold standard slit lamp and corneal specialist examination for diagnosing corneal pathologies.
8 Three ophthalmologists evaluated the images from the two cameras. The smartphone camera had a sensitivity range from 54% to 71% and a specificity range from 82% to 96%. The portable ophthalmic camera sensitivity ranged from 66% to 75%, and specificity from 91% to 98%. Similarly, this study discusses the presence of high specificity and low sensitivity percentages obtained when evaluating eye conditions from images. The standard for accuracy of telemedicine screening for diabetic retinopathy and other ophthalmic diseases requires a sensitivity percentage of at least 80%. A high sensitivity is required because it determines the accuracy of the imaging modality to capture ocular pathologies, especially in telemedicine that relies on images for remote evaluations. In the study presented in this paper, this sensitivity percentage limit was not achieved from the video evaluations.
The difference in video quality could be a factor that affected the results. For instance, the wearable headset camera system had a good resolution, however, the camera highest resolution could not be achieved because of a Raspberry Pi video memory restriction. In addition, even though LEDs aided with illumination, sometimes they created reflections and over-saturated images that affected video quality.
Moreover, the Arducam cameras focus could be adjusted, and because they were placed at eye level and near the eyes, a closer view could be obtained. However, the web camera video quality was limited by a fixed focus and a longer working distance. On the other hand, when taking videos with the web camera, the room illumination was another factor that limited video quality. Standardizing lighting conditions was challenging. The computer was left in the same position and usually patients were able to move forward and bring themselves closer to the camera. However, for some patients, the laptop had to be move toward them because they had difficulty seeing and moving around. Therefore, proper lighting conditions were not always achieved when recording with the web camera.
The correct positioning of the headset camera system was essential to obtain good videos. However, this was not always achieved. Factors that influence these differences included the physiognomy of the patients, how the headset fits on the patient's head, and the inter-pupillary distance. These are aspects that must be included in the development of the next camera prototype. During this study, the cameras were fixed to the center of each display on the headset glasses, and a ball-and-socket structure allowed for angle adjustment; as a result, sometimes the cameras were not centered correctly in the eye. For future prototypes, a motorized camera system can assist in the correct alignment and positioning of the cameras. In addition, the development of a slit-lamp illumination system will be better to improve the assessment of corneal pathologies and other anterior eye segment features. A new program will be developed to improve user interphase, a site where doctors can access real-life camera feedback and modify the camera position remotely. However, this initial prototype and study shows the application of AR/VR headsets in teleophthalmology for imaging the eye.
The comparison of the averaged video quality grade between to two imaging modalities showed that the videos allowed doctors to exclude only emergent findings. Even though the videos from the wearable system were expected to have a higher image quality grade, the averaged grade for both imaging systems was the same. This could be the result of having different graders evaluating only a set of different videos. The image grade quality is a subjective assessment, therefore, in a future study, a quantitative method for measuring image quality could be included to compare the two systems.
Future studies need to compare the evaluation performed with the new wearable camera prototype with other portable camera systems such as smartphones with camera adaptors, because these are other common devices utilized in teleophthalmology. In addition, it will be good to compare evaluations from still images obtained with the wearable headset system with evaluations from still images obtained with a slit lamp biomicroscope. In addition, to improve the study design and the sensitivity of the video evaluations, more patients with a diverse set of ocular pathologies should be considered.
In summary, this study presents the development of a portable and autonomous prototype for anterior eye segment imaging, with a potential application in teleophthalmology. The analysis of the results shows that this new system is better for eye evaluations than the current standard for teleophthalmology that comprises of computer webcams. However, this system needs further improvement to provide a similar assessment to clinical evaluations.
Supported by a National Eye Institute Center Core Grant to the University of Miami (P30 EY014801), and Research to Prevent Blindness (RPB). The funding organization had no role in the design or conduct of this research.
Disclosure: A. Diego, None; M. Abou Shousha, United States Non-Provisional Patent (Application No. #20190094552). Patent is owned by University of Miami and licensed to Heru, Inc. MAS is an officer, equity holder and sits on the Board of Directors of Heru, Inc.