In this study, we performed prospective usability testing of an early binocular OCT prototype in a population of study participants with chronic eye disease, as well as in healthy volunteers. To our knowledge, this is the first system that can perform a comprehensive eye examination – with functional and diagnostic testing in addition to conventional OCT structural imaging – in an end-to-end, automated manner.
Historically, eye examinations in hospital eye clinics have been fragmented, inefficient, and costly.
4–6,8 The binocular OCT prototype combines many routine tests into one single instrument, with the aim of improving the speed and efficiency of patient flow, in addition to providing reproducible and quantitative data for several aspects of the eye examination. Designed as an automated, patient-operated device, usability testing is indispensable to predict the likelihood of future successful implementation in eye clinics. Moreover, usability testing can identify potential user and device errors, and thus facilitate continued improvement of the device in an iterative process.
Our cohort of first time users was able to complete the full suite of tests using the binocular OCT prototype without any previous training and without any assistance during the examination. Participants commented that the device provided “clear instructions” and was “easy to use.” The majority of our cohort was familiar with operating common technologies, such as computers and smartphones, and found the prototype to be easier to use in comparison. However, those unfamiliar with technology also rated the device highly on ease of use.
We found that the subjective ratings for ease of use and appeal of the device correlated with ratings for test duration. Participants who perceived the examination took a short time, rated the ease and appeal more positively. However, subjective ratings for duration did not correlate with the total examination time. We observed no significant difference in the total examination time between participants with chronic eye disease and healthy controls. Overall, the median time taken to complete the examination was 702 seconds (11.7 minutes) for participants with chronic eye disease and 638 seconds (10.6 minutes) for healthy controls. By comparison, Callaway et al.
8 reported a mean clinic time of 28.8 minutes for patients to undergo technician work-up (including history-taking and visual acuity measurement), and acquisition of retinal OCT in the photography suite. In many real-world settings, mean diagnostic testing time is likely to greatly exceed this, particularly in public healthcare settings, which are often overburdened and under resourced. Thus, binocular OCT examination is likely to be more efficacious than current workflows as patients undergo all tests in one location in an automated manner, reducing the time patients spend travelling through the eye clinic. Nonetheless, an important aspect of iterative usability testing is to try to identify examination components that can be further improved in terms of speed.
A small proportion of examinations generated ungradable data. In the case of visual acuity and perimetry measurements, this was a consequence of the user not responding verbally when required. This occurred more frequently for the visual acuity exam in the left eye – the first eye to be tested that required a verbal response. A more complete set of results was obtained for the right eye in these users, and for the subsequent perimetry exam. This may be explained by a learning effect, where the user subsequently understood that the task required a verbal response. As similarly reported in perimetry literature, increased exposure to the device on repeated testing is likely to yield more reliable results and improve all aspects of usability.
20 Similarly, we found that the motility exam generated more reliable data when the second eye was fixating. Ungradable data for motility exams was more prevalent in users with chronic eye disease. This was likely related to poor visual acuity or reduced visual fields, affecting the ability to discern the motility target. In future iterations, this test could be improved by using a high luminance ‘fix and follow' target.
We obtained good quality automated OCT images of the anterior segment in all except one participant, and good quality posterior segment and vitreous images in the majority of participants. Given our aging populations, and the increasing prevalence of ocular comorbidities, the collection of objective, quantitative OCT data for the whole-eye is likely to be valuable for longitudinal monitoring and detection of eye disease. In addition, the ability to derive OCT measurements of vitreous activity could have applications in monitoring vitritis.
21 The quality of OCT data generated for both whole-eye imaging and pupillometry was affected by ocular misalignment – a consequence of the user moving their eyes, poor fixation, or blinking during the examination. Our prototype is susceptible to these errors due to the relatively long image acquisition time to complete simultaneous whole-eye OCT (mean 139 seconds). With advances in swept-source OCT laser technology, image acquisition speed is likely to improve in further iterations of the device, rendering the device less vulnerable to artifacts. The quality of OCT imaging, in particular, posterior segment and vitreous imaging also appeared to be affected by large angle strabismus. In these cases, the images for the fixating eye were acceptable, whereas the OCT lasers were unable to image directly through the pupil in the nonfixating heterotropic eye due to the large angle between the pupil plane and the direction of the laser.
To be functional as an automated and interactive device, it is essential that the system is responsive to the user. In the current binocular OCT prototype, this encompasses elements such as voice recognition. In our study, the sensitivity of the VRS was only 64%. This is likely related to the wide variation in articulation, volume, and regional accents of our cohort. Although the examination was undertaken in a quiet room, background noise from the machine itself may have impacted the response heard by the system. As voice recognition technology becomes more sophisticated, the error rate is likely to reduce, but may not be eliminated. Other interactive features, such as registering responses via buttons, in a similar way to many current visual field tests, may be more appropriate for some functional tests. In some populations (e.g., pediatrics), where responses may be unreliable, objective tests using OCT imaging to track responses may be more suitable. For example, visual acuity could be measured by presenting optokinetic stimuli to the user while simultaneously tracking the movement of the fovea on OCT.
Dissecting the examinations further into the time taken to complete major exam components, we found that delivering instructions took the greatest amount of time, especially in exams where the user was repeatedly reminded to fixate. Participants commented that the central target was unclear – improving its intensity is likely to improve fixation. Although participants found the instructions to be “clear,” articulating instructions in a more succinct manner will reduce the overall examination time – this is one aspect that is particularly likely to benefit from patient and public input.
The main advantage of human operators is the immediate recognition of the discussed errors and artifacts, whereas fully automated devices will require an inherent feedback mechanism to assess the quality of the generated data. This is important for determining whether tests need to be repeated, or if the user requires further or specialist examination beyond the scope of the device. This was also one concern highlighted at our PPI event. Reassurance could be provided to the patient via a visual or audio notification, or indirectly through feedback from a technician working in the clinic; however, this would be most beneficial in parallel with an in-built tool for simultaneous quality control.
User requirements encompass more than clinical effectiveness, and the ergonomics of the device must also be considered. Many of our participants commented that the device interface was physically uncomfortable. Assessing the needs of multiple types of users is essential to encourage continued use of medical devices.
10 For late-stage prototypes, extensive ergonomic testing will be essential prior to commercial release.
In summary, the results of our usability study, and related focus group testing, make it clear that patients are receptive to the concept of an automated eye examination. To be attractive to users, easy to use, and effective at performing automated eye examinations, the system will need to be quick, responsive, and comfortable. For a system that aspires to be fully automated (i.e., operated by the patient without assistance), ongoing patient and public input will be essential to guide the design of the device. Further studies will be required to validate the diagnostic accuracy of each of the tests offered by the system. Once established, binocular OCT will offer objective, quantifiable information about almost every aspect of the eye examination and has the potential to supersede many traditional but flawed testing methods. It is unlikely that the automated eye examination will be suitable for use in all patients. However, if such a system can replace some aspects of the eye examination, workflows and waiting times are likely to improve, costs are likely to reduce, and clinicians will be able to devote more time to patient care. Ultimately, this will improve the overall experience for both the patient and the clinician, and improve the overall quality of patient care.