January 2025
Volume 14, Issue 1
Open Access
Low Vision Rehabilitation  |   January 2025
Performance on Activities of Daily Living and User Experience When Using Artificial Intelligence by Individuals With Vision Impairment
Author Affiliations & Notes
  • William Seiple
    Lighthouse Guild, New York, NY, USA
    New York University Grossman School of Medicine, Department of Ophthalmology, New York, NY, USA
  • Hilde P. A. van der Aa
    Lighthouse Guild, New York, NY, USA
    Amsterdam UMC, Vrije Universiteit Amsterdam, Ophthalmology, Amsterdam, The Netherlands
    Amsterdam Public Health Research Institute, Program Quality of Care, Amsterdam, The Netherlands
  • Fernanda Garcia-Piña
    Lighthouse Guild, New York, NY, USA
  • Izekiel Greco
    Lighthouse Guild, New York, NY, USA
  • Calvin Roberts
    Lighthouse Guild, New York, NY, USA
  • Ruth van Nispen
    Amsterdam UMC, Vrije Universiteit Amsterdam, Ophthalmology, Amsterdam, The Netherlands
    Amsterdam Public Health Research Institute, Program Quality of Care, Amsterdam, The Netherlands
  • Correspondence: William Seiple, Lighthouse Guild, 250 West 64th Street, New York, NY 10023, USA. e-mail: [email protected] 
Translational Vision Science & Technology January 2025, Vol.14, 3. doi:https://doi.org/10.1167/tvst.14.1.3
  • Views
  • PDF
  • Share
  • Tools
    • Alerts
      ×
      This feature is available to authenticated users only.
      Sign In or Create an Account ×
    • Get Citation

      William Seiple, Hilde P. A. van der Aa, Fernanda Garcia-Piña, Izekiel Greco, Calvin Roberts, Ruth van Nispen; Performance on Activities of Daily Living and User Experience When Using Artificial Intelligence by Individuals With Vision Impairment. Trans. Vis. Sci. Tech. 2025;14(1):3. https://doi.org/10.1167/tvst.14.1.3.

      Download citation file:


      © ARVO (1962-2015); The Authors (2016-present)

      ×
  • Supplements
Abstract

Purpose: This study assessed objective performance, usability, and acceptance of artificial intelligence (AI) by people with vision impairment. The goal was to provide evidence-based data to enhance technology selection for people with vision loss (PVL) based on their loss and needs.

Methods: Using a cross-sectional, counterbalanced, cross-over study involving 25 PVL, we compared performance using two smart glasses (OrCam and Envision Glasses) and two AI apps (Seeing AI and Google Lookout). We refer to these as assistive artificial intelligence implementations (AAIIs). Completion and timing were quantified for three task categories: text, text in columns, and searching and identifying. Usability was evaluated with the System Usability Scale (SUS).

Results: The odds ratios (ORs) of being able to complete Text tasks were significantly higher when using AAIIs compared to the baseline. OR when performing “Searching and Identifying” tasks varied among AAIIs, with Seeing AI and Envision improving the performance of more tasks than Lookout or OrCam. Participants expressed high satisfaction with the AAIIs.

Conclusions: Despite the findings that performance on some tasks and when using some AAIIs did not result in a greater number of PVL being able to complete the tasks, there was overall high satisfaction, reflecting an acceptance of AI as an assistive technology and the promise of this developing technology.

Translational Relevance: This evidence-based performance data provide guidelines for clinicians when recommending an AAII to PVL.

Introduction
Whereas past assistive technologies (ATs) have provided reading support for people with vision loss (PVL) by manipulating the appearance of text on screens, recent software and hardware advances, innovations in computer vision, and the development of algorithms commonly known as artificial intelligence (AI) offer promising solutions. AI simulates human intelligence processes, such as learning, reasoning, visual perception, speech recognition, decision making, and language comprehension. The growing ubiquity of AI has significantly increased access to information for nearly everyone, including PVL. Moreover, deep-learning approaches to AI allow for rapid retrieval and processing of information that support processes, such as reading text, recognizing objects and persons, and describing scenes. With increased computing power and improvements in digital cameras, integrating AI into portable applications, such as “smart glasses” and cellular phone apps, holds great promise for the future of AT, offering hope for a more inclusive and accessible future. 
Recently, there has been considerable interest in the engineering literature regarding applying AI to the needs of PVL. A review of this literature in 2023 found 646 published studies on vision assistive AI in just the previous 2 and a half years.1 Diverse and innovative applications of AI in developing AT for PVL share the goal of enhancing independence and quality of life.29 Walle et al.8 and others have outlined the issues to be resolved in the use of AI for PVL, including the types of information to be obtained (e.g. object recognition, navigation, and scene description), the location of acquisition hardware (e.g. wearable, cane-mounted, and mobile phones), the types of sensors (e.g. cameras, global positioning system (GPS), position sensors, and infrared detectors), the algorithm used (e.g. YOLO, SLAM, and Kalman filters), the limits imposed by the processing load (to be done locally or remotely), the nature of the user interface (e.g. buttons, gestures, and voice), and the choice of feedback modalities (e.g. auditory and tactile). Despite these multiple obstacles, proposed solutions have been appearing rapidly. Coupling innovative AI technologies with the capabilities of wearable devices and mobile phone apps can provide users with tools for navigating and understanding the visual world. The potential of AI to transform the lives of PVL has generated enthusiasm in the vision-loss community. However, the functionality of AI in the hands of PVL must be objectively quantified. In this study, we examined how these advanced technologies perform, given the constraints of vision loss. 
Methods
Study Design
A cross-sectional, counterbalanced, cross-over study compared performance using four assistive artificial intelligence implementations (AAIIs): two smart glasses OrCam MyEye 2 Pro, version 9.21 (https://www.orcam.com/en-us/home, $4250 as of July 31, 2024), and Envision Glasses Home Edition (https://www.letsenvision.com/glasses/home, $2499 as of July 31, 2024); and two mobile phone applications Seeing AI version 1.0.1 on an iPhone 13 pro, IOS 17.6 (https://www.seeingAI.com, $0), and Lookout by Google for Android version 4.3 on a Pixel 4, Android version 11 (https://www.lookout.com, $0). All AAIIs provided audio feedback, such as describing scenes and reporting the presence and descriptions of people and objects, to allow users access to text and information about their surroundings. 
Recruitment and Inclusion Criteria
Participants were recruited from the Lighthouse Guild's low-vision clinic and rehabilitation services. We asked the staff to refer patients for screening and advertised on the research page of the Lighthouse Guild's website. Potential participants were screened based on the inclusion criteria, including being at least 18 years old, visual acuity from 20/100 (0.7 logMAR) to no light perception, and a cognitive function score of ≥10 (based on the 6-item Cognitive Impairment Test).10 Additionally, individual goals were measured with the Participatory Activity Inventory (PAI),11 and technology goals were determined using the Matching Person and Technology (MPT) assessment.12 Participants were enrolled based on their willingness to explore the use of AT to achieve their goals and commitment to attend the clinic for five sessions. Participants were excluded if they had prior proficiency with any AAII to be tested in the study. 
Study Procedure
Data were collected during five sessions – baseline performance with no AI and wearing usual refractions (session 1) and four subsequent visits, each with one of the four AAIIs (sessions 2–5). Sessions ran for approximately 90 minutes, including breaks, and were scheduled about a week apart. The order of AAII testing was counterbalanced across participants. Users practiced each AAII's range of functionality by performing training tasks. We encouraged participants to become familiar with the physical device, open the software, scroll among options, and learn appropriate camera framing techniques for each task. The practice did not use the versions of the tasks tested in the data collection part of the study, and performance on the practice tasks was not scored. 
Before the testing phase began, the researchers quantified the performance of all settings of a given AAII for each task. They determined the setting that yielded the best results for a task and AAII (Table 1). During testing, the experimenter launched the app and selected the appropriate setting for the task. 
Table 1.
 
The Best Settings for the Tasks Determined Prior to Testing
Table 1.
 
The Best Settings for the Tasks Determined Prior to Testing
Outcome Measures
We collected descriptive data at baseline on eye disease/cause of vision loss, visual acuity measured with an Early Treatment Diabetic Retinopathy Study (ETDRS) letter chart at one meter (Precision Vision ETDRS Chart No. 2112, https://precision-vision.com), onset of vision impairment, self-reported severity of visual impairment, age, sex, race, comorbid conditions, education, living situation, employment, participants’ needs and goals (measured with the PAI), and familiarity with and receptivity to AT to achieve their goals (measured with the MPT). 
Performance and timing on 14 vision-related activities of daily living (ADL) were evaluated at each session. The tasks were carefully chosen based on published literature, patients’ needs, and the capabilities of the AAII.1320 The tasks were divided into three categories based on the primary visual requirement of the tasks. All tasks were timed and scored except for reading a newspaper article, which was just scored. The criteria for scoring accuracy of an AAII when used by PVL and participants’ successful task completion when using an AAII are shown in Table 2
Table 2.
 
Scoring Criteria for the Tasks
Table 2.
 
Scoring Criteria for the Tasks
Category 1, “Text,” included tasks typical of those evaluated in vision-related ADL questionnaires. The tasks were: (1) read a newspaper article (from The New York Times Large Print Weekly, text size 2 M) and describe the article’s contents; (2) read an invoice and report the amount due (text size 2 M); (3) read a handwritten five-item grocery list and report the non-food item (text size equivalent 2 M); (4) report the name of the medication and dose instructions on a pill bottle (text size for medication 2.75 M and instructions 2 M); (5) identify a banknote; and (6) read a street sign at eight meters (text 74 M). 
Category 2, “Text in Columns,” included (7) read a table of contents from the newspaper and report the page number of a given article (e.g., on which page is the Business Section? text size 4 M); and (8) read a digital TV channel guide from Spectrum presented on a 24 inch TV at 2 meters (equivalent text size 3 M). The task was to report the channel for a given show (e.g. On which channel is Major League Baseball [MLB]?). 
Category 3, “Searching and Identifying,” included (9) find the matching sock. A sock of a given color was shown, and the task was to find that color sock in a group of four socks of different colors; (10) scan a grocery product box to identify the product from the barcode; (11) match a photograph of a person. One photograph was shown initially, and the task was to match the same photograph in a group of three photographs; (12) describe a scene. Participants stood at 1.5 meters in front of a color poster of a landscape so that the image filled the camera's field of view; (13) find a person in a room. A person stood quietly at one of five positions in the testing room; and (14) identify a room when standing at the entrance. The rooms were an office, living room, bathroom, kitchen, and store. 
To prevent learning effects, a different version of each task was used in each session (e.g. 5 different newspaper articles, 5 different medicine bottles, and 5 different street signs), and the order of the task versions was counterbalanced across AAIIs. 
At the end of each session, the participant completed two surveys: (1) a subsection of the MPT (form 4) titled Assistive Technology Predisposition Assessment. The questions were: “This AT will help me achieve my goals,” “This AT will fit well with my accustomed routines,” “This AT will physically fit in all desired environments,” and “I will feel comfortable using this AT around others”); and (2) the System Usability Scale (SUS), which has 10 items scored on a 5-point Likert scale and has often been used in studies on AT due to its good psychometric properties.21,22 After each session, participants completed free-form responses to the question, “For which tasks would you use this AT, including any tasks in addition to those tested today?” 
Statistical Analyses
Log odds ratios (ORs) with Firth Bias were calculated, indicating the odds of a person being able to perform the task when using an AAII compared with the odds of performing the task at baseline (without any AAII). Person abilities were compared with AAII abilities descriptively. To detect overall differences in timing among conditions (baseline and the 4 AAIIs) across each of the multiple tasks, nonparametric Friedman One-Way Repeated Measure Analyses of Variance by Ranks were performed (the data failed to meet the assumptions of parametric testing). A Benjamini-Hochberg correction was used to control the false discovery rate due to multiple Friedman tests.23 Post hoc (pairwise) Wilcoxon signed-rank tests were used to determine whether there were statistically significant differences among conditions. Multivariate logistic regression analyses were used to identify baseline performance and outcomes predictors. All analyses were performed in R Studio version 4.2.1, and a P value of ≤ 0.05 was used to determine statistical significance. 
Results
Participant Characteristics
The Institutional Review Board of the Lighthouse Guild in New York, NY, approved the study. Twenty-five eligible participants gave informed consent and completed the study. They were compensated for their participation. The average age of the participants was 64 years (range = 30–83 years), visual acuity ranged from 0.7 logMAR to no light perception, and text-to-speech was the most used AT at baseline (84% of participants; Table 3). Fifty-two percent were White, 40% were Black, and 8% were Asian. In response to the MPT questions about previous experience with any assistive technology, 62% were satisfied, 28% were neutral, and 10% were unsatisfied. 
Table 3.
 
Participant Characteristics (n = 25)
Table 3.
 
Participant Characteristics (n = 25)
Outcomes
Initially, the data were examined for order effects because there might have been a practice effect over the AAII conditions. There was no significant effect of order for any tasks (Friedman One-Way Repeated Measure Analyses of Variance by Ranks), confirming that the counterbalancing was effective. 
We compared the ORs for the number of participants who could complete the tasks using an AAII (“N able” in Table 4) relative to those who could do the tasks at baseline. We found that the ORs of PVL being able to complete tasks in the text category were significantly higher for 100% of the tasks when using Envision, Seeing AI, and Lookout and for 83% of the tasks when using OrCam (see Table 4). In the text in columns category, significantly more participants succeeded in finding a page number and a channel for a TV show using OrCam (100%) and Seeing AI (50%). In contrast, using the other two AAIIs did not significantly increase the ORs of completing the text in columns tasks, predominately because they read the text column-wise. For the non-text searching and identifying tasks, using Seeing AI resulted in higher ORs of being able (83%), followed by OrCam (50%), Envision (50%), and Lookout (33%). In Table 4, ORs were indicated as significantly higher in bolded black text (ɫ) and significantly lower in bolded red text (‡). Nonsignificant findings are not bolded. 
Table 4.
 
Odds Ratios of PVL Being Able to Perform a Task
Table 4.
 
Odds Ratios of PVL Being Able to Perform a Task
Timing
The time to complete the task was measured as the total time it took the AAII to capture, process, and announce the information, plus the time it took the participant to interpret and report the information provided by the AAII. The overall Friedman analyses on the timing data for tasks in the text category showed significant time differences across AAII conditions for the invoice (χ² = 34.3, P < 0.001), handwriting tasks (χ² = 75.2, P < 0.001), medicine bottle (χ² = 34.0, P < 0.001), banknote (χ² = 45.3, P < 0.001), and street sign (χ² = 43.1, P < 0.001) tasks. Timing was not measured for the article task due to the different text lengths of the articles. In the text in columns category, neither the table of contents nor the TV guide task showed significant differences across testing conditions. For tasks in the searching and identifying category, the following tasks showed significant differences across testing conditions: color matching (χ² = 60.1, P < 0.001), face matching (χ² = 31.4, P < 0.001), landscape (χ² = 83.4, P < 0.001), and room identification (χ² = 9.3, P < 0.001). The overall Friedman test outcomes were not significant for timing for the barcode task and the find a person task. 
Post hoc Wilcoxon signed-rank analyses for the tasks found to have significant main effects are shown in Table 5. For the invoice and the medicine bottle tasks, timing was significantly faster than baseline when using all AAIIs. Timing for the handwriting task was significantly faster than baseline when using Envision, Seeing AI, and Lookout, but not OrCam. For the banknote task, only Seeing AI and Lookout were significantly faster than the baseline. All but OrCam allowed significantly faster than baseline performance on the street sign task. When doing tasks in the searching and identifying category, using OrCam and Lookout resulted in equivalent or statistically slower performance than baseline, and using Seeing AI allowed equivalent or significantly faster times for the landscape and room tasks. 
Table 5.
 
Wilcoxon Signed-Rank Test Results Comparing the Performance of an AAII to Baseline, Shown Only When the Overall Friedman Test was Statistically Significant (n = 25, P ≤ 0.05)
Table 5.
 
Wilcoxon Signed-Rank Test Results Comparing the Performance of an AAII to Baseline, Shown Only When the Overall Friedman Test was Statistically Significant (n = 25, P ≤ 0.05)
Comparison of AAII's Accuracy and Participants’ Task Success
We collected data on AAII's accuracy and participants’ task success. One might expect the relationship between AAII accuracy and participants’ task success to be unitary; however, this was not always the case. In Figure 1, the dots are placed at the X-values for the percentage of tasks the AAII could do correctly (based on the criteria described in Table 2). The arrows represent the magnitude of additional tasks completed by the participants. For example, the OrCam device accurately completed an average of 55% of the text tasks; however, the participants completed 33% more. Similarly, participants could complete 10% more tasks than Envision, 5% more than Seeing AI, and 26% more than Lookout. This better performance reflects the failure of the AAII to read correctly, including missing words and phrases, mispronounced words, misplacement of line orders, etc., plus the added effects of human intelligence using context. In these cases, participants accurately inferred information about the text that the AAII did not specifically report. In contrast, no arrows extend from the dots for tasks in the searching and identifying category where language context did not play a role in task completion. That is, participants were only successful to the extent that the AAII reported the information required to complete the tasks. 
Figure 1.
 
Comparison of AAII success versus participant success.
Figure 1.
 
Comparison of AAII success versus participant success.
Predictors of Outcomes
We explored the predictive value of sex, age, visual acuity, diagnosis, and responses to the baseline PAI questions about perceived performance on similar tasks to our outcome measures. Using Multivariate Logistic Regression analyses, we found that being able to do the tasks at baseline was significantly predicted by logMAR visual acuity and lower PAI-measured perceived difficulty of doing the same tasks (Table 6). 
Table 6.
 
P Values Associated With Predicting Performance at Baseline
Table 6.
 
P Values Associated With Predicting Performance at Baseline
Table 7.
 
P Values Associated With Predicting Change in Performance
Table 7.
 
P Values Associated With Predicting Change in Performance
We also explored the predictive utility of demographic and perceived task difficulty in terms of changes in task performance with each AAII compared to baseline (improved, unchanged, or worsened). For all task categories, higher perceived difficulty significantly predicted improved performance with an AAII (Table 7); that is, more participants who reported higher perceived difficulty on the PAI administered at baseline had improved task performance when using an AAII than those reporting lower difficulty. In addition, more participants with poorer baseline visual acuity had improved task performance when using an AAII than those with better acuity. 
Graphical Representation
As we have done previously,24 symbols were used to interpret the statistical findings from Tables 2 and 3 visually. Figure 2 aims to help practitioners recommend AT to PVL by offering a framework based on the AAII’s performance. The tasks represent the goals and needs of the patient, and the symbols represent the AAII’s performance. A large dot indicates where using an AAII resulted in a significantly higher OR, and the Wilcoxon signed-rank tests of quantitative timing data showed significantly faster performance than baseline. A small dot indicates that there was only a significantly higher OR, whereas the Wilcoxon signed-rank tests showed no statistically significant difference in timing relative to the baseline. An equal sign indicates when the AAII’s performance, as measured by both statistics, was not significantly different from the baseline. A small “X” shows a significantly lower OR whereas the Wilcoxon signed-rank tests showed no statistically significant difference, and a large “X” is drawn where both showed statistically worse performance than baseline. Across the tasks, using Seeing AI resulted in the best overall performance. 
Figure 2.
 
Graphical representation of devices based on ORs and the Post hoc Wilcoxon Signed-Rank findings.
Figure 2.
 
Graphical representation of devices based on ORs and the Post hoc Wilcoxon Signed-Rank findings.
Satisfaction
Based on the MPT (form 4), participants reported that all the AAIIs mostly met their desired goals, routines, and environments and that they did not feel self-conscious using them, with a median Likert score of 5 out of 5 for Seeing AI, 4 out of 5 for OrCam, 3 for Envision, and 3 for Lookout. 
SUS data were transformed and calculated according to the standard method.21 The mean score across participants was 77.9 for Seeing AI, 62.5 for Envision, and 57.2 for OrCam and Lookout (higher scores represent higher perceived usability). Person measures were calculated using Rasch analysis (Winsteps) for each AAII SUS score and compared using a One-way Repeated Measures Analysis of Variance. There was no significant effect of the AAIIs on perceived device acceptance (F = 1.04, P = 0.38). 
Perceived Use
At the end of each testing session, participants were asked to list the tasks they would use that AAII for. Their answers were binned into categories: reading, scene description, shopping, money identification, etc. There were no significant differences in the distribution of tasks across the AAIIs, and the total occurrence for each category is plotted in Figure 3. By far, the most often reported perceived use was for reading tasks. 
Figure 3.
 
Distribution of perceived use for AAIIs.
Figure 3.
 
Distribution of perceived use for AAIIs.
Discussion
Recent strides in machine learning have paved the way for the development of AI technologies that can perform tasks traditionally requiring human intelligence. Computational Neural Network innovations have revolutionized computer vision, leading to considerable progress in image recognition, object detection, and other visual tasks. This, combined with sophisticated Large-Language models that process and generate text and capture the intricacies of language, has supported translation into real-world applications that directly assist PVL (such as Seeing AI and Lookout). The future looks promising, as modern computer methods are increasingly being harnessed for assistance to PVL, as evidenced in recent review articles.1,2527 
In the present study, we compared the success and ease of use of four AAIIs to a baseline condition with no AAII by collecting objective performance data on real-world tasks and subjective data on usability and acceptance. Our findings showed that the four AAIIs we evaluated were most effective when used for tasks in the text category. The ORs of completing tasks were significantly higher than at baseline for five of the six tasks in the text category when using OrCam and for all six tasks using Envision, Seeing AI, and Lookout. For the searching and identifying category, ORs of being able to complete tasks were significantly higher than at baseline for five of the six tasks when using Seeing AI, four tasks with Envision, three with OrCam, and only two with Lookout. 
Previous studies have explored the functionality of AI with PVL. Granquist et al.28 compared performance using the OrCam MyEye 1 to an earlier version of Seeing AI using six reading tasks. With a sample size of 7, they found that 71% of the tasks could be completed using OrCam, and 55% could be done using Seeing AI. This difference in performance was not significantly different. Likewise, participants' ease of use ratings were not significantly different between the two AAIIs. In contrast, we found that using Seeing AI allowed more participants to perform reading tasks faster, perhaps due to differences in the newer versions of each AAII. Nguyen et al.29 studied the effectiveness of OrCam MyEye 2 in 20 patients with inherited retinal disease. Participants were trained and provided with an OrCam for 5 weeks of home use. At follow-up, there were significant improvements in self-reported functioning on the National Eye Institute Visual Function Questionnaire (NEI-VFQ)’s visual functioning subscale, the reading goal of the PAI, and persons’ scores on the OrCam Function Questionnaire (OFQ). There was no comparison with other AAIIs in this study, and the amount of at-home usage was not quantified. Amore et al.30 examined the efficacy of the OrCam MyEye 2 using reading, money, color, face, and object recognition tasks. Reading tasks improved the most, with an average of 71% of the participants, but other tasks improved by only 26% (from their table 1). Bhagat et al.31 assessed accuracy and usability in an unstated number of PLV users. They reported 100% accuracy for printed text for Seeing AI, Envision, and Lookout, 90% accuracy for handwritten paragraphs for Seeing AI and Lookout, and good usability for Seeing AI and Lookout, with medium usability for Envision. Our findings are generally consistent with these past reports. 
We expected that participants with eye diseases characterized by central vision loss would benefit most from an AAII. However, diagnosis, along with sex and age, did not predict the performance at baseline or outcomes in our current study. Although we found no statistical relationship between clinical and demographic variables and outcomes, perceived task difficulty assessed with the PAI at baseline significantly predicted the objective performance of the same tasks at baseline. Similarly, Szlyk et al.32 reported an average correlation between perceived and functional performance of 0.50 ± 0.12. Latham and Usherwood16 reported an average correlation of 0.61 ± 0.15 between self-reported difficulty and rate of task completion, and Kartha et al.33 found a correlation of 0.60 between self-reported visual ability and actual task performance. In our current study, perceived ability, as recorded by PAI questions, significantly predicted performance at baseline. This finding adds to the external validity of some of the questions in the PAI. 
We assessed usability with the SUS, an instrument sensitive to differences among products and interfaces.21,22 Amore et al.30 reported that the SUS score for OrCam was 63.5, but their study design did not compare the AAIIs. The usability of the AAII we tested ranged from 77.9 for Seeing AI to a low of 57.2 for both OrCam and Lookout. 
We found that the most frequently perceived potential use of an AAII was for reading (47.4%). This observation aligns with previous reports showing that reading improvement is the most common goal of PVL when seeking rehabilitation.3436 and as a use for technology.1 In the past, this stated goal of improving reading had been interpreted as a need for training to allow patients to best use their residual vision for reading37 and for technologies that alter text through manipulation of luminance, contrast, and magnification.3845 The data in Figure 3 suggest another interpretation; that is, PVL wish to access the content of text, even if the visual system does not process the text. They overwhelmingly stated that they would use the AAII for reading, even though the technology did not always “read” in a manner typical of visual reading. That is, audio reading is a very acceptable way to access text for PVL. 
A parallel exists between the AAII findings and our previous work using head-mounted displays.24 Although image-manipulation approaches to AT significantly increased the number of reading tasks PVL could do, they provided little assistance with other ADL tasks. Understanding written material is a straightforward task as all the information is included in the text, and the function of technology is to present that information in an understandable format. On the other hand, the information needed to complete an ADL task is not always lexical, and inferences need to be made based on a separate set of variables for each task. Current technologies, including AI, do not gather information top-down and fail to assist in tasks requiring active perception. 
Strengths and Limitations
Our experimental design and choice of tasks and outcome measures allow for multiple AAIIs to be compared in small-scale studies that provide objective and subjective information, which is imperative for evaluating quickly advancing technologies. In addition, we used a cross-over design and counterbalanced the order of the tasks to guard against order and sequence effects. We included a heterogeneous group of participants, varying in age, sex, onset and severity of the visual impairment, and various eye conditions (including glaucoma, macular degeneration, cone dystrophy, and retinitis pigmentosa). This diversity contributes to the richness of our data and aligns with claims that the AAIIs are suitable for a large variety of PVL. Moreover, our study was conducted without the influence of commercial parties that could compromise the objectivity and integrity of the results. 
Participants used the AAII after training and instructions. To solely test the AAII’s accuracy and task completion, we effectively made the AAII interface fully accessible by determining the best setting for each task for each AAII before data collection began (see Table 1) and then having the experimenter launch the app and select the ideal setting for each task. It is unknown whether the patterns of results would have been different if participants had had the opportunity to use the AAII at home for an extended period. Although some published studies have reported positive results following an at-home period using the technology, others have reported little or no change in performance.14,38,4345 The differences may arise from these studies’ lack of controlled/quantified usage at home. 
Selection Bias
We included participants of various ages, diagnoses, and visual acuities, typical of patients in our low-vision service who might benefit from using AT. The clinical presentations of the participants in our study were similar to those of other studies testing AT devices in PVL.14,38,4547 There were no dropouts during the study. Participants received reimbursement for their participation. However, there remains the possibility of volunteer bias in our study.48 
Implications for Clinical Practice and Future Research
We intend for vision professionals to use our findings to guide technology recommendations based on PVL clinical characteristics and functional needs. For example, two options for PVL who retain some pattern vision and wish to access text are image enhancement (e.g. magnifiers) and AI. Our past study on head-mounted displays found that an average of 51% of participants with acquired vision loss and visual acuity better than 20/800 who could not complete the tasks at baseline could complete reading tasks using magnification.24 In our current study, 73% of the participants within this visual acuity range who could not complete the tasks at baseline could complete the text tasks when using AI. Although not directly statistically comparable, these findings suggest that AI could be recommended as an acceptable alternative to magnification in this group of PVL. This is especially the case given the differences in cost and the observation that 97% of people with visual impairment already own a smartphone.49 It is logical that PVL with light perception or no light perception would not benefit from using image enhancement approaches. In contrast, an average of 76% of PVL in these lower visual acuity categories who could not complete the text tasks at baseline completed them when using AI. In this case, a strong recommendation for the use of AI is supported. 
Furthermore, PVL wish to find assistance with non-text tasks. Based on our current and past findings, PVL with acuity better than 20/800 benefited less from using image enhancement technology or AI in tasks such as those in our searching and identifying category. In contrast, 67% of those with light perception (LP) or no light perception (NLP) who could not complete the tasks at baseline completed them when using an AAII. However, the recommendations for the use of AI here depend upon the AAII, with 90% of those unable at baseline gaining the ability to complete the searching and identifying tasks when using Seeing AI, 71% when using Envision, but only 37% when using OrCam, and 27% when using Lookout. 
These examples emphasize the value of evidence-based data in guiding recommendations for low-vision technologies. 
Acknowledgments
This work was partly supported by a grant from the American Macular Degeneration Foundation. 
Disclosure: W. Seiple, Astellas Pharmaceuticals, Inc. (C); H.P.A. van de Aa, None; F. Garcia-Pina, None; I. Greco, None; C. Roberts, None; R. van Nispen, Janssen-Cilag NV (C) 
References
Gamage B, Do T, Nicholas, , P, et al. What do blind and low-vision people want from assistive smart devices? Comparison of the literature with a focused study. ASSETS '23: Proceedings of the 25th International ACM SIGACCESS Conference on Computers and Accessibility. 2023; 30: 1–21.
Mehmood R, Khan S. Enhancing accessibility for visually impaired users through AI-based image processing. J Vis Impair Blindness. 2019; 13(6): 502–515.
Faisal F, Abdullah M. Towards AI accessibility tools for blind and visually impaired people: a review. J Artif Intell Res. 2020; 69(2): 123–145.
Khan MA, Paul P, Rashid M, Hossain M, Ahad MA. An AI-based visual aid with integrated reading assistant for the completely blind. IEEE Trans Hum Mach Syst. 2020; 50(6): 507–517. [CrossRef]
Patil HV, Bansode PK. Computer vision in AI systems for blind assistance: an overview. J Comp Vis Apps. 2020; 31(2): 233–249.
Iqbal I, Imran M, Mehmood U. AI-based assistive technologies for the visually impaired: a survey. IEEE Access. 2021; 9: 10456–10478.
Hong AL, Smith BK. Deep learning for the blind: technologies, trends and future directions. Proc IEEE, 2021; 109(4),636–648.
Walle H, de Runz C, Series B, Venturini G. A survey on recent advances in AI and vision-based methods for helping and guiding visually impaired people. Appl Sci (Basel). 2022; 12(5): 2308. [CrossRef]
Chaple S, Raut V, Patni JC, et al. Artificial intelligence on visually impaired people: a comprehensive review. In 2024 5th International Conference on Intelligent Communication Technologies and Virtual Mobile Networks (ICICV). 2024; 11: 304–308.
O'Sullivan D, O'Regan NA, Timmons S. Validity and reliability of the 6-item cognitive impairment test for screening cognitive impairment: a review. Dement Geriatr Cogn Disord. 2016; 42: 42–49. [CrossRef] [PubMed]
Bruijning JE, van Rens G, Knol D, et al. Psychometric analyses to improve the Dutch ICF Activity Inventory. Optom Vis Sci. 2013; 90: 806–819. [CrossRef] [PubMed]
Scherer MJ, Craddock G. Matching Person & Technology (MPT) assessment process. Technol Disabil. 2002; 14: 125–131. [CrossRef]
Owsley C, McGwin G, Jr, Sloane ME, et al. Timed instrumental activities of daily living tasks: relationship to visual function in older adults. Optom Vis Sci. 2001; 78: 350–359. [CrossRef] [PubMed]
Culham LE, Chabra A, Rubin GS. Clinical performance of electronic, head-mounted, low-vision devices. Ophthalmic Physiol Opt. 2004; 24: 281–290. [CrossRef] [PubMed]
Dougherty BE, Martin SR, Kelly CB, et al. Development of a battery of functional tests for low vision. Optom Vis Sci. 2009; 86: 955–963. [CrossRef] [PubMed]
Latham K, Usherwood C. Assessing visual activities of daily living in the visually impaired. Ophthalmic Physiol Opt. 2010; 30(1): 55–65. [CrossRef] [PubMed]
Finger RP, McSweeney SC, Deverell L, et al. Developing an instrumental activities of daily living tool as part of the low vision assessment of daily activities protocol. Invest Ophthalmol Vis Sci. 2014; 55: 8458–8466. [CrossRef] [PubMed]
Haymes SA, Johnston AW, Heyes AD. The development of the Melbourne low-vision ADL index: a measure of vision disability. Invest Ophthalmol Vis Sci. 2001; 42(6): 1215–1225. [PubMed]
Starke SD, Golubova E, Crossland MD, et al. Everyday visual demands of people with low vision: a mixed methods real-life recording study. J Vis. 2020; 20(9): 3. [CrossRef] [PubMed]
Tan ACS, Peterson CL, Htoon HM, et al. Development and validation of performance-based assessment of daily living tasks in age-related macular degeneration. Transl Vis Sci Technol. 2024; 13(6): 9. [CrossRef] [PubMed]
Brooke J. SUS-A quick and dirty usability scale. Usability Eval Ind. 1996; 189: 4–7. Available at: https://www.researchgate.net/publication/228593520_SUS_A_quick_and_dirty_usability_scale.
Bangor A, Kortum PT, Miller JT. An empirical evaluation of the System Usability Scale. Int J Hum Comput Interact. 2008; 24(6): 574–594. [CrossRef]
Benjamini Y, Hochberg Y. Controlling the false discovery rate—a practical and powerful approach to multiple testing. J R Stat Soc B. 1995; 57: 289–300. [CrossRef]
van der Aa HPA, Garcia-Peña F, Van Nispen R, et al. Performance on clinical outcomes, activities of daily living, and user experience on head-mounted displays for people with vision impairment. Ophthal Physiol Optics. 2024; 44(5): 840–853. [CrossRef]
Gupta L, Varma N, Agrawal S, et al. Approaches in assistive technology: a survey on existing assistive wearable technology for the visually impaired. In: Pandian A., Fernando X., Islam S.M.S. (eds) Computer Networks, Big Data and IoT. Lecture Notes on Dat Engineering and Communications Technologies, vol 66. Singapore: Springer; 2021.
Kumar S, Mishra D, Ganie S, et al. Artificial intelligence solutions for the visually impaired: a review. In: Handbook of Research on AI and Knowledge Engineering for Real-Time Business Intelligence. Hershey, PA: IGI Global; 2023: 198–207.
Chape S, Raul V, Patni JC, et al. Artificial intelligence on visually impaired people: a comprehensive review. 5th International Conference on Intelligent Communication Technologies and Virtual Mobile Networks. 2024:11(3): 304–308.
Granquist C, Sun SY, Montezuma SR, et al. Evaluation and comparison of artificial intelligence vision aids: Orcam MyEye 1 and Seeing AI. J Vis Impair Blind. 2021; 115(4): 277–285. [CrossRef]
Nguyen X, Koopman J, van Genderen MM, et al. Artificial vision: the effectiveness of the OrCam in patients with advanced inherited retinal dystrophies. Acta Ophthal. 2022; 100: e986–e9933. [CrossRef]
Amore F, Silvestri V, Guidobaldi M, et al. Efficacy and patients’ satisfaction with the ORCAM MyEye Device among visually impaired people: A multicenter study. J Med Sys. 2023; 47: 11. [CrossRef]
Bhagat S, Jishi P, Agarwal A, et al. Accessibility evaluation of major assistive mobile applications for the visually impaired. ITU J Future Evolv Technol. 2023; 4(4): 631–643. [CrossRef]
Szlyk JP, Seiple W, Fishman GA, et al. Perceived and actual performance of daily tasks: relationship to visual function tests in individuals with retinitis pigmentosa. Ophthalmology. 2001; 108(1): 65–75. [CrossRef] [PubMed]
Kartha A, Singh RK, Bradley C. Self-reported visual ability versus task performance in Individuals with ultra-low vision. Trans Vis Sci Technol. 2023; 12(10): 14. [CrossRef]
Elliott DB, Trukolo-Ilic M, Strong JG, et al. Demographic characteristics of the vision-disabled elderly. Invest Ophthalmol Vis Sci. 1997; 38: 2566–2575. [PubMed]
Owsley C, McGwin G, Jr, Lee PP, et al. Characteristics of low-vision rehabilitation services in the United States. Arch Ophthalmol. 2009; 127(5): 681–689. [CrossRef] [PubMed]
Brown JC, Goldstein JE, Chan TL, et al. Characterizing functional complaints in patients seeking outpatient low-vision services in the United States. Ophthalmology. 2014; 121(8): 1655–1662. [CrossRef] [PubMed]
Smallfield S, Clem K, Myers A. Occupational therapy interventions to improve the reading ability of older adults with low vision: a systematic review. Am J Occup Ther. 2013; 67: 288–295. [CrossRef] [PubMed]
Wittich W, Lorenzini MC, Markowitz SN, et al. The effect of a head-mounted low vision device on visual function. Optom Vis Sci. 2018; 95: 774–784. [CrossRef] [PubMed]
Miller A, Crossland MD, Macnaughton J, et al. Are wearable electronic vision enhancement systems (wEVES) beneficial for people with age-related macular degeneration? A scoping review. Ophthalmic Physiol Opt. 2023; 43(4): 680–701. [CrossRef] [PubMed]
Taylor JJ, Bambrick R, Brand A, et al. Effectiveness of portable electronic and optical magnifiers for near vision activities in low vision: a randomised crossover trial. Ophthalmic Physiol Opt. 2017; 37(4): 370–384. [CrossRef] [PubMed]
Kinateder M, Gualtieri J, Dunn MJ, et al. Using an augmented reality device as a distance-based vision aid-promise and limitations. Optom Vis Sci. 2018; 95(9): 727–737. [CrossRef] [PubMed]
Yeo JH, Bae SH, Lee SH, et al. Clinical performance of a smartphone-based low vision aid. Sci Rep. 2022; 12(1): 10752. [CrossRef] [PubMed]
Schmidt DC, Kjolholm CDB, Torner Jordana J, et al. At-home trial of four different head-mounted displays in visual rehabilitation of people with Stargardt disease. Clin Optom (Auckl). 2023; 15: 271–281. [CrossRef] [PubMed]
Ortiz A, Chung ST, Legge GE, et al. Reading with a head-mounted video magnifier. Optom Vis Sci. 1999; 76(11): 755–763. [CrossRef] [PubMed]
Weckerle P, Trauzettel-Klosinski S, Kamin G, et al. Task performance with the Low Vision Enhancement System (LVES). Vis Impair Res. 2000; 2(3): 152–162. [CrossRef]
Crossland MD, Starke SD, Imielski P, et al. Benefit of an electronic head-mounted low vision aid. Ophthalmic Physiol Opt. 2019; 39(6): 422–431. [CrossRef] [PubMed]
Peterson RC, Wolffsohn JS, Rubinstein M, et al. Benefits of electronic vision enhancement systems (EVES) for the visually impaired. Am J Ophthalmol. 2003; 136(6): 1129–1135. [CrossRef] [PubMed]
Tripepi G, Jager KJ, Dekker FW, et al. Selection bias and information bias in clinical research. Nephron Clin Pract. 2010; 115(2): c94–c99. [CrossRef] [PubMed]
Martiniello N, Eisenbarth W, Lehane C, et al. Exploring the use of smartphones and tablets among people with visual impairments: are mainstream devices replacing the use of traditional visual aids? Assist Technol. 2019; 34(1): 34–45. [CrossRef] [PubMed]
Figure 1.
 
Comparison of AAII success versus participant success.
Figure 1.
 
Comparison of AAII success versus participant success.
Figure 2.
 
Graphical representation of devices based on ORs and the Post hoc Wilcoxon Signed-Rank findings.
Figure 2.
 
Graphical representation of devices based on ORs and the Post hoc Wilcoxon Signed-Rank findings.
Figure 3.
 
Distribution of perceived use for AAIIs.
Figure 3.
 
Distribution of perceived use for AAIIs.
Table 1.
 
The Best Settings for the Tasks Determined Prior to Testing
Table 1.
 
The Best Settings for the Tasks Determined Prior to Testing
Table 2.
 
Scoring Criteria for the Tasks
Table 2.
 
Scoring Criteria for the Tasks
Table 3.
 
Participant Characteristics (n = 25)
Table 3.
 
Participant Characteristics (n = 25)
Table 4.
 
Odds Ratios of PVL Being Able to Perform a Task
Table 4.
 
Odds Ratios of PVL Being Able to Perform a Task
Table 5.
 
Wilcoxon Signed-Rank Test Results Comparing the Performance of an AAII to Baseline, Shown Only When the Overall Friedman Test was Statistically Significant (n = 25, P ≤ 0.05)
Table 5.
 
Wilcoxon Signed-Rank Test Results Comparing the Performance of an AAII to Baseline, Shown Only When the Overall Friedman Test was Statistically Significant (n = 25, P ≤ 0.05)
Table 6.
 
P Values Associated With Predicting Performance at Baseline
Table 6.
 
P Values Associated With Predicting Performance at Baseline
Table 7.
 
P Values Associated With Predicting Change in Performance
Table 7.
 
P Values Associated With Predicting Change in Performance
×
×

This PDF is available to Subscribers Only

Sign in or purchase a subscription to access this content. ×

You must be signed into an individual account to use this feature.

×