Published on in Vol 13 (2025)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/67711, first published .
Outcomes of an App-Based Intervention to Target Naming Among Individuals With Poststroke Aphasia: Virtual Randomized Controlled Trial

Outcomes of an App-Based Intervention to Target Naming Among Individuals With Poststroke Aphasia: Virtual Randomized Controlled Trial

Outcomes of an App-Based Intervention to Target Naming Among Individuals With Poststroke Aphasia: Virtual Randomized Controlled Trial

1Department of Communication Sciences and Disorders, University of Alberta, 2-70 Corbett Hall, Edmonton, AB, Canada

2Department of Speech-Language Pathology, University of Toronto, Toronto, ON, Canada

3KITE Research Institute, Toronto Rehabilitation - University Health Network, University of Toronto, Toronto, ON, Canada

4Nuromedia GmbH, Cologne, Germany

5Quality and Usability Lab, Technische Universitat Berlin, Berlin, Germany

6Center for Technology Experience, AIT Austrian Institute of Technology, Vienna, Austria

7Immersive Reality Lab, Hamm-Lippstadt University of Applied Sciences, Hamm, Germany

Corresponding Author:

Esther S Kim, PhD


Background: People with aphasia present with language and communication deficits, most notably in lexical retrieval (naming). Although positive outcomes in naming have been observed following speech-language treatment, many individuals with aphasia continue to face impairments after the acute phase of rehabilitation. Mobile app–based therapies are increasingly being used by speech-language pathologists in the rehabilitation of people with aphasia as an adjunct to or in lieu of traditional in-person therapy approaches. These apps can increase the intensity of treatment and have been shown to result in meaningful outcomes across several domains.

Objective: VoiceAdapt is a mobile therapy app addressing naming impairments, designed within a user-centered design framework. The VoiceAdapt app uses two evidence-based lexical retrieval treatments to engage people with aphasia to improve their naming abilities through interaction with the app. The purpose of this study was to conduct a randomized controlled trial to examine the preliminary clinical efficacy of training with VoiceAdapt on the language and communication outcomes of people with aphasia.

Methods: A two-arm, waitlist-controlled, crossover group randomized controlled trial was conducted at two sites within Canada. During the intervention phase, participants completed 5 weeks of independent training with the app, which involved naming practice using Semantic Features Analysis and Phonological Components Analysis. The primary outcome measure was naming performance (Boston Naming Test); secondary outcomes included measures of overall language and naming (Western Aphasia Battery-Revised), communication (Communication Effectiveness Index), and quality of life (Stroke and Aphasia Quality of Life Scale-39).

Results: A total of 37 people with aphasia in the chronic stages (average 4.6 y postonset of aphasia) participated in this study. Participants used the app for an average of 20 hours over the 5-week intervention phase. Training with VoiceAdapt resulted in an increase of 1.6 points on the Boston Naming Test (Cohen d=0.3). Evidence for improved naming was also observed on trained items, as well as subtests of naming or word-finding on the WAB-R. Training with the app also resulted in a significant increase in participants’ perceptions of their communication quality of life (increase of 0.1 points; Cohen d=0.3), but no other measures (WAB-R Aphasia Quotient, Communicative Effectiveness Index) were significant.

Conclusions: Individuals with aphasia who used the VoiceAdapt app for 5 weeks to target naming skills demonstrated measurable gains in naming and communication-based quality of life. Notably, these changes were observed in a remotely delivered program, in participants who were in the chronic stages of aphasia. These findings inform the profession on the use of app-based home therapy programs as an accessible, cost-effective option for individuals in the chronic stages of recovery who often have limited options for rehabilitation.

Trial Registration: ClinicalTrials.gov NCT04108364; https://clinicaltrials.gov/ct2/show/NCT04108364

International Registered Report Identifier (IRRID): RR2-10.2196/30621

JMIR Mhealth Uhealth 2025;13:e67711

doi:10.2196/67711

Keywords



Background

Aphasia, an acquired language disorder most commonly caused by stroke, impacts linguistic expression and comprehension in both written and oral modalities. People with aphasia may have impairments in their ability to speak, understand, read, and write. Thus, aphasia can severely limit activities and restrict participation in everyday activities [1]. Nearly one-third of individuals who experience a stroke have aphasia [2,3], and its presence has been associated with poorer outcomes in both acute and chronic periods following stroke [4,5].

Speech-language pathologists (SLPs) provide rehabilitation services for people with aphasia, often focusing on remediation of the language impairment or training of compensatory strategies. Such interventions have been shown to improve outcomes for people with aphasia [6,7], but constraints on health care resources often prevent people with aphasia from receiving these services, especially in the chronic stages. Recently, the proliferation of mobile app–based therapies available to deliver speech-language interventions has meant that people with aphasia can receive prolonged rehabilitation in an efficient, cost-effective manner. SLPs have increasingly been using such apps as adjuvant therapy approaches, or even in lieu of traditional (face-to-face) therapy, with positive outcomes observed across several domains [8,9]. One benefit of these apps is that people with aphasia can increase the intensity of their rehabilitation, a factor that is associated with greater long-term recovery [7,10-12]. In addition, app-based interventions can be self-delivered by people with aphasia and still yield positive language and communication outcomes [13-15]. Moreover, app-based interventions allow for therapy content to be adapted according to the user or patient state, due to the ability to measure multiple parameters of user interactions with touch-based devices.

The research evidence investigating the feasibility, acceptability, and clinical utility of tablet-based speech-language therapy apps for aphasia has increased in recent years, with inconclusive results. In part, this may relate to the chronicity of individuals with aphasia included in these studies. For instance, Ruiz et al [16] found no difference between an intervention group provided with tablet-based therapy in the first week poststroke and a no-treatment control group. However, Mallet et al [17,18] reported that the use of mobile therapy apps for speech-language therapy was feasible in the very early stages of acute stroke recovery and could potentially be used to bridge the gap between discharge from acute care and the start of outpatient therapy. On the other hand, evidence is emerging on the utility of app-based therapy for adults in the chronic stages of acquired aphasia. Across a growing number of studies, individuals with aphasia have shown improvement across a variety of domains, including word production, reading, naming, spoken discourse, quality of life, and overall language [13-15,19-24].

App-based therapy has the potential to increase treatment dosage (total hours), intensity (hours per week), frequency (days per week), and duration (total weeks)—important factors that contribute to positive speech-language therapy outcomes [25]. For instance, participants in a study by Alam et al [22] were found to benefit more from therapy in the condition where an app was provided as an adjunct to therapy (the “study” arm) than in the condition where there was no adjunct (the “control” arm). However, it should be noted that participants in the study arm received more therapy overall due to the addition of the app. Individuals with aphasia who self-managed their own participation using therapy apps in the absence of face-to-face therapy have also demonstrated improved language outcomes [13-15,21]. Indeed, app-based therapy in lieu of traditional therapy, particularly for individuals in the chronic stages of aphasia, could be an important component of optimizing recovery in the context of limited rehabilitation resources. However, further research to understand the impact of apps (alone and in combination with face-to-face rehabilitation) is warranted.

A recent review by Szeto et al [26] on the effect of mobile apps for stroke rehabilitation reported that aphasia therapy apps that resulted in the greatest improvements were ones that incorporated characteristics of face-to-face therapy. These included massed practice (repetition) of goal-oriented, task-specific exercises, user interactivity, and multisensory stimulation. A similar review identifying and evaluating mobile apps for speech therapy in adults with communication disorders (not limited to aphasia) revealed that there was a general lack of engaging and interactive elements in the apps [27]. Increasing the usability of therapy apps could be enhanced by designing apps with input from relevant stakeholders, following the principles of user-centered design (UCD). UCD is a systematic approach to meeting user experience goals and usability through integrating the needs and abilities of end users during the design process [28]. Although the number of therapy apps designed for aphasia is steadily increasing [27,29], very few have been designed following the principles of UCD [30,31], despite increasing evidence supporting the beneficial impacts of UCD on user experience in digital health technologies [32]. Indeed, in their review of 70 therapy apps for adults with communication disorders, Vaezipour et al [27] recommend using a co-design process involving end users (ie, people with aphasia), subject matter experts (ie, SLPs), and incorporating interactive elements and gamification to tailor the app to individual needs.

Development of the VoiceAdapt App

The VoiceAdapt app, developed by an international team of interdisciplinary researchers from the fields of computer science, usability design, and speech-language pathology, is a tablet-based app focused on improving naming abilities in individuals with aphasia. VoiceAdapt instantiates two evidence-based lexical retrieval protocols, Phonological Components Analysis (PCA) [33,34] and Semantic Feature Analysis (SFA) [35,36], into an adaptive, voice-responsive therapy app that people with aphasia can use to improve their naming skills.

Using UCD principles, the team engaged 75 end users (including people with aphasia, clinicians, and caregivers) in designing the app. Over the course of three iteration cycles, end users participated in interviews to elicit user requirements, provided feedback on mock-ups and prototypes, and evaluated the designs with a focus on user experience of multimodal speech and adaptivity [32]. The end result was a mobile speech-language training app for iOS and Android that supported regular training sessions, adaptivity, personalization, and addressed multimodal language goals (reading, understanding, speech production).

Objective

The purpose of this study is to report the efficacy of a beta version of the VoiceAdapt app, within a randomized controlled trial (RCT; ClinicalTrials.gov NCT04108364). Specifically, we examined the impact of training with the VoiceAdapt app on language and communication outcomes in individuals with chronic aphasia. Primary outcomes were naming ability, and secondary outcomes included measures of overall language, communication effectiveness, and quality of life.


Recruitment

Participants were recruited offline between August 2020 and August 2022 in two geographic regions in Canada through the Language Sciences Lab (Department of Speech-Language Pathology, University of Toronto) and Aphasia Research Lab (Department of Communication Sciences and Disorders, University of Alberta), based in Toronto, Ontario, and Edmonton, Alberta, respectively. Information about the study was disseminated to clinicians and programs serving individuals with aphasia. Interested participants reached out directly to study personnel (CW, Edmonton; LL, Toronto) and provided informed consent prior to being enrolled in the study. Communicatively accessible consent forms were used, and all study personnel were trained in supported communication techniques.

Participants enrolled in the study were at least 6 months postonset of a stroke in the left hemisphere, had aphasia of mild to moderate severity (Aphasia Quotient ≥30 on Western Aphasia Battery-Revised [WAB-R] [37]), scored below 75% correct on the Boston-Naming Test [38] at initial assessment, had a prominent verbal expression impairment (based on assessment scores and SLP determination), passed screenings of basic sensory (vision and hearing) and cognitive functioning, and spoke English as a primary language. Participants with severe aphasia were excluded from the study due to significant verbal expression impairment and limited naming abilities in this population. As the app was designed for self-directed treatment, individuals with severe aphasia profiles were unlikely to benefit from this type of intervention. Participants also had access to a computer or tablet and the internet in order to participate in web-based assessment sessions. In addition to meeting all inclusion criteria, participants were excluded from participation if they were currently engaged in individual (one-on-one) speech-language therapy or were currently using speech-language therapy apps. All study procedures (intake, informed consent, and assessments) took place using videoconferencing software, in line with recent advancements indicating the feasibility and reliability of virtual assessment protocols [39,40].

Planned enrollment for the study was 80 participants based on sample size calculations, with interim analyses planned when quarterly enrollment targets were met (ie, 20, 40, and 60 participants) [39]. The sample size calculation was based on medium effect sizes estimated in this study according to the results shown by Varley et al [41]. Group sample sizes of 72 are needed to achieve 80% power to detect a difference in proportions of 0.25 between the null hypothesis and the alternative hypothesis using a 1-tailed t test with a probable α error of 0.05. Accounting for a dropout rate of approximately 10% (based on the dropout rate of Varley et al [41]) from baseline to follow-up, we estimated we would need 79 participants (rounded up to 80). It was determined that the trial would be stopped early if evidence of a treatment effect on the primary outcome measure was observed at the time of interim analyses.

Design

The RCT was a two-arm, randomized, waitlist-controlled, crossover group design. All participants completed initial assessments and then were randomized into intervention or treatment-deferred control groups. Participants in the intervention group completed training with the VoiceAdapt for 5 weeks during Phase 1 of the study, while those in the control group continued with usual care. Following midpoint assessments for all participants, those in the treatment-deferred control group crossed over into receiving treatment, while those in the intervention group crossed over into the control phase. Final assessments took place following Phase 3, approximately 13 weeks following study enrollment (Figure 1). Although the study was originally intended to be conducted in person, the COVID-19 pandemic necessitated a move to all procedures being conducted virtually.

Figure 1. Study design.

An individual not affiliated with the study created a computer-generated randomization sequence using Excel (version 2007; Microsoft Corp), stratified by site with 1:1 allocation using random block sizes of 10. Each cell containing the allocation status was concealed using black fill in the spreadsheet. One member of the study team (CW) revealed allocation status to participants (Edmonton site) or to the research coordinator at the Toronto site (LL), who then related the status to participants following baseline assessment. Assessments were completed by SLPs (Edmonton site) or speech-language pathology graduate students under the supervision of a SLP (Toronto site) who were blinded to the treatment condition of participants. Participants were told not to reveal to assessors which group they were in.

Intervention

The VoiceAdapt app intervention comprises naming exercises based on SFA and PCA, two evidence-based anomia treatments, delivered on a mobile tablet. A color picture is presented along with a series of prompts designed to engage the people with aphasia in providing semantic features (eg, “What is it used for?” [Action]; “What does it look like?” [Properties]), or phonological components (eg, “What sound does it start with?” [First sound]; “What is a rhyming word?” [Rhyme]). People with aphasia randomized to the VoiceAdapt app treatment group were instructed to practice using the app for 1 hour per day every weekday (5 days per week) for 5 weeks for a total of 25 hours of intervention. Participants were provided with tablets with the app loaded, or could download the app from a link provided by the researchers onto their own devices. Participants selected interests at initial app setup, which then tailored the practice items to each individual. Research staff provided orientation to the app and monitored participants’ app use (number of minutes per week) through the app portal. Research staff (CW and LL) checked in with participants through email, or phone, or video calls on a weekly basis, providing practice reminders and answering questions.

Outcome Measures

The primary outcome measure was the Boston Naming Test (BNT) [38] to measure naming, secondary outcome measures were WAB-R [37] and Stroke and Aphasia Quality of Life Scale (SAQOL-39g) [42] to measure quality of life, and Communication Effectiveness Index (CETI) [43] to measure the person with aphasia’s perceptions of communication effectiveness. Both the person with aphasia and their caregiver (where applicable) completed the CETI. These outcomes were assessed at baseline and after intervention phases (Figure 1). In addition, at the final assessment session, the Study Stimulus Naming Test (SSNT), a measure of performance on trained and untrained items on the VoiceAdapt app, the System Usability Scale (SUS) [44], and a 5-question posttreatment questionnaire specific to app use were administered to all participants.

Fidelity and Reliability

To ensure fidelity of procedures across both sites, a standardized intake and assessment training protocol was developed. Assessors at both sites completed all steps in the training protocol including reading background papers [39,40], reviewing intake and consent forms and manuals for all assessments, observing the study coordinator or primary investigators (licensed SLPs) doing a complete assessment session, and finally, being observed by the study coordinators or primary investigators to ensure protocol adherence prior to completing assessments on their own.

Administration of the primary outcome measure (BNT) was video recorded through the Zoom (Zoom Video Communications, Inc) videoconferencing platform. A total of 16 assessment sessions (just under 15% of the sample) were randomly selected for assessing the reliability of scoring. Point-to-point reliability was high overall (Cohen κ=0.91).

Data Management

Study data were collected and managed using REDCap (Research Electronic Data Capture; Vanderbilt University) electronic data capture tools hosted at the University of Alberta [45,46]. REDCap is a secure, web-based software platform designed to support data capture for research studies, providing (1) an intuitive interface for validated data capture, (2) audit trails for tracking data manipulation and export procedures, (3) automated export procedures for seamless data downloads to common statistical packages, and (4) procedures for data integration and interoperability with external sources.

Ethical Considerations

This project was approved by the University of Toronto Health Sciences Research Ethics Board (Protocol Number 37980) and the University of Alberta Health Research Ethics Board (Pro00091587). All participants provided informed consent following an explanation of research procedures in a communicatively accessible manner by research staff who were trained in supported communication techniques. All data were anonymized and deidentified. Participants were not provided with monetary compensation for their participation in this study.


Participants

A total of 38 participants with aphasia (13 female) were recruited and randomized across the two sites. In total, 20 individuals were randomly assigned to receive training with VoiceAdapt first, while 18 individuals were randomly assigned to the deferred treatment (control) condition. One individual randomized to the control group dropped out of the study shortly after starting treatment due to frustrations with the automatic speech recognition (ASR) aspects of the app. As the withdrawn participant only contributed two baseline assessments and did not contribute any treatment information, their data were not included in analyses. A total of 37 participants completed both phases of the study (Figure 2). The trial ended early due to results from interim analyses, as well as slower-than-expected recruitment due to challenges following the COVID-19 pandemic (see details below). Participants had a mean age of 62.2 (SD 13.9) years and 15.1 (SD 3.2) years of education and were an average of 4.6 (SD 4.2) years postonset of aphasia. No differences in age, education, years postonset, or baseline WAB-AQ/BNT were observed between groups (Table 1). Site differences were present for age and education across sites, with participants recruited through the University of Alberta being younger and having less education than those recruited through the University of Toronto. The number of participants across allocation conditions and mean demographic information are presented in Table 1.

Figure 2. Participant flowchart.
Table 1. Participant mean age, education, years post-onset, WAB-AQa, and BNTb scores across allocationsc.
CharacteristicIntervention (n=20; 6F), mean (SD)Control (n=17; 7F), mean (SD)Total (n=37; 13F), mean (SD)
Age (years)59.9 (15.3)64.8 (11.9)62.2 (13.9)
Education14.7 (2.6)15.5 (3.7)15.1 (3.2)
Years postonset3.5 (2.5)5.8 (5.3)4.6 (4.2)
WAB-AQa68.3 (15.2)73.6 (14.0)70.7 (14.7)
BNTb25.0 (12.1)23.5 (12.5)24.3 (12.1)

aWAB-AQ: Western Aphasia Battery-Revised Aphasia Quotient.

bBNT: Boston Naming Test.

cIntervention refers to participants randomized to receive VoiceAdapt training first; control refers to participants randomized to the treatment-deferred control group.

Analytic Strategy

The 2×2 crossover trial with one baseline was analyzed using Bayesian inference to estimate the parameters of a mixed-effects regression model. Using mixed-effects regression can account for hierarchical sampling of participants within sites. Using this crossover design allowed estimation of treatment effect, time effect, sequence effect, and carryover effect. Bayesian inference was chosen for accurate multiple comparisons [47]. For each outcome, we modeled treatment effect, time effect, and the outcome at baseline covariate as a normal distribution with unknown variance using a noninformative flat prior for the mean parameter and a noninformative inverse-gamma prior for the variance parameter that was sampled independently of the other parameters. This Bayesian model was fit using the adaptive Metropolis-Hastings Markov chain Monte Carlo simulation method implemented in Stata (version 16; StataCorp LLC). The proportion of the posterior distribution of the treatment effect that was greater than zero was interpreted as the probability that the outcome increased. For such interval hypothesis testing, a probability of 0.95 or higher was considered strong evidence of improvement during VoiceAdapt training. In these data, we found no statistical evidence for a sequence effect or a carryover effect, so these were dropped from our model. We also checked for an effect of app use and found no statistical evidence for an effect, so it was also dropped from the model. All analyses were carried out on the 37 participants who completed both phases of the study unless otherwise indicated.

Our initial sample size calculation was based on using frequentist approaches. There were several realities that necessitated a change to Bayesian analyses. Due to recruitment challenges in the virtual context, we needed to do interim analyses more frequently than outlined in our protocol. Bayesian analysis allows researchers to conduct interim analyses without being penalized because decisions at analysis points are based on the posterior probabilities of treatment effects [47]. Bayesian analysis also provides a straightforward means of analyzing multiple outcomes multiple times. While our power analysis was based on the effect size of one strong primary outcome, BNT, in reality, our different outcomes had differential effects based on our interim analyses. Given that we found significant results with our reduced sample size, and the recruitment challenges referenced above, we decided to end the trial early.

Primary Outcome

The primary outcome was the BNT to measure change in naming performance as a result of training with the VoiceAdapt app. Participants increased an average of 1.6 points on the BNT (equal-tailed 95% credible interval −0.3 to 3.6; Cohen d=0.3) when training with VoiceAdapt (Table 2).

Table 2. Median and equal-tailed 95% credible interval of the posterior probability distribution of each parameter of the cross-over model for the primary outcome (BNTa).
ParameterBNT, median (95% credible interval)
Fixed-effect parametersb
VoiceAdapt effect1.6 (−0.3 to 3.6)c
Time effect1.5 (−0.4 to 3.3)
Outcome at baseline covariate1.0 (0.8 to 1.1)
Constant2 (−3 to 6)
Random-effect parameters
Among participant variance21 (8 to 39)
Residual variance17 (10 to 28)

aBNT: Boston Naming Test.

bProportion of Bayesian posterior probability distribution of VoiceAdapt effect greater than zero, Cohen d: 0.954.

cCohen d=0.3.

Other Outcomes: Naming Ability

Additional analyses with other measures related specifically to naming abilities were examined further using the same analysis procedures. These included the Object Naming subtest on the WAB-R (20 item naming test); the Naming and Word Finding subscore of the WAB-R (comprising combined scores from Object Naming, Word Fluency, Sentence Completion, and Responsive Speech subsets); and the SSNT, a 322 item naming measure comprising both trained and untrained words programmed into the VoiceAdapt app. As seen in Table 3, training with VoiceAdapt resulted in an increase of 3 points on the Object Naming subtest (equal-tailed 95% credible interval 1-5) and an increase of 2.6 points on the Naming and Word Finding subtest (equal-tailed 95% credible interval 0.15-5), representing small effects (Cohen d=0.4 and 0.3, respectively). Interestingly, the SSNT demonstrated the greatest change as a result of training with VoiceAdapt, with a median change score of 13 (equal-tailed 95% credible interval 7-19; Cohen d=0.9).

Table 3. Median and equal-tailed 95% credible interval of the posterior probability distribution of each parameter of the cross-over model for additional outcomes (SSNTa, WAB-Rb Object Naming, WAB-R Naming, and Word Finding total).
ParameterSSNT Object Naming (WAB-R) Naming and Word Finding (WAB-R)b
Fixed-effect parameters
VoiceAdapt effect, median (95% credible interval)13 (7 to 19)c3 (1 to 5)d2.6 (0.15 to 5)e
Proportion of Bayesian posterior probability distribution of VoiceAdapt effect greater than zero, Cohen d0.9990.9980.980
Time effect, median (95% credible interval)13 (6 to 19)−0.6 (−2.6 to 1.4)−0.6 (−3.0 to 1.9)
Outcome at baseline covariate, median (95% credible interval)1.05 (0.9 to 1.2)1.0 (0.8 to 1.1)1.0 (0.9 to 1.1)
Constant, median (95% credible interval)1 (20 to 18)1 (−5 to 7)1 (9 to 6)
Random-effect parameters, median (95% credible interval)
Among participant variance375 (188 to 668)23 (10 to 43)31 (11 to 60)
Residual variance168 (103 to 277)19 (11 to 30)28 (17 to 45)

aSSNT: Study Stimulus Naming Test.

bWAB-R: Western Aphasia Battery-Revised.

cCohen d=0.9.

dCohen d=0.4.

eCohen d=0.3.

Secondary Outcomes

There was no significant effect of VoiceAdapt training on overall language abilities, as evidenced by a mean change in WAB-R Aphasia Quotient of 0.4 (Cohen d=0.01). VoiceAdapt training resulted in a significant increase in participants’ perceptions on the SAQOL-39g Communication Quality of Life subscale (increase of 0.1 points; Cohen d=0.3), representing a small effect, but not on overall Quality of Life (SAQOL-39g mean score decrease of 0.01 points; Cohen d=0.1). Measures of communication effectiveness as measured by the participant (CETI-participant mean score=−5; Cohen d=0.1) or their partners (CETI-partner mean score=0.1; Cohen d=0.2) were not found to be significantly improved by VoiceAdapt training (Table 4).

Table 4. Median and equal-tailed 95% credible interval of the posterior probability distribution of each parameter of the cross-over model for the secondary outcomes (WAB-Ra aphasia quotient, SAQOL-39gb mean score, SAQOL-39g Communication score, CETIc, and CETI-Partner)d.
ParameterWAB-R AQ SAQOL-39g meanSAQOL-39Comm CETI CETI-Partner
Fixed-effect parameters
VoiceAdapt effect, median (95% credible interval)0.4 (−1.1 to 1.8)e−0.005 (−0.1 to 0.1)f0.1 (−0.1 to 0.3)g−5 (−12 to 4)f0.1 (−7 to 7)h
Proportion of Bayesian posterior probability distribution of VoiceAdapt effect greater than zero, Cohen d0.7070.4680.8460.1420.506
Time effect, median (95% credible interval)0.4 (−1.1 to 1.8)0.01 (−0.1 to 0.1)−0.05 (−0.3 to 0.1)1 (−7 to 10)9 (2 to 16)
Outcome at baseline covariate, median (95% credible interval)1.0 (0.9 to 1.1)0.9 (0.7 to 1.1)0.8 (0.5 to 1.0)0.6 (0.4 to 0.9)0.8 (0.5 to 1)
Constant, median (95% credible interval)−1 (−7 to 4)0.4 (−0.4 to 1.2)0.9 (0.1 to 1.6)44 (22 to 66)18 (−3 to 38)
Random-effect parameters, median (95% credible interval)
Among participant variance7 (1 to 15)0.1 (0.05 to 0.2)0.2 (0.1 to 0.3)231 (37 to 493)137 (12 to 311)
Residual variance10 (6 to 17)0.06 (0.03 to 0.1)0.2 (0.1 to 0.3)309 (175 to 526)151 (76 to 289)

aWAB-R: Western Aphasia Battery-Revised.

bSAQOL-39g: Stroke and Aphasia Quality of Life Scale.

cCETI: Communication Effectiveness Index.

dSample includes three longitudinal assessments of 37 participants, except for CETI Partner outcome because 7 participants did not have a participating partner.

eCohen d=0.01.

fCohen d=0.1.

gCohen d=0.3.

hCohen d=0.2.

Participant Perceptions of VoiceAdapt App

Participants used the app for an average of 19.5 (SD 8.6) hours over the course of the 5-week VoiceAdapt intervention phase. The SUS [44], a 10-item questionnaire, was used to evaluate participant perceptions of the ease of using the VoiceAdapt app. Each question has one of five response options, ranging from strongly disagree to strongly agree; items are scored (accounting for positive or negative valence) and multiplied by 2.5 to yield an overall usability score out of 100. The average SUS score for all participants was 61.2 (SD 18.9). The mean score for each question ranged from 2.2 to 2.6, indicating participants felt neutral or agreed with statements pertaining to the usability of the app. In addition to the SUS, a 5-question posttreatment questionnaire was administered to all participants, asking about their experiences with the app. Mean responses to the posttreatment questionnaire are provided in Table 5.

Table 5. Average responses to posttreatment questionnaires by study participants.
QuestionAverage response (1=strongly disagree; 5=strongly agree)
I really liked the app4.1
The app was easy to use3.9
Training with the app helped my communication3.9
I would use this app if it were still available3.8
I would be willing to pay a fee for the VoiceAdapt app3.2

Principal Findings

This RCT was designed to assess the efficacy of the VoiceAdapt app, a tablet-based naming therapy program for people with aphasia. The app implemented evidence-based treatment principles into an adaptive, personalizable app that addressed multimodal language stimulation with the aim of stimulating language production, specifically single-word naming.

Training with the VoiceAdapt app resulted in improved single-word naming performance. The change in naming after treatment on the primary outcome measure (ie, BNT) was minimal, though statistically robust. Converging evidence for improved naming also came from observed improvements on subtests of the WAB-R related specifically to naming abilities (Naming and Word Finding, Object Naming), as well as the SSNT, a set of 322 items (including trained and untrained words). Notably, improvements on the SSNT were accompanied by a large effect size. These results extend the literature base on the efficacy of SFA and PCA for improving naming [48,49], as well as the growing number of studies reporting positive outcomes from app-based instantiations of these therapy protocols [15,18,23].

Importantly, participants’ perception of their quality of life as it related specifically to their communication abilities (SAQOL-39g Communication score) also revealed a small but significant improvement after VoiceAdapt training, indicating that over and above changes in actual numbers of words produced after training, participants felt more positive about their communicative abilities (eg, speaking in general, finding the right word, and being understood by others). In contrast to the participants’ perceived communication quality of life, communicative effectiveness as measured by the CETI was not perceived to have improved after VoiceAdapt treatment by either the participants or their family members. The concept of communicative effectiveness extends beyond performance on linguistic tasks to situational interactions in the real world [50]. Implementing exercises or activities in future apps that provide the opportunity to improve participation (eg, in conversations) and social integration, perhaps by incorporating digital networking for peer support [51] or virtual reality [52], may in the future lead to generalization to domains of communication effectiveness, overall functioning, and well-being.

Despite improvements in naming and communication quality of life, secondary outcomes encompassing overall language and communication or quality of life did not change significantly. In part, this may be related to the broad domains they were measuring. For instance, the WAB-R AQ calculation includes language comprehension, repetition, and spoken discourse in addition to naming. Similarly, the mean SAQOL-39g score includes evaluation of activities of daily living (eg, preparing food and getting dressed), physical activities (eg, walking and standing), and emotional or mental health items, in addition to communication items. Therefore, despite current recommendations for reporting outcomes for aphasia treatment research [53], it may have been unrealistic to expect change on these two comprehensive measures when only one domain, naming, was implemented in the training app. Incorporating different and varied exercises in the future that target additional language domains may lead to generalization to other language domains.

Recent reviews of the effectiveness of speech-language therapy have reported that intensity (dosage, frequency, and duration) of treatment is associated with the largest gains [25], particularly for individuals in the chronic phases of aphasia. However, in this study, app use was not found to be associated with improvement. In addition, although participants were tasked with practicing for 1 hour daily, average app use was less than that. Nonetheless, our participants, who were an average of 4.6 years (SD 4.2) postonset of aphasia, self-managed their therapy with the app with periodic check-ins from the intervention team, supporting prior reports [15,21,26] that tablet-based therapy apps can be an effective option for people with aphasia in the chronic stages to continue to work on their language [54].

One distinctive feature of the VoiceAdapt app was in its design, incorporating UCD principles with input from end users (therapists, people with aphasia, etc), consistent with findings demonstrating that UCD can significantly enhance user engagement and intervention outcomes in technology-based health apps [27,55]. The app included several features recommended in Vaezipour et al’s [27] review of mobile apps. Specifically, the app was tailored to participants’ interests in that they were able to select categories of items they wished to practice. Participants also interacted with the app to select multimodal cues as needed, providing individualized support according to their performance. Despite this, participants’ ratings of app favorability were in the midrange (mean SUS score 61.2, SD 18.9). Other studies investigating therapy apps have reported higher scores [56,57], but there may be other factors contributing to these higher scores (eg, more clinician involvement and different app content). Nonetheless, improvements could be made to further enhance the effectiveness and usability of the app. One limitation of the app was that the ASR software built into the app was not always accurate, particularly given the common co-occurrence of accompanying motor speech impairments with poststroke aphasia. Indeed, frustrations with ASR contributed to the one participant’s withdrawal we had in this study. As this is a rapidly evolving area of study and development, future iterations of the app may perform better in this regard. Incorporation of additional features demonstrated to improve engagement and interactivity [26,27], such as social interaction and gamification, could also result in an improved app in the future. Importantly, future additions and designs should incorporate the wishes, needs, and recommendations of people with aphasia [31].

Limitations and Future Directions

The planned methods, including recruitment activities, assessments, and treatment, were all modified to be delivered virtually due to the COVID-19 pandemic. This may have affected recruitment, as our sample size was less than expected. However, positive results in naming were still obtained. Future research will have to determine whether greater and more extensive improvements might be obtained with a larger sample size. In addition, people with aphasia, on average, did not practice for the full dosage prescribed. This, too, leaves a question for future research of whether greater gains might be obtained with more practice.

The study design did not include a maintenance assessment for all participants postintervention (although individuals in the treatment condition had one maintenance assessment at approximately 6 wk postintervention). This is recognized as a limitation, as long-term maintenance of treatment gains is an important outcome in aphasia rehabilitation. Finally, the nature of this study did not allow for participants to be blinded to the condition. However, all individuals completing assessments at each time point were blinded to participant condition. Future studies could examine the efficacy of training using the VoiceAdapt app compared to other speech-language therapy apps on improving naming performance in individuals with aphasia.

Conclusions

In this RCT, 5 weeks of training in naming single words with the VoiceAdapt app resulted in measurable gains in naming and communication-based quality of life, in contrast to no training in a group of participants in the chronic stages of poststroke aphasia. These changes were seen in a remotely delivered program, in participants who were no longer receiving speech therapy and who practiced their exercises at home using the app.

The findings inform the profession on telerehabilitation practices and monitoring app–based home therapy programs, with a strong recommendation to incorporate UCD methodologies for improved user adherence and therapeutic efficacy [32,55]. In addition, the improvements seen in naming and communicative functioning suggest that instantiating evidence-based treatments such as the one done with this app and which can be found in some others (eg, Tactus and Constant Therapy) can be recommended as an efficacious, easily accessible, and cost-effective method for individuals who are in the chronic stages of recovery and who often have limited options for rehabilitation.

Acknowledgments

The authors would like to acknowledge Anna-Maria Bieber, Stephanie Schwarz, and Till Bieg for their support of the user-centered design of the app, Jevgenijs Danilins for his contributions to app development, and March of Dimes Canada for their support of participant recruitment. This research was funded under the JTC 2017 launched by the Joint Programming Initiative “More Years Better Lives” (JPI MYBL), which was supported by Horizon2020, the EU Framework for Research and Innovation (grant 643850). The VoiceAdapt Consortium was funded by the Canadian Institutes of Health Research grant #155164; the Austrian Research Promotion Agency (FFG), project 861960, and the German Federal Ministry of Education and Research (grant 16SV7995K).

Authors' Contributions

ESK and ER designed the study with the assistance of LL, JNV-A, SM, RS, and RPS. All authors designed the VoiceAdapt app. PM developed the VoiceAdapt app and backend. ESK, ER, LL, and CW recruited participants and managed the randomized controlled trial. SS completed statistical analyses in consultation with ESK and ER. All authors read and approved the final manuscript.

Conflicts of Interest

None declared.

Checklist 1

CONSORT eHEALTH (V 1.6.1) checklist.

PDF File, 10772 KB

  1. Hilari K. The impact of stroke: are people with aphasia different to those without? Disabil Rehabil. 2011;33(3):211-218. [CrossRef] [Medline]
  2. Dickey L, Kagan A, Lindsay MP, Fang J, Rowland A, Black S. Incidence and profile of inpatient stroke-induced aphasia in Ontario, Canada. Arch Phys Med Rehabil. Feb 2010;91(2):196-202. [CrossRef] [Medline]
  3. Flowers HL, Silver FL, Fang J, Rochon E, Martino R. The incidence, co-occurrence, and predictors of dysphagia, dysarthria, and aphasia after first-ever acute ischemic stroke. J Commun Disord. 2013;46(3):238-248. [CrossRef] [Medline]
  4. Lazar RM, Boehme AK. Aphasia as a predictor of stroke outcome. Curr Neurol Neurosci Rep. Sep 19, 2017;17(11):83. [CrossRef] [Medline]
  5. Stransky ML, Jensen KM, Morris MA. Adults with communication disabilities experience poorer health and healthcare outcomes compared to persons without communication disabilities. J Gen Intern Med. Dec 2018;33(12):2147-2155. [CrossRef] [Medline]
  6. Breitenstein C, Grewe T, Flöel A, et al. Intensive speech and language therapy in patients with chronic aphasia after stroke: a randomised, open-label, blinded-endpoint, controlled trial in a health-care setting. Lancet. Apr 15, 2017;389(10078):1528-1538. [CrossRef] [Medline]
  7. Ali M, VandenBerg K, Williams LJ, et al. Predictors of poststroke aphasia recovery. Stroke. 2021. [CrossRef]
  8. Des Roches CA, Kiran S. Technology-based rehabilitation to improve communication after acquired brain injury. Front Neurosci. 2017;11:382. [CrossRef] [Medline]
  9. Braley M, Pierce JS, Saxena S, et al. A virtual, randomized, control trial of a digital therapeutic for speech, language, and cognitive intervention in post-stroke persons with aphasia. Front Neurol. 2021;12:626780. [CrossRef] [Medline]
  10. Bhogal SK, Teasell R, Speechley M. Intensity of aphasia therapy, impact on recovery. Stroke. Apr 2003;34(4):987-993. [CrossRef] [Medline]
  11. Brady MC, Kelly H, Godwin J, Enderby P, Campbell P. Speech and language therapy for aphasia following stroke. Cochrane Database Syst Rev. Jun 1, 2016;2016(6):CD000425. [CrossRef] [Medline]
  12. Cherney LR, Patterson JP, Raymer A, Frymark T, Schooling T. Evidence-based systematic review: effects of intensity of treatment and constraint-induced language therapy for individuals with stroke-induced aphasia. J Speech Lang Hear Res. Oct 2008;51(5):1282-1299. [CrossRef] [Medline]
  13. Braley M, De Oliveira E, Munsell M, et al. A phase ii randomized, virtual, clinical trial of speech therapy app for speech, language, and cognitive intervention in stroke. Arch Phys Med Rehabil. Nov 2020;101(11):e62. [CrossRef]
  14. Palmer R, Dimairo M, Cooper C, et al. Self-managed, computerised speech and language therapy for patients with chronic aphasia post-stroke compared with usual care or attention control (Big CACTUS): a multicentre, single-blinded, randomised controlled trial. Lancet Neurol. Sep 2019;18(9):821-833. [CrossRef]
  15. Stark BC, Warburton EA. Improved language in chronic aphasia after self-delivered iPad speech therapy. Neuropsychol Rehabil. Jul 2018;28(5):818-831. [CrossRef] [Medline]
  16. Ferrete Ruíz EJ, de Torres Chacón MDLR, Perez Sánchez S, et al. Breaking the communication barrier using new technologies in aphasic patients post-stroke. Revista Científica de la Sociedad de Enfermería Neurológica (English ed). Jul 2023;58:14-20. [CrossRef]
  17. Mallet KH, Shamloul RM, Corbett D, et al. RecoverNow: feasibility of a mobile tablet-based rehabilitation intervention to treat post-stroke communication deficits in the acute care setting. PLoS One. 2016;11(12):e0167950. [CrossRef] [Medline]
  18. Mallet KH, Shamloul RM, Lecompte-Collin J, Winkel J, Donnelly B, Dowlatshahi D. Telerehab at home: mobile tablet technology for patients with poststroke communication deficits—a pilot feasibility randomized control trial. J Speech Lang Hear Res. Feb 13, 2023;66(2):648-655. [CrossRef] [Medline]
  19. Des Roches CA, Balachandran I, Ascenso EM, Tripodis Y, Kiran S. Effectiveness of an impairment-based individualized rehabilitation program using an iPad-based software platform. Front Hum Neurosci. 2014;8:1015. [CrossRef] [Medline]
  20. Kurland J, Liu A, Stokes P. Effects of a tablet-based home practice program with telepractice on treatment outcomes in chronic aphasia. J Speech Lang Hear Res. May 17, 2018;61(5):1140-1156. [CrossRef] [Medline]
  21. Heide J, Netzebandt J, Ahrens S, Brüsch J, Saalfrank T, Schmitz-Antonischki D. Improving lexical retrieval with LingoTalk: an app-based, self-administered treatment for clients with aphasia. Front Commun. 2023;8:1210193. [CrossRef]
  22. Alam N, Munjal S, Panda NK, Kumar R, Gupta S. Efficacy of Jellow app as an adjunct to stimulation therapy in improvement in language and quality of life in patients with chronic Broca’s Aphasia. Disabil Rehabil Assist Technol. Jul 2023;18(5):596-602. [CrossRef] [Medline]
  23. Ballard KJ, Etter NM, Shen S, Monroe P, Tien Tan C. Feasibility of automatic speech recognition for providing feedback during tablet-based treatment for apraxia of speech plus aphasia. Am J Speech Lang Pathol. Jul 15, 2019;28(2S):818-834. [CrossRef]
  24. Routhier S, Bier N, Macoir J. Smart tablet for smart self-administered treatment of verb anomia: two single-case studies in aphasia. Aphasiology. Nov 14, 2014;30(2-3):1-21. [CrossRef]
  25. Brady MC, Myzoon A, Vandenberg K, REhabilitation and recovery of peopLE with Aphasia after StrokE (RELEASE) Collaborators. Dosage, intensity, and frequency of language therapy for aphasia: a systematic review-based, individual participant data network meta-analysis. Stroke. Mar 2022;53(3):956-967. [CrossRef] [Medline]
  26. Szeto SG, Wan H, Alavinia M, Dukelow S, MacNeill H. Effect of mobile application types on stroke rehabilitation: a systematic review. J Neuroeng Rehabil. Jan 24, 2023;20(1):12. [CrossRef] [Medline]
  27. Vaezipour A, Campbell J, Theodoros D, Russell T. Mobile apps for speech-language therapy in adults with communication disorders: review of content and quality. JMIR Mhealth Uhealth. Oct 29, 2020;8(10):e18858. [CrossRef] [Medline]
  28. Bevan N, Carter J, Earthy J, Geis T, Harker S. New ISO standards for usability, usability reports and usability measures. 2016. Presented at: International Conference on Human Computer Interaction 2016:268-278; Toronto, Canada. [CrossRef]
  29. Repetto C, Paolillo MP, Tuena C, Bellinzona F, Riva G. Innovative technology-based interventions in aphasia rehabilitation: a systematic review. Aphasiology. Sep 2020;25:1-24. [CrossRef]
  30. Griffel J, Leinweber J, Spelter B, Roddam H. Patient-centred design of aphasia therapy apps: a scoping review. Aphasie Verwandte Geb. 2019;46(2):6-21. URL: https://clok.uclan.ac.uk/id/eprint/30913/ [Accessed 2025-08-11]
  31. Kamaruddin AI, Masrom M, Aziz MAA, Mohd Noor N. End-user needs and perspectives on speech therapy apps for aphasia. Presented at: 2023 IEEE 2nd National Biomedical Engineering Conference (NBEC):127-132; Melaka, Malaysia. [CrossRef]
  32. Schatz R, Kim E, Bieber A, et al. User-centered design of a multimodal training app for people with aphasia. 2020. Presented at: Annual Convention of the American Speech-Language-Hearing Association (ASHA); Nov 19-21, 2020; San Diego, CA. URL: https://voigt-antons.de/publication/2020-11-01-OC17 [Accessed 2025-08-11]
  33. Leonard C, Rochon E, Laird L. Treating naming impairments in aphasia: findings from a phonological components analysis treatment. Aphasiology. Sep 2008;22(9):923-947. [CrossRef]
  34. Simic T, Leonard C, Laird L, Stewart S, Rochon E. The effects of intensity on a phonological treatment for anomia in post-stroke aphasia. J Commun Disord. 2021;93:106125. [CrossRef] [Medline]
  35. Boyle M, Coelho CA. Application of semantic feature analysis as a treatment for aphasic dysnomia. Am J Speech Lang Pathol. Nov 1995;4(4):94-98. [CrossRef]
  36. Efstratiadou EA, Papathanasiou I, Holland R, Archonti A, Hilari K. A Systematic review of semantic feature analysis therapy studies for aphasia. J Speech Lang Hear Res. May 17, 2018;61(5):1261-1278. [CrossRef] [Medline]
  37. Kertesz A, Raven JC. WAB-R: Western Aphasia Battery-Revised. PsychCorp; 2007.
  38. Goodglass H, Kaplan E, Weintraub S. Boston Naming Test. Lea & Febiger; 1983. URL: https://psycnet.apa.org/doiLanding?doi=10.1037%2Ft27208-000 [Accessed 2025-08-11]
  39. Kim ES, Laird L, Wilson C, et al. Implementation and effects of an information technology-based intervention to support speech and language therapy among stroke patients with aphasia: protocol for a virtual randomized controlled trial. JMIR Res Protoc. Jul 2, 2021;10(7):e30621. [CrossRef] [Medline]
  40. Dekhtyar M, Braun EJ, Billot A, Foo L, Kiran S. Videoconference administration of the western aphasia battery-revised: feasibility and validity. Am J Speech Lang Pathol. May 8, 2020;29(2):673-687. [CrossRef] [Medline]
  41. Varley R, Cowell PE, Dyson L, Inglis L, Roper A, Whiteside SP. Self-administered computer therapy for apraxia of speech: two-period randomized control trial with crossover. Stroke. Mar 2016;47(3):822-828. [CrossRef] [Medline]
  42. Hilari K, Byng S, Lamping DL, Smith SC. Stroke and Aphasia Quality of Life Scale-39 (SAQOL-39): evaluation of acceptability, reliability, and validity. Stroke. Aug 2003;34(8):1944-1950. [CrossRef] [Medline]
  43. Lomas J, Pickard L, Bester S, Elbard H, Finlayson A, Zoghaib C. The communicative effectiveness index: development and psychometric evaluation of a functional communication measure for adult aphasia. J Speech Hear Disord. Feb 1989;54(1):113-124. [CrossRef] [Medline]
  44. Brooke J. SUS: a “quick and dirty” usability scale. In: Jordan PW, Thomas B, Weerdmeester BA, McLelland IL, editors. Usability Evaluation In Industry. Taylor and Francis; 1996:189-194.
  45. Harris PA, Taylor R, Thielke R, Payne J, Gonzalez N, Conde JG. Research electronic data capture (REDCap)—a metadata-driven methodology and workflow process for providing translational research informatics support. J Biomed Inform. Apr 2009;42(2):377-381. [CrossRef] [Medline]
  46. Harris PA, Taylor R, Minor BL, et al. The REDCap consortium: building an international community of software platform partners. J Biomed Inform. Jul 2019;95:103208. [CrossRef] [Medline]
  47. Gelman A, Tuerlinckx F. Type S error rates for classical and Bayesian single and multiple comparison procedures. Comput Stat. Sep 2000;15(3):373-390. [CrossRef]
  48. Quique YM, Evans WS, Dickey MW. Acquisition and generalization responses in aphasia naming treatment: a meta-analysis of semantic feature analysis outcomes. Am J Speech Lang Pathol. Mar 11, 2019;28(1S):230-246. [CrossRef]
  49. Sze WP, Hameau S, Warren J, Best W. Identifying the components of a successful spoken naming therapy: a meta-analysis of word-finding interventions for adults with aphasia. Aphasiology. Jan 2, 2021;35(1):33-72. [CrossRef]
  50. Doedens WJ, Meteyard L. Measures of functional, real-world communication for aphasia: a critical review. Aphasiology. Apr 2, 2020;34(4):492-514. [CrossRef]
  51. Kurfess C, Corsten S, Nickel MT, Knieriemen M, Kreiter D, Lauer N. Peer-to-peer support: digital networking in aphasia to improve quality of life (PeerPAL). Front Commun. Sep 7, 2023;8:1187233. [CrossRef]
  52. Devane N, Behn N, Marshall J, Ramachandran A, Wilson S, Hilari K. The use of virtual reality in the rehabilitation of aphasia: a systematic review. Disabil Rehabil. Nov 6, 2023;45(23):3803-3822. [CrossRef]
  53. Wallace SJ, Worrall L, Rose T, et al. A core outcome set for aphasia treatment research: the ROMA consensus statement. Int J Stroke. Feb 2019;14(2):180-185. [CrossRef] [Medline]
  54. Macoir J, Lavoie M, Routhier S, Bier N. Key factors for the success of self-administered treatments of poststroke aphasia using technologies. Telemed J E Health. Aug 2019;25(8):663-670. [CrossRef] [Medline]
  55. Greinacher R, Kojic T, Meier L, Parameshappa RG, Moller S, Voigt-Antons JN. Impact of tactile and visual feedback on breathing rhythm and user experience in VR exergaming. 2020. Presented at: 2020 Twelfth International Conference on Quality of Multimedia Experience (QoMEX):635-640; Athlone, Ireland. [CrossRef]
  56. Best B, Campbell J, Roxbury T, Worthy P, Copland DA. Exploring the usability and feasibility of a mobile music listening application for people living in the community with post-stroke aphasia. Disabil Rehabil. Jan 2024;46(2):344-353. [CrossRef] [Medline]
  57. Simic T, Leonard C, Laird L, Cupit J, Höbler F, Rochon E. A usability study of internet-based therapy for naming deficits in aphasia. Am J Speech Lang Pathol. Nov 1, 2016;25(4):642-653. [CrossRef] [Medline]


ASR: automatic speech recognition
BNT: Boston Naming Test
CETI: Communication Effectiveness Index
PCA: Phonological Components Analysis
RCT: randomized controlled trial
REDCap: Research Electronic Data Capture
SAQoL-39g: Stroke and Aphasia Quality of Life Scale - 39
SFA: Semantic Features Analysis
SLP: speech-language pathologist
SSNT: Study Stimulus Naming Test
SUS: System Usability Scale
UCD: user-centered design
WAB-R: Western Aphasia Battery-Revised


Edited by Lorraine Buis; submitted 18.10.24; peer-reviewed by John Pierce, Lili Zhang; final revised version received 16.06.25; accepted 22.06.25; published 18.08.25.

Copyright

© Esther S Kim, Laura Laird, Carlee Wilson, Steven Stewart, Philip Mildner, Sebastian Moller, Raimund Schatz, Robert P Spang, Jan-Niklas Voigt-Antons, Elizabeth Rochon. Originally published in JMIR mHealth and uHealth (https://mhealth.jmir.org), 18.8.2025.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mHealth and uHealth, is properly cited. The complete bibliographic information, a link to the original publication on https://mhealth.jmir.org/, as well as this copyright and license information must be included.