Published on in Vol 7, No 8 (2019): August

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/14250, first published .
Assessing Mobile Phone Digital Literacy and Engagement in User-Centered Design in a Diverse, Safety-Net Population: Mixed Methods Study

Assessing Mobile Phone Digital Literacy and Engagement in User-Centered Design in a Diverse, Safety-Net Population: Mixed Methods Study

Assessing Mobile Phone Digital Literacy and Engagement in User-Centered Design in a Diverse, Safety-Net Population: Mixed Methods Study

Original Paper

1Division of General Internal Medicine, Department of Medicine, University of California, San Francisco, San Francisco, CA, United States

2Department of Psychiatry, Zuckerberg San Francisco General Hospital, University of California, San Francisco, San Francisco, CA, United States

3Center for Vulnerable Populations, Zuckerberg San Francisco General Hospital, University of California, San Francisco, San Francisco, CA, United States

4School of Social Welfare, University of California, Berkeley, Berkeley, CA, United States

Corresponding Author:

Sarah S Nouri, MD, MPH

Division of General Internal Medicine

Department of Medicine

University of California, San Francisco

1545 Divisadero Street

Box 0320

San Francisco, CA, 94143

United States

Phone: 1 2149126555

Email: sarah.nouri@ucsf.edu


Background: Health care systems are rapidly deploying digital tools for disease management; however, few studies have evaluated their usability by vulnerable populations. To understand the barriers to app usage among vulnerable populations, we employed user-centered design (UCD) methods in the development of a new text messaging app.

Objective: The study aimed to describe variations in patients’ engagement in the app design process, focusing on limited health literacy (LHL), limited English proficiency (LEP), and limited digital literacy (LDL).

Methods: We conducted 20 in-depth semistructured interviews with primary care patients at a public health care system, used open-ended discussions and card sorting tasks to seek input about mobile phones and text messaging, and used open coding to categorize the patterns of mobile phone usage and to evaluate engagement in the card sorting process. We examined qualitative differences in engagement by examining the extensiveness of participant feedback on existing and novel text messaging content and calculated the proportion of patients providing extensive feedback on existing and novel content, overall and by health literacy, English proficiency, and digital literacy.

Results: The average age of the 20 participants was 59 (SD 8) years; 13 (65%) were female, 18 (90%) were nonwhite, 16 (80%) had LHL, and 13 (65%) had LEP. All had depression, and 14 (70%) had diabetes. Most participants had smartphones (18/20, 90%) and regularly used text messaging (15/20, 75%), but 14 (70%) of them reported having difficulty texting because of inability to type, physical disability, and low literacy. We identified 10 participants as specifically having LDL; 7 of these participants had LEP, and all 10 had LHL. Half of the participants required a modification of the card sorting activity owing to not understanding it or not being able to read the cards in the allotted time. The proportion of participants who gave extensive feedback on existing content was lower in participants with limited versus adequate English proficiency (4/13, 30% vs 5/7, 71%), limited versus adequate health literacy (7/16, 44% vs 3/4, 75%), and limited versus adequate digital literacy (4/10, 40% vs 6/10, 60%); none of these differences were statistically significant. When examining the proportion of patients who gave extensive feedback for novel messaging content, those with LHL were less engaged than those with adequate health literacy (8/16, 50% vs 4/4, 100%); there were no statistical differences by any subgroup.

Conclusions: Despite widespread mobile phone use, digital literacy barriers are common among vulnerable populations. Engagement in the card sorting activity varied among participants and appeared to be lower among those with LHL, LEP, and LDL. Researchers employing traditional UCD methods should routinely measure these communication domains among their end-user samples. Future work is needed to replicate our findings in larger samples, but augmentation of card sorting with direct observation and audiovisual cues may be more productive in eliciting feedback for those with communication barriers.

JMIR Mhealth Uhealth 2019;7(8):e14250

doi:10.2196/14250

Keywords



Background

Despite the potential promise of health-related mobile apps in improving the health of individuals with chronic diseases [1-4], few studies evaluate their usability and feasibility, and even fewer do so among vulnerable populations [5-7]. Existing data suggest that populations with limited health literacy (LHL), limited digital literacy (LDL), and limited English proficiency (LEP) are less likely to use health information technology (HIT), including mobile apps [6,8-11]. Given that smartphone ownership rates are similar in low income, minority populations compared with the general population of the United States, there are likely other barriers contributing to this decreased use of health apps [12]. To better understand these barriers and to reduce the disparities in the use of HIT for the management of chronic diseases, it is critical to include the vulnerable populations in the process of app development [13,14].

Even more specifically, there is a need to advance the field in terms of appropriate design science methods for working with vulnerable populations, including racial or ethnic minorities and those with low socioeconomic status and educational attainment [7]. There is similarly a need to understand the use and effectiveness of these methods in patients with mental health disorders that are associated with lower activation and engagement, such as depression [15]. Standard user-centered design (UCD) methods, such as prototyping and card sorting, often use rapid-fire sessions with end users to generate and prioritize a large amount of potential content for a digital health program or intervention [16-18]. Although the goal of these methods is to understand the experiences, beliefs, and preferences of end users, they could represent a cognitively demanding approach for the participants—especially in terms of hypothetical discussions about future health behaviors and sifting through large volumes of potential content.

Objectives

We sought to understand which design methods worked well within a larger study employing UCD methods to develop a text messaging app aimed at increasing physical activity among patients with comorbid diabetes and depression. In this paper, we evaluate data from 2 sets of semistructured interviews conducted during the early phases of the app development and describe the patterns of mobile phone usage and variations in engagement with design methods by health literacy, English proficiency, and digital literacy of patients recruited from a public health care system.


Research Setting and Sampling Procedure

Data for this study comprise 20 transcripts of semistructured interviews collected during 2016 to 2018 over 2 phases. All patients were recruited from primary care clinics from the public health care system for the city and county of San Francisco. A total of 10 transcripts are from patients who had completed a previous trial evaluating automated text messages as an adjunct to cognitive behavioral therapy for depression (MoodText trial, NCT01083628) [19], and 10 are from a separate group of patients participating in early design sessions to develop a text messaging app to increase physical activity in patients with diabetes and depression (Diabetes and Depression Text Messaging Intervention [DIAMANTE] trial, NCT03490253). Inclusion criteria were as follows: age ≥18 years, English- or Spanish-speaking, ownership of any type of mobile phone, and a diagnosis of depression. Exclusion criteria were active suicidal ideation with a plan and active severe psychosis.

Data Collection Procedure

Short questionnaires were administered to the participants during recruitment to assess sociodemographic factors (including age, gender, race and ethnicity, education level, income, employment status, and English proficiency), health status, and health literacy. Semistructured, in-depth interviews were conducted with all 20 participants in either English or Spanish. All interviews were 1.5 hours in duration, and the interviews in Spanish were conducted by study staff who were native speakers. Interview guides for all participants included questions about mobile phone usage, physical activity, and feedback on sample text messages and text messaging interventions (interview guides and card sorting instructions are available by request). Participants also completed a closed card sorting activity. Card sorting is a method used to explore how people group concepts, and this has previously been used in the development of text messaging interventions [20]. Participants were given a set of note cards with sample text messages written at a sixth-grade reading level in either English or Spanish, depending on their preferred language, and were asked to sort the cards into 3 piles depending on whether they liked, disliked, or felt neutral about the messages on the cards. They were then asked to explain why they liked or disliked these messages. If they did not provide a reason for liking or disliking the messages, they were probed once more by the interviewer. If the participants did not understand the card sorting activity or had difficulty reading the cards, card sorting was modified such that the interviewer read cards aloud and elicited feedback on each sample text message one at a time. On completion of the card sorting activity, participants were asked if they had any additional ideas for potential text message content or structure of the text messaging intervention.

Qualitative Analysis

The qualitative analysis was done with open coding of all transcripts using inductive and deductive coding techniques (Dedoose version 8.1.8). One author (SN) created the initial codebook based on the interview questions. A second author (PA) read a subset of transcripts and coded using the original codebook. SN, PA, and CL met frequently to review the codebook, discuss the emerging themes, and resolve disagreements. For this study, we specifically coded for (1) patterns of mobile phone usage and (2) engagement in the design methods. This involved categorizing both the thematic ideas that emerged from the interviews, as well as the extensiveness of participation in (1) the card sorting activity to provide feedback on the existing text messaging content and (2) the semistructured interview to provide novel suggestions for new text messaging content. Feedback on existing content was further categorized as extensive if it included not only what the participants liked or disliked but also why they liked it and how they felt it would help them. If the feedback on the existing content consisted only of what the participants liked or disliked, it was determined to be minimal. Novel suggestions were extensive if they provided detailed content for possible text messages and also offered new types of messages. If novel suggestions were repetitions of text messages already developed by the study and shared with the participant, or if they were unrelated to the purpose or content of the study, they were determined to be minimal or misaligned.

Exploratory Analyses

For our secondary analyses, we were interested in the differences in key themes by participant demographics, specifically, health literacy, English proficiency, and digital literacy. Health literacy was measured using the single-item assessment by Chew et al, “How confident are you filling out medical forms by yourself?” As previously validated in a population similar to that of this study, adequate health literacy was defined by answering “extremely,” whereas LHL was defined by answers “not at all,” “a little bit,” “somewhat,” or “quite a bit” [21-23]. English proficiency was also measured using a validated single-item assessment, collapsing Spanish-speaking participants with those that reported “moderate,” “little,” and “very little ability” with speaking English [24]. Owing to the lack of a validated survey measure of digital literacy that is widely used for mobile phone usage, we created a definition of digital literacy based on the interview responses about the current use of mobile phones. We categorized participants with very limited use or engagement and/or self-reported difficulty in using their mobile phones as having LDL. We based our definition in Kayser et al’s concept of eHealth literacy [25], incorporating both capabilities and experience/engagement in using technologies. We did not include access to technology in our definition as all participants owned a mobile phone. We then quantitatively compared the prevalence of LDL by participants’ health literacy and English proficiency using Fisher exact tests to assess for differences by these 2 independent variables.

Finally, in an exploratory way, we calculated the sum of the number of unique feedback statements or novel suggestions per participant given during the card sorting task. We calculated the means and ranges of the total number of statements given and then preliminary Wilcoxon rank sum tests were conducted to evaluate the differences in these frequencies by health literacy, English proficiency, and digital literacy. We also calculated the number of participants who provided extensive versus minimal feedback statements and novel suggestions and used Fisher exact tests to evaluate the differences in these proportions by health literacy, English proficiency, and digital literacy. Analyses were conducted using Stata/SE 15.0.


Participant Characteristics

The average age of the participants was 59 (SD 8) years; 65% (13/20) were female, 90% (18/20) were nonwhite, 65% (13/20) had LEP (50% (10/20) Spanish-speaking), 45% (9/20) had a high school education or less, and 80% (16/20) had LHL (Table 1). None of the participants were employed full time; 45% (9/20) were disabled, and 20% (4/20) were unemployed. A total of 70% (14/20) of the participants had type 2 diabetes mellitus, and 65% (13/20) self-rated their health status as fair or poor. The proportion of participants with LEP (80% vs 50%) and with diabetes (90% vs 50%) was higher in the DIAMANTE trial than the proportion of participants in the MoodText trial; otherwise there were no differences in the demographic or health data by study.

Table 1. Participant characteristics (N=20).
CharacteristicsValues
Age (years), mean (SD)59.0 (7.7)
Gender, n (%)

Men7 (35)

Women13 (65)
Income (USD), n (%)

<$20,00011 (55)

>$20,0002 (10)

Other/refused7 (35)
Employment status, n (%)

Part time4 (20)

Unemployed4 (20)

Disabled9 (45)

Retired3 (15)
Race or ethnicity, n (%)

White2 (10)

Hispanic or Latino11 (55)

Black or African American4 (20)

Asian or Pacific Islander2 (10)

Other1 (5)
Limited English proficiency, n (%)13 (65)
Limited health literacy, n (%)16 (80)
Education, n (%)

None or primary school7 (35)

High school graduate or GEDa2 (10)

Some college or technical school9 (45)

College graduate or graduate degree2 (10)
Diabetes, n (%)14 (70)
Depression, n (%)20 (100)
Health status, n (%)

Fair or poor13 (65)
 Good7 (35)

aGED: General Education Development.

Mobile Phone Use and Digital Literacy

We uncovered major categories of mobile phone usage related to texting networks, mobile phone carriage, and difficulty with the basic features of mobile phones (ie, text messaging). Overall, nearly all participants (18/20) had a smartphone and used text messaging regularly with a network of family and friends (15/20). A total of 5 participants described having limited texting networks, texting only family members or research studies. Most participants (17/20) reported carrying their mobile phones with them whenever leaving home.

Despite the overall high rate of mobile phone usage, many participants were defined as having LDL. Participants were categorized as having LDL if they had limited texting networks, frequently did not carry their mobile phones or turned them off for long periods of time (limited mobile phone carriage), or had difficulty using their phones or sending text messages owing specifically to unfamiliarity with or difficulty typing or using the microphone feature. Overall, 10 participants were categorized as having LDL; others were categorized as having adequate digital literacy. Those who described difficulty using their phones attributed this primarily to being new smartphone owners and still learning how to use their phones. Several also cited their phones being turned off or forgetting to check their phones, low literacy, and having to use the microphone feature on their smartphone as hindrances to text messaging:

Now that I just got this phone, and before I was able to send text messages but now I’ve forgotten. [...] I didn’t want a high-tech phone […] but my kids gifted me a phone and said, “you have to learn, mom.”
[translated from Spanish]
I didn’t check my phone or my phone was off for the day. Sometimes I’ll turn my phone off and I forget to turn it back on.
Yes, because I don’t use the keys and sometimes when you speak into the microphone, it doesn’t write the words that you want to say correctly. My daughter calls me out on that. So, if I’m going to send a message that’s not going to be correct like it should be, I’d feel bad.
[translated from Spanish]

Of the 10 participants with LDL, 7 had LEP and 10 had LHL. There was a similarly high proportion of participants with LEP and/or LHL within each of the mobile phone usage domains we used to define the LDL (Table 2). Specifically, the majority of participants with limited texting network, limited mobile phone carriage, and difficulty using their phone or texting had LEP and/or LHL. Fisher exact test results were not statistically significant.

The other most commonly mentioned barriers to mobile phone usage that were not related to digital literacy included being too busy and forgetting to respond, having a physical impairment (eg, arthritis) that made typing difficult, and not being in the mood to respond to text messages:

Um, I got real busy, I didn’t hear - I didn’t check my phone or my phone was off for the day. Sometimes I’ll turn my phone off and I forget to turn it back on and, um, then it’s late and then I’m sleepy so I won’t answer.
It has such small letters and I have a problem with my hands because I have Raynaud’s syndrome, where your hands fall asleep.
[translated from Spanish]
Sometimes, well, I don’t pay attention and sometimes I do. Not every day. Sometimes I’m in a bad mood [laughs], sorry for saying it.
[translated from Spanish]
Table 2. Mobile phone digital literacy, as well as domains used to measure digital literacy by English proficiency and health literacy.
Mobile phone digital literacyTotal number of patients, N (%)LEPa, n (%)LHLb, n (%)Both LEP and LHL, n (%)
Subdomains of mobile phone digital literacy

Limited texting network5 (25)4 (80)4 (80)4 (80)

Limited mobile phone carriage3 (15)3 (100)3 (100)3 (100)

Difficulty using phone and/or texting10 (50)7 (70)10 (100)7 (70)
Overall limited digital literacy10 (50)7 (70)10 (100)7 (70)

aLEP: limited English proficiency.

bLHL: limited health literacy.

Engagement in Card Sorting Activity

Engagement was examined in 2 ways. First, we determined the frequency of feedback given about the existing content and the frequency of novel suggestions provided. We also used our qualitative coding to identify the extensiveness of both feedback on the existing content and novel suggestions.

Feedback on Existing Text Message Content

Feedback on text message content was elicited by a card sorting activity—half of the participants required a modification of the activity as they either did not understand it or had difficulty reading the cards in the allotted time.

The most common positive feedback was for the text messaging content to encourage participants to reflect on their behaviors or thoughts, give concrete ideas or advice, or be highly positive and motivating. The most common negative feedback was for the content that was viewed as repetitive. Several exemplar quotes demonstrated these positive and negative feedback themes:

But most of the times when I received the messages, they came at a good time and it just helped me to evaluate, um, you know, my thinking and my feelings, which if it hadn’t came, then I wouldn’t have been thinking about, you know, different things
[theme: encourage reflection]
These are all concrete suggestions of things that you can try because they will probably improve your mood just because you try. […] So I think it’s just little suggestions or ways that you could help make yourself feel better. Less negative. Less down.
[theme: concrete advice]
I also liked the way they were positive.
[translated from Spanish; theme: positive messages]
They’re a little repetitive.
[translated from Spanish; theme: repetitive messages]

For each participant, the number of different reasons they gave for liking or disliking sample text messages was summed, and this defined the frequency of feedback.

We also evaluated the extensiveness of feedback provided by each participant. Half of the participants provided extensive feedback, as evidenced by these comments:

[They] just make you stop for a minute. I think that’s maybe the key to why it works so well because you know the message comes in and you’re like well why am I doing this? I’m not being very positive today and I could do something better. I could think of another way to handle this. Maybe I should try that.
Well, they spark the idea of getting started. They spark the idea of why - I mean, they have started to continue in giving me I guess motivational nudges here and there.
It says “every time you think you can’t or you won’t be able to [do something], remember the times you were able to.” Yes, there are times that these messages help me, this type of message helps me because often I say, “I can’t,” but occasionally I’ve read these messages and I say, “yes I can. “I’ll try it no matter what,” “I’ll try it,” and then I’ve done it and I have examples.
[translated from Spanish]

On the other hand, participants who gave feedback that was minimal, provided little, if any, varied feedback, and did not explain why they liked or disliked the text messages even after being probed by the interviewer:

Yeah. Not interested. Okay.
I liked all of them.
[translated from Spanish]

For quantitative analyses (Table 3), we calculated the frequency of feedback by coding each instance of feedback on the sample text messages. We found that the frequency of feedback was lower in participants with LEP, LHL, and LDL than in those with English proficiency, adequate health literacy, or adequate digital literacy, although none of these differences were statistically significant. It appears that the participants who were proficient in English provided more extensive feedback compared with those with LEP (5/7, 71% vs 4/13, 30%), as did those with adequate versus LHL (3/4, 75% vs 7/16, 44%) and those with adequate versus LDL (6/10, 60% vs 4/10, 40%). These differences approached but did not achieve statistical significance.

Table 3. Frequency and extensiveness of the feedback on existing text messaging content by English proficiency, health literacy, and digital literacy.
Patient characteristicFrequency indicated by median number of unique feedbackExtensiveness indicated by number who provided extensive feedback, n (%)
English proficiency


Adequate (n=7)5.55 (71)
 Limited (n=13)44 (30)
Health literacy


Adequate (n=4)5.53 (75)
 Limited (n=16)47 (44)
Digital literacy


Adequate (n=10)56 (60)
 Limited (n=10)44 (40)
Novel Suggestions

Participants were asked whether they had other ideas for text messages that were not presented during the card sorting activity, as well as whether they had any other general feedback on the text messages or the text messaging interventions. The majority of suggestions were for text message content:

Um, or, um, “I know you’re feeling down, so let’s go for a walk and maybe that will pick you up a little bit. I’ve tried that before and it happens, so let’s do it.”
Hmm. “Rise and shine!” Uh, let’s see, okay. “Rise and shine, get off your behind!”

A fewer number of participants provided feedback on broader, structural content of the text messaging interventions:

It could be a little cartoon running, an example of someone walking or running.
[translated from Spanish]
There are thousands of illnesses in the world. So, first you need to know which illnesses someone has, and what difficulties they may have in doing physical activity. There are many types of physical activity. Because after my surgery my exercise was to stretch, to lift my arms with a pound of rice. It’s moving the body. From here down I couldn’t move because of my knee. Stretching, nothing more. Of course. This way, you have to see what the patient has to motivate them.
[translated from Spanish]

We evaluated the extensiveness of novel suggestions as well. Extensive novel suggestions were defined as providing detailed content for possible text messages and also offering new types of messages:

And I would send them [a message] saying that life is very beautiful. Life is a thing of beauty and one should be happy and put aside the negative to give life to the positive.
[translated from Spanish]
That the gym is not the only place where one can exercise, because you have this mentality of “I’ll sign up but it’s very expensive.” If there are other options for exercise, that’s excellent.
[translated from Spanish]

Suggestions were categorized as minimal or misaligned, if they were vague or identical to the messages presented during card sorting or were unrelated to the physical activity content focus of the text messaging intervention (eg, text messages with the results of laboratory tests):

Probably reminding uh reminding me for doing me or exercise, walking, whatever.
I think that if you're going to send me one of these messages, it’s to warn me about something that happened or is about to happen. I mean to say […] let me give an example. The doctor saw […] my results and wanted to tell me how they went.
[translated from Spanish]

We coded each separate answer by participants as novel suggestions, and the frequency of novel suggestions for each participant was defined as the sum of the coded excerpts. The median number of novel suggestions was 2 per participant (range 0-9; Table 4), and there were no significant differences by English proficiency, health literacy, or digital literacy. All participants with adequate health literacy provided extensive novel suggestions, compared with only half of those with LHL (Table 4). There were no differences in the extensiveness of novel suggestions by English proficiency or digital literacy.

Table 4. Frequency and extensiveness of novel suggestions by English proficiency, health literacy, and digital literacy.
Patient characteristicFrequency indicated by median number of suggestionsExtensiveness indicated by number that provided extensive suggestions, n (%)
English proficiency


Adequate (n=7)2.54 (57)

Limited (n=13)26 (56)
Health literacy


Adequate (n=4)24 (100)

Limited (n=16)28 (50)
Digital literacy


Adequate (n=10)25 (50)

Limited (n=10)25 (50)

Principal Findings

Our study found that the participants representing vulnerable populations engaged differently in the UCD process that we employed and described difficulty with mobile phone usage and text messaging. Although nearly all participants had a smartphone that they carried with them throughout the day and a large majority had wide texting networks, over three-quarters still described difficulty with text messaging. This discrepancy suggests that smartphone ownership, even with daily personal use, does not accurately predict comfort or ability in using the basic features of phones such as text messaging. This finding is supported by emerging data demonstrating limited use (eg, voice communication only) of mobile phones by vulnerable populations [26], as well as by efforts in mobile interaction design to focus on low literacy or illiterate populations [27].

We found variations in engagement in our design process by health literacy, English proficiency, and digital literacy. Digital literacy, defined broadly by the US Department of Education as “digital problem solving” and measured by assessments of basic computer competence, affects 16% of the US population; those who have LDL tend to be older, nonwhite, foreign born, and less educated than those who are digitally literate [28]. This definition, however, does not incorporate the use of other ubiquitous technologies, including smartphones, and has not been evaluated in the context of health and health care. Therefore, we used a de novo classification from the participants’ responses to calculate our own definition of mobile phone digital literacy, informed by previous conceptual models in this space. Most individuals we identified as having LDL also had LHL and LEP, suggesting some overlap in these constructs. User characteristics such as age, education, and employment are considered in frameworks for developing information technology systems, and it is likely that health literacy and English proficiency are similarly important user characteristics in health-related digital literacy [25,29]. In fact, health literacy has been included as a component of digital literacy in some conceptual models [30].

Despite following best practice guidelines for card sorting, such as limiting the number of cards and categories and providing uniform, clear directions on how to conduct the task [17,31,32], this method had to be modified in half of the interviews and was therefore not fully effective in this study population. Eliciting novel suggestions was also difficult, with most participants only providing 1 or 2 suggestions. The latter is consistent with studies of engagement in qualitative methods that suggest that end users are often unable to provide anticipatory feedback [16]. However, we expected more consistency in the feedback on existing materials, as end users are usually more engaged in this type of feedback. This may be because in obtaining feedback via a card sorting method, we created the additional challenge of having to provide feedback on hypothetical messages within a set time limit, which has been shown to be a more challenging task [16]. It is also possible that the participant’s response bias owing to the interviewer’s demand characteristics contributed to the differences in feedback by health and digital literacy and English proficiency, though we tried to minimize this by conducting interviews in Spanish with native Spanish-speaking interviewers [33]. Future work is needed to replicate our findings in larger samples; however, our findings suggest that augmenting card sorting with different methods in UCD, such as task analysis and usability testing with direct observation, may be more effective in obtaining feedback from vulnerable populations [18,25,29]. Furthermore, observation in environments in which end users are both comfortable and will be using the digital health intervention may be more productive in eliciting feedback and matching the intervention with the end users’ needs and preferences.

In the development of digital health interventions to improve outcomes in patients with diabetes, others have also employed user-centered or participatory design and have obtained and successfully incorporated feedback in their development processes [34,35]. Although these studies have enrolled diverse populations, they do not describe whether the feedback they received differed within subgroups of their populations. Our exploratory analyses showed differences in engagement in specific subgroups of patients, suggesting that there may be other factors contributing to these differences. LHL, specifically, has been associated with lower patient activation and engagement in clinical settings, and this may be relevant in research settings as well [36-38]. Our findings underscore the importance of measuring health literacy, English proficiency, and digital literacy and considering their effects in UCD. This is critical as simply recruiting these vulnerable populations to user-centered or participatory design research rather than engaging them in feedback processes in a meaningful way carries the risk of exacerbating existing disparities in the use of HIT [8].

Limitations

There are several limitations to this study. In defining engagement, we assessed the extensiveness of feedback statements and novel suggestions; however, the provision of minimal feedback may have been owing to disinterest in the content being discussed rather than purely a lack of engagement. Given our small sample sizes that are typical in design research [31,32], many comparisons did not reach statistical significance; nevertheless, our quantitative data did allow us to highlight these differences. In addition, we focused specifically on mobile phone–related digital literacy rather than a more comprehensive digital literacy assessment, given the focus of our future text messaging intervention. Furthermore, 10 of the participants had already participated in a trial evaluating a text messaging app and therefore had more familiarity with this type of intervention. Nevertheless, neither successful completion of the card sorting task nor the engagement differed by study. Finally, although we chose to look specifically at health literacy, English proficiency, and digital literacy as the predictors of engagement, there are almost certainly other factors contributing to the differences in engagement noted in these populations, including education, race, culture, patient activation, and the nature and homogeneous content of the interviews [37,39]. Notably, all the study participants had depression, which has been associated with lower patient activation [40]; however, despite this, we were still able to detect differences in engagement within subgroups of the population.

Conclusions

Engagement in our design process varied by health literacy, English proficiency, and digital literacy. The participants of our study represent a diverse population—in race, employment, education, literacy and language, and general health status—that is rarely captured in usability studies for HIT. We believe this to be a major strength of this paper, both in describing the variations in patterns of mobile phone usage and in evaluating the engagement in UCD methods. Our findings highlight the need for a better understanding of how to consider, define, and incorporate digital literacy when developing HIT, as well as continued efforts in better engaging vulnerable populations in research. Our future work will report on the process for incorporating format and content feedback into our final intervention.

Acknowledgments

This project was funded by the Agency for Healthcare Research and Quality (1R01HS025429-01), a career development award by the National Institute of Mental Health (5K23MH0944442), a career development award by the National Cancer Institute (K24CA212294), and a research training grant by the National Research Service Award (T32HP19025). The sponsors did not play a role in the design, methods, subject recruitment, data collection, analysis, and preparation of manuscript.

Conflicts of Interest

None declared.

  1. Rathbone AL, Prescott J. The use of mobile apps and SMS messaging as physical and mental health interventions: systematic review. J Med Internet Res 2017 Aug 24;19(8):e295 [FREE Full text] [CrossRef] [Medline]
  2. Anglada-Martinez H, Riu-Viladoms G, Martin-Conde M, Rovira-Illamola M, Sotoca-Momblona JM, Codina-Jane C. Does mhealth increase adherence to medication? Results of a systematic review. Int J Clin Pract 2015 Jan;69(1):9-32. [CrossRef] [Medline]
  3. Whitehead L, Seaton P. The effectiveness of self-management mobile phone and tablet apps in long-term condition management: a systematic review. J Med Internet Res 2016 May 16;18(5):e97 [FREE Full text] [CrossRef] [Medline]
  4. Wu Y, Yao X, Vespasiani G, Nicolucci A, Dong Y, Kwong J, et al. Mobile app-based interventions to support diabetes self-management: a systematic review of randomized controlled trials to identify functions associated with glycemic efficacy. JMIR Mhealth Uhealth 2017 Mar 14;5(3):e35 [FREE Full text] [CrossRef] [Medline]
  5. Sarkar U, Gourley GI, Lyles CR, Tieu L, Clarity C, Newmark L, et al. Usability of commercially available mobile applications for diverse patients. J Gen Intern Med 2016 Dec;31(12):1417-1426 [FREE Full text] [CrossRef] [Medline]
  6. O'Connor S, Hanlon P, O'Donnell CA, Garcia S, Glanville J, Mair FS. Understanding factors affecting patient and public engagement and recruitment to digital health interventions: a systematic review of qualitative studies. BMC Med Inform Decis Mak 2016 Sep 15;16(1):120 [FREE Full text] [CrossRef] [Medline]
  7. Stowell E, Lyson MC, Saksono H, Wurth RC, Jimison H, Pavel M, et al. Designing and Evaluating mHealth Interventions for Vulnerable Populations: A Systematic Review. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems. 2018 Presented at: CHI'18; April 21-26, 2018; Montreal, QC, Canada p. 1-17. [CrossRef]
  8. Mackert M, Mabry-Flynn A, Champlin S, Donovan EE, Pounders K. Health literacy and health information technology adoption: the potential for a new digital divide. J Med Internet Res 2016 Oct 4;18(10):e264 [FREE Full text] [CrossRef] [Medline]
  9. Chakkalakal RJ, Kripalani S, Schlundt DG, Elasy TA, Osborn CY. Disparities in using technology to access health information: race versus health literacy. Diabetes Care 2014;37(3):e53-e54 [FREE Full text] [CrossRef] [Medline]
  10. Broderick J, Devine T, Langhans E, Lemerise AJ, Lier S, Harris L. National Academy of Medicine. 2014. Designing Health Literate Mobile Apps   URL: https://nam.edu/wp-content/uploads/2015/06/HealthLiterateApps.pdf [accessed 2019-03-29] [WebCite Cache]
  11. Sarkar U, Karter AJ, Liu JY, Adler NE, Nguyen R, Lopez A, et al. The literacy divide: health literacy and the use of an internet-based patient portal in an integrated health system-results from the diabetes study of northern California (DISTANCE). J Health Commun 2010;15(Suppl 2):183-196 [FREE Full text] [CrossRef] [Medline]
  12. Pew Research Center. 2019. Mobile Fact Sheet   URL: https://www.pewinternet.org/fact-sheet/mobile/ [accessed 2018-10-17] [WebCite Cache]
  13. Aguilera A, Lyles CR. The case for jointly targeting diabetes and depression among vulnerable patients using digital technology. JMIR Diabetes 2017 Jan 17;2(1):e1 [FREE Full text] [CrossRef] [Medline]
  14. Ramirez V, Johnson E, Gonzalez C, Ramirez V, Rubino B, Rossetti G. Assessing the use of mobile health technology by patients: an observational study in primary care clinics. JMIR Mhealth Uhealth 2016 Apr 19;4(2):e41 [FREE Full text] [CrossRef] [Medline]
  15. Doherty G, Sharry J, Bang M, Alcañiz M, Baños R. Technology in Mental Health. In: CHI'08 Extended Abstracts on Human Factors in Computing Systems. 2008 Presented at: CHI EA'08; April 5-10, 2008; Florence, Italy p. 3965-3968. [CrossRef]
  16. Yardley L, Spring BJ, Riper H, Morrison LG, Crane DH, Curtis K, et al. Understanding and promoting effective engagement with digital behavior change interventions. Am J Prev Med 2016 Nov;51(5):833-842. [CrossRef] [Medline]
  17. Usability. What & Why of Usability: Card Sorting   URL: https://www.usability.gov/how-to-and-tools/methods/card-sorting.html [accessed 2019-03-29] [WebCite Cache]
  18. Dopp AR, Parisi KE, Munson SA, Lyon AR. A glossary of user-centered design strategies for implementation experts. Transl Behav Med 2018 Dec 7 (epub ahead of print). [CrossRef] [Medline]
  19. Aguilera A, Bruehlman-Senecal E, Demasi O, Avila P. Automated text messaging as an adjunct to cognitive behavioral therapy for depression: a clinical trial. J Med Internet Res 2017 May 8;19(5):e148 [FREE Full text] [CrossRef] [Medline]
  20. Farris KB, Salgado TM, Batra P, Piette JD, Singh S, Guhad A, et al. Confirming the theoretical structure of expert-developed text messages to improve adherence to anti-hypertensive medications. Res Social Adm Pharm 2016;12(4):578-591 [FREE Full text] [CrossRef] [Medline]
  21. Chew LD, Bradley KA, Boyko EJ. Brief questions to identify patients with inadequate health literacy. Fam Med 2004 Sep;36(8):588-594 [FREE Full text] [Medline]
  22. Powers BJ, Trinh JV, Bosworth HB. Can this patient read and understand written health information? J Am Med Assoc 2010 Jul 7;304(1):76-84. [CrossRef] [Medline]
  23. Sarkar U, Schillinger D, López A, Sudore R. Validation of self-reported health literacy questions among diverse English and Spanish-speaking populations. J Gen Intern Med 2011 Mar;26(3):265-271 [FREE Full text] [CrossRef] [Medline]
  24. Karliner LS, Napoles-Springer AM, Schillinger D, Bibbins-Domingo K, Pérez-Stable EJ. Identification of limited English proficient patients in clinical care. J Gen Intern Med 2008 Oct;23(10):1555-1560 [FREE Full text] [CrossRef] [Medline]
  25. Kayser L, Kushniruk A, Osborne RH, Norgaard O, Turner P. Enhancing the effectiveness of consumer-focused health information technology systems through ehealth literacy: a framework for understanding users' needs. JMIR Hum Factors 2015 May 20;2(1):e9 [FREE Full text] [CrossRef] [Medline]
  26. Thies IM. SIGCHI Social Impact Award Talk -- Designing for Low-Literate Users. In: Proceedings of the 2017 CHI Conference Extended Abstracts on Human Factors in Computing Systems. 2017 Presented at: CGI EA'17; May 6-11, 2017; Denver, Colorado, USA p. 8-9. [CrossRef]
  27. Belay EG, McCrickard DS, Besufekad SA. Designing Mobile Interaction for Low-Literacy (D-MILL). In: Proceedings of the First African Conference on Human Computer Interaction. 2016 Presented at: AfriCHI'16; November 21-25, 2016; Nairobi, Kenya p. 251-255. [CrossRef]
  28. Mamedova S, Pawlowski E. National Center for Education Statistics (NCES). 2018. A Description of US Adults Who Are Not Digitally Literate   URL: https://nces.ed.gov/pubs2018/2018161.pdf [accessed 2019-03-29] [WebCite Cache]
  29. Kushniruk A, Turner P. A framework for user involvement and context in the design and development of safe e-health systems. Stud Health Technol Inform 2012;180:353-357. [CrossRef] [Medline]
  30. Norman CD, Skinner HA. eHealth literacy: essential skills for consumer health in a networked world. J Med Internet Res 2006 Jun 16;8(2):e9 [FREE Full text] [CrossRef] [Medline]
  31. Paul CL. A modified Delphi approach to a new card sorting methodology. J Usability Stud 2008;4(1):7-30 [FREE Full text]
  32. Wood JR, Wood LE. Card sorting: current practices and beyond. J Usability Stud 2008;4(1):1-6 [FREE Full text]
  33. Dell N, Vaidyanathan V, Medhi I, Cutrell E, Thies W. Yours is Better: Participant Response Bias in HCI. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 2012 Presented at: CHI'12; May 5-10, 2012; Austin, TX, USA p. 1321-1330. [CrossRef]
  34. Mayberry LS, Berg CA, Harper KJ, Osborn CY. The design, usability, and feasibility of a family-focused diabetes self-care support mhealth intervention for diverse, low-income adults with type 2 diabetes. J Diabetes Res 2016;2016:7586385 [FREE Full text] [CrossRef] [Medline]
  35. Zachary WW, Michlig G, Kaplan A, Nguyen NT, Quinn CC, Surkan PJ. Participatory design of a social networking app to support type II diabetes self-management in low-income minority communities. Proc Int Symp Hum Factors Ergon Healthc 2017 Jun;6(1):37-43 [FREE Full text] [CrossRef] [Medline]
  36. Aboumatar HJ, Carson KA, Beach MC, Roter DL, Cooper LA. The impact of health literacy on desire for participation in healthcare, medical visit communication, and patient reported outcomes among patients with hypertension. J Gen Intern Med 2013 Nov;28(11):1469-1476 [FREE Full text] [CrossRef] [Medline]
  37. Martin LT, Schonlau M, Haas A, Derose KP, Rosenfeld L, Buka SL, et al. Patient activation and advocacy: which literacy skills matter most? J Health Commun 2011;16(Suppl 3):177-190 [FREE Full text] [CrossRef] [Medline]
  38. McCormack L, Thomas V, Lewis MA, Rudd R. Improving low health literacy and patient engagement: a social ecological approach. Patient Educ Couns 2017 Jan;100(1):8-13. [CrossRef] [Medline]
  39. Bledsoe KL, Hopson RK. Conducting ethical research and evaluation in underserved communities. In: Mertens DM, Ginsberg PE, editors. The Handbook of Social Research Ethics. Thousand Oaks, California: Sage Publications; 2013:391-406.
  40. Blakemore AH, Hann M, Howells K, Panagioti M, Sidaway M, Reeves D, et al. Patient activation in older people with long-term conditions and multimorbidity: correlates and change in a cohort study in the United Kingdom. BMC Health Serv Res 2016 Oct 18;16(1):582 [FREE Full text] [CrossRef] [Medline]


DIAMANTE: Diabetes and Depression Text Messaging Intervention
HIT: health information technology
LDL: limited digital literacy
LEP: limited English proficiency
LHL: limited health literacy
UCD: user-centered design


Edited by G Eysenbach; submitted 05.04.19; peer-reviewed by L Richardson, M Tremblay, S Chokshi; comments to author 01.06.19; revised version received 11.06.19; accepted 11.06.19; published 29.08.19

Copyright

©Sarah S Nouri, Patricia Avila-Garcia, Anupama Gunshekar Cemballi, Urmimala Sarkar, Adrian Aguilera, Courtney Rees Lyles. Originally published in JMIR Mhealth and Uhealth (http://mhealth.jmir.org), 29.08.2019.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mhealth and uhealth, is properly cited. The complete bibliographic information, a link to the original publication on http://mhealth.jmir.org/, as well as this copyright and license information must be included.