RefCheck Maintenance Notice

On Monday, December 3, 2018, from 16:00-18:00 EST, RefCheck will be undergoing maintenance. RefCheck is the process where, during copyediting, all references are extracted from the manuscript file, parsed, matched against various databases (eg, PubMed and CrossRef), and automatically corrected. For more information on RefCheck, please visit our Knowledge Base.

Who will be affected?

Advertisement

Citing this Article

Right click to copy or hit: ctrl+c (cmd+c on mac)

Published on 07.09.18 in Vol 6, No 9 (2018): September

Preprints (earlier versions) of this paper are available at http://preprints.jmir.org/preprint/11450, first published Jun 29, 2018.

This paper is in the following e-collection/theme issue:

    Original Paper

    A Mobile Health Intervention for HIV Prevention Among Racially and Ethnically Diverse Young Men: Usability Evaluation

    1School of Nursing, Columbia University, New York, NY, United States

    2Division of Adolescent Medicine, Ann & Robert H Lurie Children's Hospital of Chicago, Chicago, IL, United States

    3Birmingham AIDS Outreach, Birmingham, AL, United States

    4Department of Pediatrics, Feinberg School of Medicine, Northwestern University, Chicago, IL, United States

    Corresponding Author:

    Hwayoung Cho, PhD, RN

    School of Nursing

    Columbia University

    560 West 168th Street

    New York, NY, 10032

    United States

    Phone: 1 2123056201

    Email:


    ABSTRACT

    Background: Mobile health (mHealth) apps have the potential to be a useful mode of delivering HIV prevention information, particularly for young men (13-24 years) who account for 21% of new HIV diagnoses in the United States. We translated an existing evidence-based, face-to-face HIV prevention curriculum into a portable platform and developed a mobile Web app: MyPEEPS Mobile.

    Objective: The purpose of this study was to assess the usability of MyPEEPS Mobile from both expert and end user perspectives.

    Methods: We conducted a heuristic evaluation with five experts in informatics to identify violations of usability principles and end user usability testing with 20 young men aged 15 to 18 years in New York, NY, Birmingham, AL, and Chicago, IL to identify potential obstacles to their use of the app.

    Results: Mean scores of the overall severity of the identified heuristic violations rated by experts ranged from 0.4 and 2.6 (0=no usability problem to 4=usability catastrophe). Overall, our end users successfully completed the tasks associated with use case scenarios and provided comments/recommendations on improving usability of MyPEEPS Mobile. The mean of the overall Post-Study System Usability Questionnaire scores rated by the end users was 1.63 (SD 0.65), reflecting strong user acceptance of the app.

    Conclusions: The comments made by experts and end users will be used to refine MyPEEPS Mobile prior to a pilot study assessing the acceptability of the app across diverse sexual minority young men in their everyday lives.

    JMIR Mhealth Uhealth 2018;6(9):e11450

    doi:10.2196/11450

    KEYWORDS



    Introduction

    Background

    With the rapid proliferation of mobile phone ownership across the world, use of mobile technologies in health care has expanded [1]. More than 325,000 mobile health (mHealth) apps were available globally on Apple iTunes and Google Play in 2017, and the number of mHealth apps continues to increase [2]. The ubiquitous nature of mobile phones brings convenience to everyday lives and creates opportunities to deliver health interventions in a portable format with enhanced privacy, increasing accessibility to the health interventions particularly tailored for stigmatized and disenfranchised populations [3-5].

    Figure 1. Four YMSM (young men who have sex with men) avatars on the MyPEEPS Mobile app.
    View this figure

    mHealth technology specifically has the potential to be a useful delivery mode of health information because it allows for the dissemination of information quickly and broadly [6-8].

    For the success of health information technologies, usability must be considered from the start of system development [9], yet few mHealth apps have undergone rigorous usability evaluation prior to their dissemination [10]. Usability factors remain one of the major obstacles to adoption of mHealth technologies because mHealth apps produced with poor quality are difficult to use, or are misused, which can lead to unintended consequences [11-13]. Therefore, usability evaluations are necessary to identify usability violations, guide system modification, and enhance technology acceptance by end users [14]. To provide the most effective and thorough usability evaluation results, a combination of usability evaluation techniques, including both experts and intended end users, during the evaluations is recommended [15,16].

    Study Context: MyPEEPS Intervention

    In 2016, of the 39,782 people in the United States newly infected with HIV, 21% were youth ages 13 to 24 years and 81% of incident cases among these youth were diagnosed in young men who have sex with men (YMSM), disproportionately occurring in African-American/black and Latino/Hispanic men [17]. An original Male Youth Pursuing Education, Empowerment & Prevention around Sexuality (MyPEEPS) intervention is a theory-driven (ie, social cognitive theory) [18], manualized HIV prevention curriculum developed for racially and ethnically diverse YMSM to address the need for an evidence-based HIV prevention intervention for this population [19]. The group-based, in-person intervention was found to be efficacious on reducing sexual risk, specifically sexual risk while under the influence of alcohol or drugs, in a 12-week feasibility trial. Nonetheless, participant engagement proved challenging due to travel distance and logistics around scheduling a group-based intervention.

    The use of mobile apps has been a popular way, particularly for YMSM, to get health information, connect with gay friends, and seek sex partners [20]. With the great promise of mHealth technology, we translated the existing face-to-face intervention into a mobile platform using an iterative design process [21,22]. The mobile Web app, MyPEEPS Mobile, was implemented by software developers at Little Green Software. MyPEEPS Mobile is guided by four YMSM avatars (ie, Philip aka P, Artemio, Nico, and Tommy; Figure 1) who manage their sexual health against a backdrop of personal, family-based, and relational challenges, and deliver the HIV prevention information to the end users. MyPEEPS Mobile consists of 21 activities divided into four modules or “PEEPScapades.” The activities include didactic content, graphical reports, videos, and true/false and multiple-choice quizzes. A user is required to complete the activities in consecutive order. On completing each activity, the user receives a trophy as a reward to promote continued participation. The purpose of this study was to assess the usability of the mHealth intervention, MyPEEPS Mobile, from the perspectives of experts and end users.


    Methods

    Overview

    We conducted two types of rigorous usability evaluations of MyPEEPS Mobile. First, we conducted a heuristic evaluation with informatics experts to identify violations of usability principles. Next, end user usability testing was conducted with target users, young men who are attracted to other men, to identify obstacles to their use of MyPEEPS Mobile. The Institutional Review Board (IRB) of Columbia University Medical Center in New York, NY, served as the central IRB for this study and approved all study activities.

    Heuristic Evaluation

    Sample Selection

    Five informaticians were invited via email to participate in a heuristic evaluation of MyPEEPS Mobile. The sample size was chosen in accordance with Nielsen’s recommendation to include three to five heuristic evaluators, as no additional information is likely to be produced with a larger sample [23]. Qualifications of the experts included (1) at least a Master’s degree in the field of informatics and (2) training in human-computer interaction. These qualifications were essential since the quality of the heuristic evaluation is dependent on the skills and experience of the usability experts [24].

    Procedures

    Heuristic evaluators were given a description of the full functionality of MyPEEPS Mobile. Each heuristic evaluator completed each of the 21 activities within the app (Textbox 1) at least once.


    Textbox 1. Summary of the 21 activities on MyPEEPS Mobile.
    View this box

    Experts were instructed to think-aloud as they evaluated the app. The process was recorded using a TechSmith Morae Recorder [25], which enables the researcher to record and analyze the audio recording and screenshots captured during the heuristic evaluation. Following completion of the tasks, heuristic evaluators were asked to rate the severity of the violations using an online version of the Heuristic Evaluation Checklist developed by Bright et al [26], based on Nielsen’s 10 heuristics [27]. Each heuristic was evaluated by one or more items and the overall severity of the identified heuristic violations were rated into five categories: no problem (0), cosmetic problem only (1), minor problem (2), major problem (3), and usability catastrophe (4). The evaluators were also asked to provide additional comments regarding the user interface. After the surveys were completed, evaluators received US $150 as compensation for their time.

    Data Analysis

    All experts’ comments about usability problems on the evaluation form and from the Morae recordings were compiled and reviewed by two research team members. Discrepancies in coding the data according to the usability factors of Nielsen’s 10 heuristics were discussed until consensus was achieved. Mean severity scores were calculated for each heuristic principle.

    End User Usability Testing

    Sample Selection

    For end user usability testing, potential participants were recruited from local community organizations through the use of passive and active methods (ie, convenience sampling; flyers, posting on social media, and direct outreach at community-based organizations) in New York, NY; Birmingham, AL; and Chicago, IL. Eligibility criteria were (1) between 13 and 18 years of age, (2) self-identified as male, (3) male sex assigned at birth, (4) understand and read English, (5) living within the metropolitan area of one of the three cities, (6) ownership of a mobile phone, (7) sexual interest in men and having either kissed another man or plans on having sex with a man in the next year, and (8) self-reported HIV-negative or unknown status. A sample of 20 participants was anticipated to be sufficient because prior research suggests an increasing benefit with samples up to 20 in usability testing (ie, the minimum percentage of problems identified rose from 82% up to 95% when the number of users was increased from 10 to 20) [28].

    Procedures

    All participants were given a brief explanation of MyPEEPS Mobile. The first 10 participants were provided with use case scenario version 1 (Textbox 2); once data saturation was achieved, the remaining 10 participants were provided with use case scenario version 2 (Textbox 3). Participants were asked to complete tasks using MyPEEPS Mobile on an iOS simulator for Windows computers. While participants were doing the tasks, the computer screen was video recorded using iMotions software (iMotions Biometric Research Platform 6.0, iMotions A/S, Copenhagen), which enables researchers to present images or screen/scene recordings and synchronize data from a variety of hardware platforms, if needed (eg, eye-tracking data), simultaneously. After the participants completed the tasks, they were then asked to watch a recording of their task performance on the computer screen. Participants were encouraged to retrospectively think-aloud and asked to verbalize their thoughts about the tasks they completed while watching a replay of the screen recordings. The process, including participants’ verbal comments, was audio recorded using Morae [25]. As part of the usability assessment, participants were asked to rate the app’s usability using the third version of the Post-Study System Usability Questionnaire (PSSUQ) [29] administered via Qualtrics (Provo, UT, USA) following the testing of MyPEEPS Mobile. The third version of the PSSUQ is a 16-item survey instrument to assess system usability on a scale ranging from 1 (strongly agree) to 7 (strongly disagree) including a neutral midpoint. A lower score on the PSSUQ indicates higher perceived usability of the app. The study visit took approximately 2 hours and participants were compensated US $40 to US $50 for their time. Interested individuals were consented for participation with a waiver of parental permission for minors.


    Textbox 2. Use case scenario version 1 (N=10).
    View this box

    Textbox 3. Use case scenario version 2 (N=10).
    View this box
    Data Analysis

    Data analysis was based on the audio/video recordings collected by Morae [25] and iMotions software. Participants’ verbalizations from the audio recordings were transcribed verbatim. Notes of critical incidents, characterized by comments, silence, repetitive actions, and error messages, were compiled from the recordings. Content analysis, a technique for making replicative and valid inferences from data, was performed by two research team members by reviewing the transcripts and critical incidents to identify common usability concerns. A third reviewer consulted in instances of uncertainty or discrepancy in the content analysis. Results from the PSSUQ were analyzed using Stata SE 14 (StataCorp LP, College Station, TX, USA) to calculate the descriptive statistics to complement the findings from the usability assessment.


    Results

    Heuristic Evaluation

    The mean age of the heuristic evaluators was 46.2 (SD 8.9) years, and mean years of experience in informatics that they had was 13.0 (SD 4.5) years. All the heuristic evaluators were female, 60% (n=3) were Asian, and 40% (n=2) were white. Mean scores and sample comments from the heuristic evaluation were organized into Nielsen’s 10 usability heuristics (Table 1) [27,30]. The mean scores of the overall severity of the identified heuristic violations ranged from 0.4 and 2.6, in which scores closest to 0 indicate a more usable app.

    The heuristic principle identified as the most in need of refinement was “user control and freedom” (mean 2.60, SD 1.14). Experts pointed out that MyPEEPS Mobile did not allow users the ability to move forward and backward: the “Back to Map” button only appears at the beginning of each activity. The second heuristic most identified for improvement was visibility of system status (mean 2.20, SD 0.45). A total of 21 activities divided into four PEEPScapades (Textbox 1) were displayed along a virtual “map” within MyPEEPS Mobile (Figure 2). The heuristic evaluators indicated that it was unclear which PEEPscapade they were in on the map, and the app should keep users informed about what was going on. Moreover, heuristic evaluators identified that the white navigation arrows used to advance through the “Testing with Tommy” activity and related comics (eg, illustrating what symptoms to look for when it comes to getting tested and treated for sexually transmitted diseases) were not clearly visible (Figure 3).

    In response to the usability factor “help and documentation” (mean 1.60, SD 0.89), one expert pointed to the lack of an instruction manual on how to navigate the app as a major concern. To improve “match between system and the real world” (mean 1.40, SD 0.89), experts recommended that an individual’s five important identity traits in the “My Bulls-I” activity (ie, move most important at the top followed by second through fifth; Figure 4) and response options for risk level in the “Well Hung” activity (ie, move no risk to far left side, followed by greater levels of risk to the right; Figure 5) be listed in a natural/logical order.

    End User Usability Testing

    The mean age of the end users was 17.4 (SD 0.88, range 15-18) years. Demographic characteristics including race, ethnicity, current student status, and education level are reported in Table 2. Sexual orientation was characterized using a gradient scale ranging from exclusively gay/homosexual to exclusively heterosexual, to capture the fluidity of sexuality at the time of the survey. Descriptive statistics on technology use, including type of mobile phone, social media sites, and apps are reported in Table 3. The majority of participants (85%, n=17) reported almost constant internet use (more than several times a day). The same percentage (85%, n=17) of participants reported using mobile devices (eg, mobile phone, tablet, and cell phone) as opposed to using laptop/desktop (15%, n=3) to access the internet in the past month. The mean duration of participants’ use of mobile apps on a mobile phone per day was 9.40 (SD 5.52) hours.

    Table 1. Mean severity scoresa and sample comments from the heuristic evaluation.
    View this table
    Figure 2. Map on MyPEEPS Mobile.
    View this figure
    Figure 3. Comics with unclear navigation.
    View this figure

    Positive Comments

    Overall, our end users successfully completed the tasks associated with the use case scenarios (Textboxes 2 and 3) and provided positive comments regarding the use of MyPEEPS Mobile. For example, participants liked the design and layout of MyPEEPS Mobile. One participant stated:

    The basic structure is that there’s pretty much an outline and you have four boys/men who are kind of the characters that kind of take you along this journey to HIV prevention and sexual health for MSM. This is sort of like fun. I like cartoons, videos, and quizzes.

    Also, participants liked the ease of the overall app use. One participant stated:

    I think it was pretty easy once I got the hang of moving to the side. I would just click the number and then I would start the quiz. It was simple. It was pretty quick.

    Recommendations

    General App Use

    Participants provided recommendations to improve usability by expressing their frustrations in general use of the app. For example, several participants commented that receiving error messages over and over frustrated them. They suggested that error messages may be more helpful if an explanation is provided for wrong answers, and they preferred to be provided with a correct answer after two wrong attempts. Moreover, participants identified several terms might be difficult to understand for younger participants (eg, ages 13-15 years). They suggested that explanations be provided for potentially unfamiliar terms (eg, related to sex work, such as “client/tricks”). In addition to the supplementary explanation, participants recommended that the key sexual health terms/keywords (eg, give head/get head/top/bottom) be bolded for emphasis. Several participants reported an issue with the “Previous” button at the end of an activity; instead of taking them to the previous page, it would erroneously take them back to the beginning of the activity.

    Figure 4. My Bulls-I activity.
    View this figure
    Figure 5. Well Hung activity.
    View this figure
    Table 2. Demographic characteristics of participants (N=20).
    View this table
    Specific Activities Within the App

    Participants also provided comments on specific activities within MyPEEPS Mobile. For instance, several participants noted an issue with the features and functions on an activity, Underwear Personality Quiz, in which participants were introduced to four YMSM avatars, and the four avatars’ personality traits were shared with a “gossip” link (Figure 6). Participants had difficulty recalling if they had viewed each avatar’s gossip page. They recommended that an indication mark (eg, checkmark) on the top right corner of each avatar be shown when the avatar’s gossip has been viewed.

    A running theme of the app was sexual risk reduction and goal-setting through an activity called the BottomLine. In this activity, participants were challenged to articulate how much risk they were willing to accept for different sexual acts. They were asked to continually reconsider these limits as they progressed through the app (Textbox 1; the activity appears four times throughout the app). At the end, participants were presented with the activity #21, BottomLine Overview, which shows them a chronological overview of how their BottomLine changed as they progressed through the app. The participants were then encouraged to continue to stick to their sexual health goals. Many participants felt that the activity required too much reading (ie, full history of the end user’s BottomLine changes for each sexual act), and/or they did not recognize that the display reflected their own changes (ie, selected responses when they previously completed the four BottomLine activities). One participant commented, “It was like a receipt. It was so long! Oh, I thought it was other people’s [BottomLines].” Rather than a full report, participants suggested that we show only the most current BottomLine responses and highlight where they made changes. One participant stated:

    I need only the most current bottom line. I feel like if you click on the bottom line link, you should show like the most current one, and then below that maybe like a link to see the previous ones.

    Another participant commented:

    Just divide it. Put like all of what I first answered my bottom like, and have it separate in the first square, and then, read it down in a different square, like point out some changes I’ve made to my bottom line since the last time. So, I can compare to the ones on top to see how different they are so they’re not all like mixed.
    Table 3. Technology use by participants (N=20).
    View this table
    Figure 6. Underwear Personality Quiz.
    View this figure
    Table 4. Post-Study System Usability Questionnaire (PSSUQ) scoresa (N=20).
    View this table

    Participants’ perceived usability scores rated using PSSUQ [29] are reported in Table 4. Mean of the overall PSSUQ score was 1.63 (SD 0.65), reflecting strong user acceptance of MyPEEPS Mobile.


    Discussion

    Principal Findings

    Despite the proliferative use of mobile technologies in health care, few mHealth apps have been released with consideration of their quality through comprehensive and rigorous usability evaluations. Given a lack of evidence-based mHealth apps for HIV risk reduction in high-risk populations, we developed a mobile HIV prevention app for young men ages 13 to 18 years and assessed the app’s usability through a heuristic evaluation with informatics experts and end user testing to identify potential usability issues.

    Usability factors remain obstacles to mobile technology adoption. Usability evaluations are foundational to the success of achieving systems that meet human-computer interaction principles and in many cases improve their application in a real-world setting [31]. In the context of rigorous usability testing of the systems, it is critical to choose the most appropriate evaluation methods that meet the study aims and ultimately achieve the goals of the systems. We employed two usability evaluation methods most commonly used in usability studies (ie, a heuristic evaluation and end user testing) to capture different usability perspectives from experts and end users [32]. Similar to prior research, usability experts were more likely to identify usability problems related to general interface features working in a natural and logical order [33,34], whereas end users identified those related to impact on task performance interacting with the app [16]. For example, contrary to feedback received from experts in the heuristic evaluation regarding the usability factor “match between system and the real world,” our end users did not identify the natural/logical order of response options as a problem. Given the natural/logical order matched to the real world, no risk (0) should start on the left ending with higher risk (3) on the right in the Well Hung activity (Figure 5). An example of the attitude of end users regarding the ordering issues was expressed by one participant, who stated, “It was something that I didn’t really pay attention to. I would say like I don’t mind it. It doesn’t matter.” Inclusion of intended end users in the usability testing, in addition to the heuristic experts, enabled us to identify both logic and flow issues as well as functionality most important to end users to support overall engagement with the app [35].

    Use case scenarios play an important role in usability evaluations, impacting the quality of usability testing [36]. The use case scenarios should be formulated to facilitate determination of system usability by researchers. Guided by the objectives of usability testing, use case scenarios should include key tasks that can provide valid usability data related to users’ experience with the app use. In this study, we utilized two versions of use case scenarios to capture specific aspects of representative tasks as well as “big picture” issues related to the goals of the app. For example, in the first version of use case scenarios we included tasks that examine end users’ performance on every type of learning activity including comics, animated videos, and games. Given that the running theme throughout the app was sexual risk reduction and goal-setting via the BottomLine activity, in the next version of use case scenarios we included all tasks to test end users’ engagement in BottomLine activities in addition to the learning activities we included in the first version. Using these different versions of robust use case scenarios in end user usability testing enabled us to identify areas where usability was a potential concern and to obtain users’ valuable comments on their overall app use (ie, version 1) as well as specific task performance (ie, version 2), which is a strength of our usability study.

    Although findings from our usability evaluations yielded specific feedback regarding ways to improve users’ experience, the overall usability scores rated by the PSSUQ were high, which indicated that our app was perceived as highly usable. The results support the potential for high acceptability of MyPEEPS Mobile. The promising usability of the app provides a foundation for user satisfaction in the planned randomized controlled trial which aims to reduce sexual risk behavior in a high-risk population.

    Limitations

    The generalizability of the results may be limited by the study sample, settings, and inclusion/exclusion criteria. Our targeted population was diverse YMSM living within the metropolitan area in New York, Chicago, and Birmingham, and those who had either kissed another man or planned on having sex with a man in the next year. Results may differ in transgender groups, other groups who live in rural areas, or those who have more/less experience in sexual activities. Although the age range for inclusion in this study was between 13 and 18 years, our participants were between 15 and 18 years of age, which may differ in younger adolescent MSM (eg, 13-14 years of age). A limitation of this study was related to the participants’ self-reported data such as end users’ perceived usability scores, which may be influenced and could bias the results.

    A limitation of this study was that we conducted the usability evaluations on a computer as opposed to a mobile device. We chose to conduct the evaluations on a computer so that we would be able to analyze the data more effectively using Morae and iMotions. Understanding that there may be differences in computer and mobile device user interactions, we employed an iOS simulator on the computer so that the app can be utilized in the same manner as on a mobile phone. Although still a limitation, the use of the iOS simulator minimized its impact.

    Conclusions

    We tested the usability for an evidence-based HIV prevention mobile app intended for diverse YMSM through a heuristic evaluation with informatics experts and end user testing. The use of the two usability assessment methods for a mHealth app added value to this study by producing reliable results of a user interface from experts as well as user interaction with the app from end users. Findings from our rigorous usability evaluations will be used to refine the content, organization, and workflow of MyPEEPS Mobile. Following these refinements, we will conduct a 6-week pilot study to assess end users’ acceptability of the app before beginning a multicity, 12-month efficacy study. Our work highlights the importance of utilizing a rigorous usability approach to refine a mHealth app before it is deployed in a high stakes environment.

    Acknowledgments

    This research was supported by the National Institute of Minority and Health Disparities of the National Institutes of Health (NIH) under award number U01MD11279 (multiple principal investigators including RS and RG). The content is solely the responsibility of the authors and does not necessarily represent the official views of the NIH.

    Conflicts of Interest

    None declared.

    References

    1. US Department of Health and Human Services. US Food and Drug Administration. 2018. Mobile medical applications   URL: https://www.fda.gov/MedicalDevices/DigitalHealth/MobileMedicalApplications/default.htm [accessed 2018-08-23] [WebCite Cache]
    2. Research2Guidance. mHealth App Economics 2017: Current Status and Future Trends in Mobile Health. 2017 Nov.   URL: https:/​/www.​uzelf.org/​wp-content/​uploads/​2017/​12/​R2G-mHealth-Developer-Economics-2017-Status-And-Trends.​pdf [accessed 2018-08-23] [WebCite Cache]
    3. Amanda L, Rich L, Scott C, Kristen P. Teens and Mobile Phones. Washington, DC: Pew Internet & American Life Project; 2010 Apr 10.   URL: http:/​/www.​pewinternet.org/​files/​old-media/​/Files/Reports/2010/PIP-Teens-and-Mobile-2010-with-topline.​pdf [accessed 2018-08-23] [WebCite Cache]
    4. Kleinke JD. Dot-gov: market failure and the creation of a national health information technology system. Health Aff (Millwood) 2005;24(5):1246-1262. [CrossRef] [Medline]
    5. Devi BR, Syed-Abdul S, Kumar A, Iqbal U, Nguyen P, Li YJ, et al. mHealth: an updated systematic review with a focus on HIV/AIDS and tuberculosis long term management using mobile phones. Comput Methods Programs Biomed 2015 Nov;122(2):257-265. [CrossRef] [Medline]
    6. Roess A. The promise, growth, and reality of mobile health - another data-free zone. N Engl J Med 2017 Nov 23;377(21):2010-2011. [CrossRef] [Medline]
    7. Sullivan PS, Grey JA, Simon RB. Emerging technologies for HIV prevention for MSM: what we have learned, and ways forward. J Acquir Immune Defic Syndr 2013 Jun 1;63 Suppl 1:S102-S107 [FREE Full text] [CrossRef] [Medline]
    8. Kay M. mHealth: New Horizons for Health Through Mobile Technologies: Second Global Survey on eHealth. Geneva: World Health Organization; 2011.
    9. Goldberg L, Lide B, Lowry S, Massett HA, O'Connell T, Preece J, et al. Usability and accessibility in consumer health informatics current trends and future challenges. Am J Prev Med 2011 May;40(5 Suppl 2):S187-S197. [CrossRef] [Medline]
    10. Zapata B, Fernández-Alemán JL, Idri A, Toval A. Empirical studies on usability of mHealth apps: a systematic literature review. J Med Syst 2015 Feb;39(2):1. [CrossRef] [Medline]
    11. Zahra F, Hussain A, Mohd H. Usability factors of mobile health application for chronic diseases. In: AIP Conference Proceedings 1761. 2016 Aug 12 Presented at: ICAST’16 International Conference on Applied Science and Technology; April 11-13, 2016; Kedah, Malaysia p. 020108   URL: https://aip.scitation.org/doi/pdf/10.1063/1.4960948
    12. Yusof MM, Stergioulas L, Zugic J. Health information systems adoption: findings from a systematic review. Stud Health Technol Inform 2007;129(Pt 1):262-266. [Medline]
    13. Schnall R, Mosley JP, Iribarren SJ, Bakken S, Carballo-Diéguez A, Brown IW. Comparison of a user-centered design, self-management app to existing mHealth apps for persons living with HIV. JMIR Mhealth Uhealth 2015 Sep 18;3(3):e91 [FREE Full text] [CrossRef] [Medline]
    14. Issa T, Isaias P. Usability and human computer interaction (HCI). In: Sustainable Design. London: Springer; 2015:19-36.
    15. Yen P, Bakken S. A comparison of usability evaluation methods: heuristic evaluation versus end-user think-aloud protocol - an example from a web-based communication tool for nurse scheduling. AMIA Annu Symp Proc 2009 Nov 14;2009:714-718 [FREE Full text] [Medline]
    16. Jaspers MWM. A comparison of usability methods for testing interactive health technologies: methodological aspects and empirical evidence. Int J Med Inform 2009 May;78(5):340-353. [CrossRef] [Medline]
    17. Centers for Disease Control and Prevention. 2018. HIV among youth   URL: https://www.cdc.gov/hiv/group/age/youth/index.html [accessed 2018-08-23] [WebCite Cache]
    18. Bandura A. Social cognitive theory exercise control of HIV infection. In: DiClemente R, Peterson J, editors. Preventing AIDS: Theories and Methods of Behavioral Interventions. New York: Plenum Press; 1994:25-59.
    19. Hidalgo MA, Kuhns LM, Hotton AL, Johnson AK, Mustanski B, Garofalo R. The MyPEEPS randomized controlled trial: a pilot of preliminary efficacy, feasibility, and acceptability of a group-level, HIV risk reduction intervention for young men who have sex with men. Arch Sex Behav 2015 Feb;44(2):475-485 [FREE Full text] [CrossRef] [Medline]
    20. Macapagal K, Moskowitz DA, Li DH, Carrión A, Bettin E, Fisher CB, et al. Hookup app use, sexual behavior, and sexual health among adolescent men who have sex with men in the United States. J Adolesc Health 2018 Jun;62(6):708-715. [CrossRef] [Medline]
    21. Schnall R, Kuhns L, Hidalgo M, Hirshfield S, Pearson C, Radix A, et al. Development of MyPEEPS Mobile: a behavioral health intervention for young men. Stud Health Technol Inform 2018;250:31. [Medline]
    22. Schnall R, Kuhns L, Bullock K, Powell D, Thai J, Pearson C, et al. Participatory end-user feedback to update MyPEEPS: a theory-driven evidence based intervention for YMSM. 2018 Nov 10 Presented at: American Public Health Association (APHA) Annual Meeting & Expo; Nov 10-14, 2018; San Diego, CA.
    23. Nielsen J, Molich R. Heuristic evaluation of user interfaces. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 1990 Apr 01 Presented at: CHI '90 SIGCHI Conference on Human Factors in Computing Systems; April 1-5, 1990; Seattle, WA p. 249-256.
    24. Po S, Howard S, Vetere F, Skov M. Heuristic evaluation and mobile usability: bridging the realism gap. In: Mobile Human-Computer Interaction - MobileHCI 2004. Berlin: Springer; 2004:49-60.
    25. TechSmith. Morae Usability and Web Site Testing. Michigan, USA: TechSmith Corporation; 1995.
    26. Bright TJ, Bakken S, Johnson SB. Heuristic evaluation of eNote: an electronic notes system. AMIA Annu Symp Proc 2006:864 [FREE Full text] [Medline]
    27. Nielsen J. Heuristic evaluation. In: Usability Inspection Methods. New York: John Wiley & Sons; 1994:25-62.
    28. Faulkner L. Beyond the five-user assumption: benefits of increased sample sizes in usability testing. Behav Res Methods Instrum Comput 2003 Aug;35(3):379-383. [Medline]
    29. Lewis JR. Psychometric Evaluation of the PSSUQ Using Data from Five Years of Usability Studies. Int J Hum-Comput Int 2002 Sep;14(3-4):463-488. [CrossRef]
    30. Nielsen J. Finding usability problems through heuristic evaluation. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 1992 May 03 Presented at: CHI '92 SIGCHI Conference on Human Factors in Computing Systems; May 3-7, 1992; Monterey, CA p. 373-380.
    31. Norgaard M, Hornb K. What do usability evaluators do in practice? An explorative study of think-aloud testing. In: Proceedings of the 6th Conference on Designing Interactive Systems. 2006 Presented at: DIS '06 6th conference on Designing Interactive systems; June 26-28, 2006; University Park, PA p. 209-218.
    32. Lai T, Larson E, Rockoff ML, Bakken S. User acceptance of HIV TIDES--Tailored Interventions for Management of Depressive Symptoms in persons living with HIV/AIDS. J Am Med Inform Assoc 2008;15(2):217-226 [FREE Full text] [CrossRef] [Medline]
    33. Lathan CE, Sebrechts MM, Newman DJ, Doarn CR. Heuristic evaluation of a web-based interface for internet telemedicine. Telemed J 1999;5(2):177-185. [CrossRef] [Medline]
    34. Lai T. Iterative refinement of a tailored system for self-care management of depressive symptoms in people living with HIV/AIDS through heuristic evaluation and end user testing. Int J Med Inform 2007 Oct;76 Suppl 2:S317-S324 [FREE Full text] [CrossRef] [Medline]
    35. Schnall R, Bakken S, Rojas M, Travers J, Carballo-Dieguez A. mHealth technology as a persuasive tool for treatment, care and management of persons living with HIV. AIDS Behav 2015 Jun;19 Suppl 2:81-89 [FREE Full text] [CrossRef] [Medline]
    36. Russ A, Saleem J. Ten factors to consider when developing usability scenarios and tasks for health information technology. J Biomed Inform 2018 Feb;78:123-133. [CrossRef] [Medline]


    Abbreviations

    GED: General Equivalency Diploma
    IRB: Institutional Review Board
    mHealth: mobile health
    MyPEEPS: Male Youth Pursuing Education, Empowerment & Prevention around Sexuality
    PSSUQ: Post-Study System Usability Questionnaire
    YMSM: young men who have sex with men


    Edited by G Eysenbach; submitted 29.06.18; peer-reviewed by C Adewole, S Canidate, K Bond, D Levine; comments to author 25.07.18; revised version received 03.08.18; accepted 07.08.18; published 07.09.18

    ©Hwayoung Cho, Dakota Powell, Adrienne Pichon, Jennie Thai, Josh Bruce, Lisa M Kuhns, Robert Garofalo, Rebecca Schnall. Originally published in JMIR Mhealth and Uhealth (http://mhealth.jmir.org), 07.09.2018.

    This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mhealth and uhealth, is properly cited. The complete bibliographic information, a link to the original publication on http://mhealth.jmir.org/, as well as this copyright and license information must be included.