Accessibility settings

Published on in Vol 14 (2026)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/64787, first published .
Spanish Version of the mHealth App Usability Questionnaire (S-MAUQ): Translation, Adaptation, and Validation Study

Spanish Version of the mHealth App Usability Questionnaire (S-MAUQ): Translation, Adaptation, and Validation Study

Spanish Version of the mHealth App Usability Questionnaire (S-MAUQ): Translation, Adaptation, and Validation Study

1Department of Languages and Computer Science, University of Málaga, C/ Doctor Ortiz Ramos s/n (Campus de Teatinos), Málaga, Spain

2Department of Physiotherapy, University of Málaga, Málaga, Spain

*all authors contributed equally

Corresponding Author:

Antonio L Carrillo-León, PhD


Background: In the current digital landscape, ensuring optimal usability is one of the most crucial factors determining the success of any mobile app. Questionnaire-based usability evaluations represent a highly prevalent methodology for this purpose. To date, questionnaires have been developed to assess the general system usability; however, there are hardly any questionnaires specifically designed to assess the usability of mobile health (mHealth) apps. The most widespread, the mHealth App Usability Questionnaire (MAUQ), has been developed in 4 versions according to the type of app (interactive or standalone) and the target user (patient or provider).

Objective: The objective of this study was to translate and validate the English version of the MAUQ (standalone, for patients) into a Spanish version (S-MAUQ).

Methods: The methodology used here follows that proposed by Sousa and Rojjanasrirat, which comprises 4 stages. The initial stage of the process entails a translation, harmonization, and adaptation procedure. The second and third entailed content validation (by 10 experts) and face validation (by 12 target users), respectively, which were conducted to evaluate the relevance and clarity of the questionnaire items. The item-level content validity index, scale content validity index (S-CVI), item-level face validity index, and scale face validity index (S-FVI), as well as the modified kappa statistic (κ) were used to evaluate interrater agreement among the raters, considering the probability of agreement by chance (Pc). The fourth and final stage of the process involved the assessment of the questionnaire’s reliability. A sample of 61 young adult participants installed an mHealth app (the Yazio app), used it, and responded to the S-MAUQ. The Cronbach α value for the entire questionnaire and its subscales were then calculated.

Results: For the second stage, the S-CVI was initially 0.778. We removed items #14 and #15 from the Spanish version as they were unclear and not relevant. The S-CVI changed to 0.881. The third stage had an S-FVI of 0.927, indicating that the items are clear and straightforward for the nonexpert target user to understand. Furthermore, with each κ value >0.74, the validity of the instrument is supported. The fourth stage demonstrated the reliability of the S-MAUQ with a Cronbach α value of 0.87.

Conclusions: The final version of the S-MAUQ met the validation criteria, demonstrating reliability and validity that are comparable with those of the original version. Consequently, the S-MAUQ is suitable for evaluating the usability of mHealth apps for young Spanish adults. Further research involving larger and more diverse samples is recommended.

JMIR Mhealth Uhealth 2026;14:e64787

doi:10.2196/64787

Keywords



mHealth is an acronym for mobile health, which is used to describe the practice of medicine and public health that is supported by mobile devices [1]. Such mobile devices encompass mobile phones, tablets, and wearable devices, including smart watches. mHealth apps can collect health information for practitioners, researchers, and patients alike and can monitor patients’ vital signs in real time. They facilitate direct care using mobile telemedicine and are used in the training of and collaboration with health workers [2].

The use of mHealth apps is increasing at a considerable rate [3], as evidenced by the growing number of users or patients [4]. Moreover, the emergence of innovative eHealth and mHealth technologies has the potential to significantly improve the cost-effectiveness of the health care system [5].

Presently, these health apps are typically developed either by researchers or, in the majority of cases, by commercial companies, without collaboration between these two groups. The lack of interaction between researchers and commercial developers in the field of health-related apps has resulted in a situation where commercially available apps have not been subjected to scientific validation and apps that have been developed from research projects are not commercially available [6].

It is imperative that mHealth apps are designed in a manner that ensures optimal usability. Such apps must be straightforward to use and capable of achieving their intended objectives in an efficient manner. Well-designed mHealth apps have demonstrated cost-effectiveness through the education and empowerment of patients as well as through improved treatment adherence. However, some reviews have highlighted the absence of criteria for assessing the validity of mHealth apps.

mHealth apps can be classified into two principal categories: those designed for patients and those designed for health care providers. Patient-focused apps are designed to assist with the management of health behaviors, whereas those oriented toward health care providers facilitate services such as prescriptions and education. Furthermore, mHealth apps can be classified into two additional categories: interactive apps, which facilitate communication with health care providers, and standalone apps, which include features such as reminders but lack provider interaction [7].

The optimal usability of apps is characterized by several key attributes, including efficiency, satisfaction, learnability, memorability, and low error rates [7,8]. However, the intrinsic constraints of mobile devices, notably their limited display area, can present obstacles to the attainment of these objectives [9].

The People at the Center of Mobile App usability model considers the user, task, and context of use to evaluate usability. This includes the cognitive load associated with tasks such as exercising while using the app, as discussed by Harrison et al [10]. To minimize errors, usability should be assessed using the following criteria: effectiveness, efficiency, satisfaction, learnability, memorability, and cognitive load [10].

A few usability scales have been evaluation to ascertain their relative importance in the context of the development of general software systems and mobile apps. For example, the Mobile App Rating Scale (MARS) [11], its user version (uMARS) [12], System Usability Scale [13,14], and Post-Study System Usability Questionnaire [15]. However, these scales are not appropriate for evaluating mHealth apps, as they were not developed with mHealth apps as a specific focus.

One of the most pivotal usability scales for assessing the usability of mHealth apps is the mHealth App Usability Questionnaire (MAUQ) [7]. This assesses the ease of use, interface, satisfaction, and usefulness of mHealth apps for end users.

The standalone for patient version of MAUQ has been translated, among other languages, into Malay, German, and Italian [16-18]. It is noteworthy that, while this research was being developed, another related study was published (a Spanish version of MAUQ standalone, for patient [19]) but adapted to breastfeeding support apps. However, it has too many differences from the original and some incorrect formulas, as will be addressed in the discussion section.

The objective of our study was to translate and validate the English version of the MAUQ (standalone, for patients) into a Spanish version (S-MAUQ) for use in future mHealth app research and development.

The importance of the translation into Spanish and its corresponding validation lies in two fundamental points: on the one hand, the large size of the Spanish-speaking community and, on the other, the cultural diversity of the type of user of these apps. Without this linguistic adaptation, these questionnaires would be out of reach for many users.


Overview of the Questionnaire

As previously stated, the MAUQ, developed and validated by Zhou and colleagues [7], is designed to assess the usability of mHealth apps among patients and health care providers. There are two versions of MAUQ, one for interactive apps and one for stand-alone apps. Each also addresses two of its potential users: patient or health care professional. We will work on MAUQ for standalone apps intended for patients. Always consider that the term “patient” refers to a person who uses an mHealth app to manage their health either for maintenance or improvement.

The questionnaire is comprised of 3 subscales or dimensions: ease of use (items #1 to #5), interface and satisfaction (items #6 to #12), and usefulness (items #13 to #18). Participants are required to indicate their level of agreement with each item on a 7-point Likert scale, with 1 representing strong disagreement and 7 representing strong agreement. The overall usability of the app is determined by calculating the total score and the mean score of all statements. A higher overall mean score is indicative of superior app usability. Conversely, a mean score <4 indicates suboptimal usability.

The original MAUQ (standalone, for patient version) exhibited robust internal consistency, with an overall Cronbach α value of 0.914. This indicates a high degree of correlation between the questionnaire items, indicating good reliability. The study was conducted in 2020 [7].

The methodology used to translate and validate the original MAUQ (standalone, for patient version) into Spanish followed the approach proposed by Sousa and Rojjanasrirat [20], with the objective of ensuring equivalence between the original and translated versions. The process is comprised of several steps (illustrated in Figure 1), which can be grouped into the following 4 stages: translation and adaptation, content validation, face validation, and internal consistency reliability.

Figure 1. The flow of translation, cross-cultural adaptation, and validation process for the Spanish version of the mHealth App Usability Questionnaire (S-MAUQ).

Translation and Adaptation Process

The preliminary phase of the translation process entailed the participation of two certified translators, each with a specialized understanding of both Spanish culture and linguistics. These professionals engaged in a forward translation from English to Spanish. One translator had a background in computer science, while the other had a background in health sciences. This resulted in the generation of two distinct documents designated T1 and T2.

Subsequently, the documents (T1 and T2) underwent a harmonization process. A third translator, a Spanish language teacher with more than 5 years of teaching experience, was tasked with identifying any ambiguities; discrepancies; and differences in words, sentences, grammar, and meaning. This was accomplished through a comparison of T1 and T2 with the original MAUQ. Any discrepancies and ambiguities were discussed and resolved through a process of consensus decision-making involving the third translator, the first two translators, and two members of the research team. The process culminated in the production of an initial translated version of the MAUQ document (T3).

In the third phase of the process, the translated version of the MAUQ (T3) was provided to 2 independent native Spanish speakers who had acquired English as a second language. One individual (a computer science lecturer) was a content expert in the instrument’s construct, while the other had no prior knowledge of the instrument (an English language teacher). The 2 participants then produced independent back-translated versions of the MAUQ (T4 and T5).

Last, a committee was established comprising 3 researchers, the translators involved in the initial and concluding stages, and the original developers of the MAUQ (when necessary). The committee conducted a comparative analysis of the documents with the original MAUQ, engaging in discussions via email and web-based meetings. Any ambiguities or inconsistencies in cultural meaning and expression of words and sentences and the format of the answers were addressed and resolved through consensus among the committee members. This process resulted in the development of the prefinal Spanish questionnaire, designated as S-MAUQ’.

Content Validation and Face Validation

Content validation is achieved by a panel of experts who assess the relevance and clarity of questionnaire items for measuring the usability of health-related mobile apps. Face validation is accomplished by a set of nonexperts or target users who specifically assess the clarity of the items.

Participants

For the content validation, a panel of 10 experts was formed. Our panel included 3 mobile app developers, 2 senior lecturers with a PhD in computer science, 2 usability experts, and 3 PhD senior lecturers and health-related practitioners.

For the face validation, 12 nonexpert participants were involved.

Recruitment

For the content validation, the 10 expert participants were carefully selected in order to adequately cover all areas of expertise involved. After being informed of the study’s objective, all participants signed the corresponding informed consent form.

For the face validation, users were randomly selected from an openly chosen sample. The inclusion criteria were as follows: (1) native Spanish speaker, (2) older than 18 years, and (3) familiar with smartphones. The exclusion criteria were being (1) health-related practitioners, (2) mobile app developers, or (3) computer science students or professionals.

Participants were asked to complete an anonymous web form that included the informed consent, gender, date of birth, and level of education.

Setting

The group of expert participants completed their evaluations remotely and sent them to us via email. The nonexpert participants completed their evaluation online via an anonymous web form.

Study Procedures

In the content validation stage, each expert is required to assign a score on a scale of 1 to 5 to each item. A score of 1 indicates that the item was not relevant or clear, while a score of 5 indicates that the item was highly relevant and clear. Subsequently, scores of 1 to 3 were recategorized as 0, indicating irrelevance or unclarity, while scores of 4 and 5 were recategorized as 1, indicating relevance or clarity. The final categorization was conducted with the objective of obtaining dichotomous values. In numerous other studies, the range extends from 1 to 4. In contrast, in our study, the range was from 1 to 5. However, the current quality standards are more demanding, as a rating of 3 was considered unsatisfactory. In addition, there were 3 instances of unsatisfactory performance, as opposed to only 2 instances of acceptable performance.

On the other hand, in the face validation, which was less demanding with the nonexpert evaluators, the scores ranged from 1 (indicating that the item was not understandable) to 4 (indicating that the item was very understandable). Scores of 1 and 2 were recategorized as 0 (not understandable), and scores of 3 and 4 were recategorized as 1 (understandable). This was less demanding because the chosen scores were taken directly without considering the middle 3 negative values as we did with the experts.

The calculations for the 2 types of evaluation were as described in the following statements.

If we have N as the number of raters or evaluators; AC as the number of expert raters who concur that the items are relevant, unambiguous, and intelligible; and AF as the number of nonexpert raters who concur that the items are clear and comprehensible, the item-level content validity index (I-CVI) and the item-level face validity index (I-FVI) were calculated by dividing the propoortion of raters who agreed that an item was clear and relevant to the domain by the total number of raters.

In conclusion, it can be proposed that:

ICVI=ACN and IFVI=AFN

Items with an I-CVI or I-FVI of at least 0.79 were relevant, unambiguous, and understandable to the target users, indicating their potential inclusion in the questionnaire [21,22].

The scale content validity index (S-CVI) and scale face validity index (S-FVI) can be calculated as the means of the corresponding I-CVI and I-FVI. This is accomplished by dividing the sum of each item by the total number of items [23].

Moreover, it is essential to determine the level of agreement among the 10 experts. This can be accomplished by calculating the modified kappa statistic [21,24,25], which can be expressed as follows:

κCVI=I-CVIPc1Pc and κFVI=I-FVIPc1Pc

In this formula, PC represents the probability of chance occurrence. The initial determination of the probability of agreement by chance for each item was determined using the following formula (the formula for a binomial random variable) [21]:

Pc=(NA)0.5N=N!A! (NA)!0.5N

This equation has its origins in the field of probability theory, specifically in the binomial distribution. The probability of obtaining precisely A successes in N independent Bernoulli trials (with the same rate p, in this case 1/2) is given by the probability mass function (or discrete probability density function):

Pc=(NA) pk(1p)NA and for p=1/2 → Pc=(NA) 0.5N

Graphically, the interpretation criteria for the modified kappa statistic, as outlined by Wynd et al [26] and Cicchetti and Sparrow [27], is shown in Figure 2.

Figure 2. Interpretation criteria for the modified kappa statistic.

Reliability (Internal Consistency) Validation

Participants

For the internal consistency validation, the participants were selected from among students enrolled in the School of Engineering at the University of Malaga and the Campus of International Excellence (Andalucía Tech). The required sample size was calculated as 38 participants for the purposes of reliability testing, with an expected Cronbach α of 0.80 and an expected precision of 0.1 at 95% CI [28]. Given the projected rate of nonparticipation of 20%, the final sample size was determined to be 48.

Recruitment

Participants were recruited through direct invitations from the involved professors during class time. Those who agreed to take part in the study were required to complete a web-based informed consent form. Inclusion criteria were an absence of any known medical illness and ownership of a smartphone.

Setting

Participants performed the tasks on their smartphones and in their own classrooms and used a Google Form to submit their answers to the S-MAUQ.

Study Procedures

Participants were instructed to install the Yazio app on their mobile phones. Yazio, a standalone mHealth app for patients, was selected for the usability study due to its popularity. It is one of the most widely used nutrition apps in the world, helping people lead healthier lives globally. Its availability on both Android and iPhone devices [29,30] combined with its similarity to MyFitnessPal—an app frequently used in previous studies but that requires additional resources to be used—made it a suitable choice. A concise overview of the Yazio app, along with the essential steps for preliminary configuration, was displayed on the classroom projector. This procedure should be carried out exclusively at the point of the initial execution of the app following its installation.

Subsequently, participants were required to undertake the tasks detailed in Textbox 1.

Textbox 1. Sheet of tasks to be performed using the Yazio app.
  1. Create initial profile when running the app for the first time (take a screenshot of the screen that appears at the end of the process, after discarding the tutorial)
  2. Use the functions available in the Daily and/or Profile menu to add records or data for today:
    1. Add the following foods consumed during breakfast
      • 1 black coffee
      • 2 slices of bread (bimbo) without crust
      • Natural orange juice, 200 mL
      • Olive oil, 1 tablespoon
      • Product with barcode 8410095040362 (or scan it from the product code photo)
    2. Modify (in the profile) daily water amount goal: 2.5 liters
    3. Add water consumed: 1.75 liters
    4. Add activities (exercise completed) by pressing “More”
      • “walk, general” 30 minutes
      • “fast walk (6 km/h)” 10 minutes
    5. Add 2 notes or indications about how your day is going
      • “I slept well”
      • “bored”
    6. (Finally, take the necessary screenshots, 2 or 3, to collect the data entered)
  3. Identify and note (via screenshot):
    1. The number of calories consumed during the day (due to ingested foods and drinks)
    2. Calories burned (from exercising, etc.)
    3. Calories we have left to consume to reach the day’s goal
  4. Explore and find a recipe that meets each of the following conditions:
    1. 1 vegan breakfast (between 150 and 200 kcal) and mark it as a favorite
    2. 1 vegan lunch (between 350 and 400 kcal) and mark it as a favorite
    3. See the recipes you have marked as favorites (and take a screenshot of them)

First, it is recommended that an initial profile be created and the desired weight goals be determined. Second, to add records or data pertaining to the current day, the functions available from the Daily and/or Profile menus need to be used. Third, it is requested that the number of calories consumed throughout the day be identified and recorded. This should include the calories derived from food and drink, the calories expended through exercise, and the remaining calories to be consumed to reach the daily goal. Fourth, it is required to conduct a comprehensive investigation and identify a recipe that aligns with the specific criteria outlined in step 4 in Textbox 1.

Following the completion of the tasks, the participants were invited to respond to the S-MAUQ in person using a Google Form.

The internal consistency of the S-MAUQ was evaluated by calculating the Cronbach α value for the entire questionnaire and its 3 subscales. A higher α value indicates greater internal reliability, and a value of ≥0.70 is generally considered acceptable as an indication of good internal reliability [31,32]. All statistical analyses were conducted using SPSS (IBM Corp), and the results were verified using Excel (Microsoft Corp).

Figure 1 provides an overview of the process of translation and cross-cultural adaptation as well as the subsequent validation stage for S-MAUQ’.

Ethical Considerations

No financial compensation was provided to participants at any stage of the study. Participation was voluntary and anonymous, except for members of the expert panel responsible for content validation, who were intentionally selected for their expertise. All participants (including the expert panel) were provided and signed an informed consent online, which included the possibility of opting out. The study adhered to the Declaration of Helsinki and was approved by the Bioethics Committee of the Medical University of Málaga (17‐2022-H).


Content Validation

Please refer to Table 1, which presents the calculated values (using Excel) of the I-CVI for relevancy and clarity, the (Pc), and the modified kappa agreement (κ) by the 10 experts for each item of the S-MAUQ. The S-CVI was calculated to be 0.778. However, as can be seen, there were 2 items (#3 and #6) with κ<0.4 (interpreted as poor agreement) and 2 others (#15 and #18) with 0.4<κ<0.6 (interpreted as fair). Considering these findings, we sought to further refine the instrument in accordance with the recommendations of the expert panel.

Table 1. Content validity index of item relevancy and modified kappa agreement by 10 experts in round 1.
ItemsRating of 4 or 5, nRating of 1 to 3, nItem-level content validity indexaProbability of chance occurrencebModified kappac,dInterpretation
11001.0000.0011.000Excellent
2910.9000.0100.899Excellent
3550.5000.2460.337Poor
4820.8000.0440.791Excellent
5910.9000.0100.899Excellent
6550.5000.2460.337Poor
71001.0000.0011.000Excellent
8820.8000.0440.791Excellent
9730.7000.1170.660Good
10820.8000.0440.791Excellent
11910.9000.0100.899Excellent
12910.9000.0100.899Excellent
13820.8000.0440.791Excellent
14730.7000.1170.660Good
15640.6000.2050.497Fair
16910.9000.0100.899Excellent
17730.7000.1170.660Good
18640.6000.2050.497Fair

aScale content validity index=0.778.

bComputed using the formula Pc = [ N! / (A! * (N–A)!) ] * 0.5N, where Pc is the probability of chance occurrence, n is the number of experts, and A is the number of experts who agree the items are relevant/clear.

cComputed using the formula κ=(item-level content validity index – Pc) / (1 – Pc), where Pc is the probability of chance occurrence.

dKappa was interpreted according to the following criteria: excellent, κ>0.74; good, 0.6≤κ≤0.74; fair, 0.40≤κ≤0.59; poor, κ<0.40.

The results of the second round are presented in Table 2, which indicates an S-CVI of 0.850. However, 2 items (#14 and #15) demonstrated a markedly low level of relevance and clarity (I-CVI), accompanied by a similarly low κ (0.4≤κ≤0.6, interpreted as “fair”). Following a thorough review with our panel of experts, it was determined that both items should be removed from the preliminary Spanish questionnaire (S-MAUQ). In conclusion, the S-CVI was calculated to be 0.881 (without considering the 2 items that were subsequently eliminated).

Table 2. Content validity index of item relevancy and modified kappa agreement by 10 experts in round 2.
ItemsRating of 4 or 5, nRating of 1 to 3, nItem-level content validity indexaProbability of chance occurrencebModified kappac,dInterpretation
11001.0000.0011.000Excellent
2910.9000.0100.899Excellent
3820.8000.0440.791Excellent
4820.8000.0440.791Excellent
51001.0000.0011.000Excellent
61001.0000.0011.000Excellent
71001.0000.0011.000Excellent
8820.8000.0440.791Excellent
9730.7000.1170.660Good
10910.9000.0100.899Excellent
111001.0000.0011.000Excellent
121001.0000.0011.000Excellent
13910.9000.0100.899Excellent
14640.6000.2050.497Fair
15640.6000.2050.497Fair
16820.8000.0440.791Excellent
17820.8000.0440.791Excellent
18730.7000.1170.660Good

aScale content validity index=0.850.

bComputed using the formula Pc = [ N! / (A! * (N–A)!) ] * 0.5N, where Pc is the probability of chance occurrence, n is the number of experts, and A is the number of experts who agree the items are relevant/clear.

cComputed using the formula κ=(item-level content validity index – Pc) / (1 – Pc), where Pc is the probability of chance occurrence.

dKappa was interpreted according to the following criteria: excellent, κ>0.74; good, 0.6≤κ≤0.74; fair, 0.40≤κ≤0.59; poor, κ<0.40.

Face Validation

Please refer to Table 3 for the calculated values of the I-FVI, the Pc, and the modified kappa agreement (κ) for each item. These were determined by the 12 target users. The S-FVI was calculated to be 0.927. A score >0.79 indicates that the items in the questionnaire are sufficiently clear and understandable for the nonexpert target user [27]. Furthermore, the modified kappa agreement for each item in S-MAUQ demonstrated excellent concordance (κ>0.74).

Table 3. Face validity index of item understandability and modified kappa agreement by 12 target users.
ItemsRating of 3 or 4, nRating of 1 or 2, nItem-level face validity indexaProbability of chance occurrencebModified kappacInterpretation
11201.0000.0001.000Excellent
21201.0000.0001.000Excellent
3930.7500.0540.740Excellent
41110.9170.0030.916Excellent
51110.9170.0030.916Excellent
61110.9170.0030.916Excellent
71110.9170.0030.916Excellent
81110.9170.0030.916Excellent
91020.8330.0160.831Excellent
101020.8330.0160.831Excellent
111201.0000.0001.000Excellent
121201.0000.0001.000Excellent
131201.0000.0001.000Excellent
14ef
15e
161201.0000.0001.000Excellent
171110.9170.0030.916Excellent
181110.9170.0030.916Excellent

aScale face validity index=0.927.

bComputed using the formula Pc = [ N! / (A! * (N–A)!) ] * 0.5N, where Pc is the probability of chance occurrence, n is the number of users, and A is the number of users who agree the items are understandable.

cComputed using the formula: =(item-level face validity index – Pc) / (1 – Pc), where Pc is the probability of chance occurrence.

dKappa was interpreted according to the following criteria: excellent, κ>0.74; good, 0.6≤κ≤0.74; fair, 0.40≤κ≤0.59; poor, κ<0.40.

eRemoved from the questionnaire.

fNot applicable.

Reliability (Internal Consistency) Validation

A total of 61 participants completed the web-based consent form and the S-MAUQ questionnaire for the purposes of the reliability study (the calculated sample size was 48). The participants were between the ages of 18 years and 22 years, and all were students (41 men, 20 women). The mean usability score (normalized to 1) for the Yazio app, as determined by the S-MAUQ, was 0.855 (SD 0.088). This indicates a high level of usability.

The Cronbach α value for the entire S-MAUQ was 0.87, and the values for the 3 subscales were 0.74 for the ease of use dimension (Q1-Q5), 0.74 for the interface and satisfaction dimension (Q6-Q12), and 0.64 for the usefulness dimension (Q13-Q16). The new numbering scheme was used for Q13-Q16, omitting the 2 deleted items.

Furthermore, the Cronbach α value for the complete questionnaire remained highly consistent when an item was deleted (see Table 4), indicating that the S-MAUQ exhibited good internal reliability.

Table 4. The internal consistency of the item’s total statistics.
ItemsScale mean if item deletedaScale variance if item deletedCorrected item total correlationCronbach α if item deletedb
189.4687.0860.5320.854
289.0785.1290.6350.850
389.1085.1900.5700.852
489.0787.5620.4700.857
589.1584.0610.4620.859
688.9089.3230.3450.863
789.4987.7540.4390.858
888.9789.4320.4030.860
989.1387.9490.4360.858
1089.2885.4040.5930.851
1189.7479.7970.5990.851
1289.1385.9490.7020.849
1389.2088.2940.4420.858
14cd
15c
1689.2087.9610.4510.858
1790.0085.0330.4290.860
1889.3385.3240.5250.854

aOverall scale mean=95.21.

bOverall scale Cronbach α=0.87 (high internal consistency); ease of use subscale (Q1..Q5) Cronbach α=0.74; interface and satisfaction subscale (Q6..Q12) Cronbach α=0.74; usefulness subscale (Q13, Q16..Q18) Cronbach α=0.64.

cRemoved from the questionnaire.

dNot appliclable.


Principal Findings

Ensuring optimal usability of Spanish mobile apps is one of the most crucial factors for their use. Questionnaires are a common method for evaluating usability. This study outlined the process of translating, adapting, and validating the English version of the standalone MAUQ questionnaire for the patient into Spanish.

To this end, we followed the 4-step process proposed by Sousa and Rojjanasrirat [20], as it was outlined in the Methods section.

During the translation and adaptation phase, several meetings were held to ensure conceptual and linguistic equivalence for key terms such as interface, consistency, coherence, social setting, and health and well-being services. Final wording was established by consensus, often in consultation with the original questionnaire’s authors.

In the content validation phase, the expert panel initially did not reach full agreement on several items. A second round of review narrowed the discrepancies but left 2 items (#14 and #15) unresolved. Both items were judged to be conceptually ambiguous and difficult to interpret, even for experts. Following a series of email consultations with the original authors on the meaning of and convenience of these items and their overlap with others, we confirmed the experts’ results and their suggestions that these items could be removed from the final Spanish version.

During the face validation phase, the calculated FVI and κ coefficient yielded an S-FVI that substantially exceeded the recommended threshold, thereby ensuring the quality of the translation. The results of the study indicated that the items on the questionnaire were comprehensible and suitable for the target population.

Finally, the reliability analysis demonstrated strong internal consistency, with an overall Cronbach α of 0.87. The Spanish version maintained the original 3-dimensional structure of the English questionnaire, permitting direct cross-language comparison. Only the “usefulness” dimension was slightly affected, with an acceptable Cronbach α of 0.64 for exploratory studies. Overall, the S-MAUQ exhibited sound psychometric properties, indicating that it is a reliable and valid tool for assessing the usability of mobile health apps among Spanish-speaking patients. The final iteration of the S-MAUQ is presented in Table 5 for reference.

Table 5. The final Spanish version of the mHealth App Usability Questionnaire (S-MAUQ).
Item numberQuestion
1La aplicación era fácil de usar.
2Me resultó fácil aprender a utilizar la aplicación.
3Al pasar de una pantalla a otra, la navegación era consistente, similar, previsible.
4La aplicación me permitió usar todas las funciones que necesité (tales como visualizar e introducir información, responder a recordatorios, etc).
5Cuando cometí un error usando la aplicación, fue fácil y rápido deshacerlo y volver a intentarlo.
6Me gusta el diseño y aspecto de la aplicación.
7La presentación de la información en la aplicación estaba bien organizada, y me permitió encontrar lo que buscaba.
8La aplicación mostraba adecuadamente el progreso de mi actividad mientras interactuaba con ella.
9Me sentiría cómodo/a utilizando esta aplicación en público, delante de otras personas.
10El tiempo invertido en realizar las tareas deseadas me ha parecido adecuado.
11Volvería a utilizar esta aplicación.
12En general, estoy satisfecho/a con esta aplicación.
13Creo que la aplicación puede ser útil para mi salud y bienestar.
14 (#16 in MAUQ)Esta aplicación tiene todas las funciones que esperaba.
15 (#17 in MAUQ)He podido realizar algunas tareas incluso cuando la conexión a internet era mala o no estaba disponible.
16 (#18 in MAUQ)Esta aplicación me proporcionó una forma aceptable de recibir servicios de salud, tales como el acceso a material educativo, el seguimiento de mis actividades, o la realización de autoevaluaciones.

Comparison With Prior Research

Our findings align with prior validations of the MAUQ conducted to assess a variety of mobile apps developed for different purposes and user groups, including the MyFitnessPal in Malaysia for evaluating caloric intake, weight goals, and nutritional guidance [16]; the Left-Handed Doctor mHealth app in China for medical consultations and public health guidance [33]; the Enable app in Germany for cancer patients to register electronic patient-reported outcomes and monitor treatment side effects [17]; and the DIGICOG-MS app in Italy for cognitive assessment of individuals with multiple sclerosis [18] and patients with osteoporosis in the Persian study [34]. Collectively, our findings suggest that both the MAUQ and S-MAUQ demonstrate stable content validity across diverse contexts and could be applied to a wide range of mHealth and eHealth applications designed for different populations.

It is noteworthy that a Spanish adaptation [19] developed in parallel with our own work has not adhered to the MAUQ standard; instead, it was adapted for breastfeeding. Several items have undergone alterations and been reordered, and one item was eliminated. The original dimensions were modified from 3 to 4, but the results indicate a significant lack of internal consistency in 2 final subscales. Furthermore, there was an erroneous interpretation and implementation of the formula applied to calculate the level of relevance and clarity of the questionnaire items. Consequently, it was not possible to achieve comparable outcomes in different languages.

We also compared our work with the aforementioned studies that focus on the “standalone, for patients” version of the MAUQ. The Malaysian version of the MAUQ [16] underwent an analysis similar to ours, maintaining the 3-dimensional structure but without eliminating any element. The German version of the MAUQ [17] also performed a study similar to ours and to that of the Malaysian version. However, the creators of the German version considered eliminating 1 item but finally decided not to do so in order to maintain consistency with the original. The Italian version [18] did not undergo content or face validation; they had a panel of 6 experts who directly compared a translated version with the original questionnaire and made modifications to have a more understandable Italian version.

Limitations

In our study, as in other similar work [16], we surveyed a large sample of native Spanish-speaking students to provide a representative sample of the young adult demographic. In future research, criterion validation studies should be conducted to predict different outcomes for different domains and types of users of the S-MAUQ.

Conclusions

We addressed a significant gap in the field of health app evaluation in Spanish by translating and adapting the MAUQ, one of the most recognized usability questionnaires in this area. Despite the need to eliminate 2 items due to their lack of relevance and demonstrated confusion, S-MAUQ is a reliable tool in Spanish for evaluating the usability of mHealth apps.

Acknowledgments

We would like to thank Dr. L. Zhou for granting us permission to translate the mHealth App Usability Questionnaire (MAUQ) into the Spanish version, his help during the translation process, and clarifying some of the items in the questionnaire. Thank you also for conveying to us the interest and the various requests that he received from individuals and research groups interested in having the questionnaire available in Spanish. We also thank the panel of experts and the other participants involved in this study.

Funding

Funding for the open access charge (article processing fee [APF]): Plan Propio de la UMA - Grupo de investigación SICUMA (TIC160) - Máster de Formación Permanente de Fisioterapia en Disfunciones de Suelo Pélvico (Master Ref. 81424587), Universidad de Málaga.

Data Availability

The datasets generated or analyzed during this study are available from the corresponding author on reasonable request.

Authors' Contributions

Conceptualization - ALCL, JFC, EDM

Formal analysis - ALCL, JFC, EDM

Investigation - ALCL, JFC, EDM

Methodology - ALCL, JFC, EDM

Supervision - ALCL, JFC, EDM

Validation - ALCL, JFC, EDM

Writing – original draft - ALCL, JFC

Writing – review & editing - ALCL, JFC, EDM

Conflicts of Interest

None declared.

  1. Adibi S. A technology road map. In: Mobile health. Vol 5. Springer; 2015. [CrossRef]
  2. Pimmer C, Tulenko K. The convergence of mobile and social media: affordances and constraints of mobile networked communication for health workers in low- and middle-income countries. Mobile Media & Communication. May 2016;4(2):252-269. [CrossRef]
  3. Ali EE, Chew L, Yap KYL. Evolution and current status of mhealth research: a systematic review. BMJ Innov. Jan 2016;2(1):33-40. [CrossRef]
  4. Lunden I. 6.1B Smartphone users globally by 2020, overtaking basic phone subscriptions. TechCrunch. Jun 2, 2015. URL: https:/​/techcrunch.​com/​2015/​06/​02/​6-1b-smartphone-users-globally-by-2020-overtaking-basic-fixed-phone-subscriptions/​?guccounter=12015 [Accessed 2023-08-10]
  5. de la Torre-Díez I, López-Coronado M, Vaca C, Aguado JS, de Castro C. Cost-utility and cost-effectiveness studies of telemedicine, electronic, and mobile health systems in the literature: a systematic review. Telemed J E Health. Feb 2015;21(2):81-85. [CrossRef] [Medline]
  6. Ball E, Newton S, Rohricht F, et al. mHealth: providing a mindfulness app for women with chronic pelvic pain in gynaecology outpatient clinics: qualitative data analysis of user experience and lessons learnt. BMJ Open. Mar 12, 2020;10(3):e030711. [CrossRef] [Medline]
  7. Zhou L, Bao J, Setiawan IMA, Saptono A, Parmanto B. The mHealth App Usability Questionnaire (MAUQ): development and validation study. JMIR Mhealth Uhealth. Apr 11, 2019;7(4):e11500. [CrossRef] [Medline]
  8. Nielsen J. Usability Engineering. Morgan Kaufmann; 1993. ISBN: 0-12-518405-0
  9. Zhang D, Adipat B. Challenges, methodologies, and issues in the usability testing of mobile applications. Int J Hum Comput Interact. Jul 2005;18(3):293-308. [CrossRef]
  10. Harrison R, Flood D, Duce D. Usability of mobile applications: literature review and rationale for a new usability model. J Interact Sci. 2013;1(1):1. [CrossRef]
  11. Stoyanov SR, Hides L, Kavanagh DJ, Zelenko O, Tjondronegoro D, Mani M. Mobile app rating scale: a new tool for assessing the quality of health mobile apps. JMIR Mhealth Uhealth. Mar 11, 2015;3(1):e27. [CrossRef] [Medline]
  12. Stoyanov SR, Hides L, Kavanagh DJ, Wilson H. Development and validation of the user version of the Mobile Application Rating Scale (uMARS). JMIR Mhealth Uhealth. Jun 10, 2016;4(2):e72. [CrossRef] [Medline]
  13. Brooke J. SUS—a quick and dirty usability scale. In: Jordan PW, Thomas B, Weerdmeester BA, McClelland AL, editors. Usability Evaluation in Industry. Taylor and Francis; 1996. [CrossRef]
  14. Lewis JR. The System Usability Scale: past, present, and future. International Journal of Human–Computer Interaction. Jul 3, 2018;34(7):577-590. [CrossRef]
  15. Lewis JR. Psychometric evaluation of the PSSUQ using data from five years of usability studies. Int J Hum Comput Interact. Sep 2002;14(3-4):463-488. [CrossRef]
  16. Mustafa N, Safii NS, Jaffar A, et al. Malay version of the mHealth App Usability Questionnaire (M-MAUQ): translation, adaptation, and validation study. JMIR Mhealth Uhealth. Feb 4, 2021;9(2):e24457. [CrossRef] [Medline]
  17. Moorthy P, Weinert L, Harms BC, Anders C, Siegel F. German version of the mHealth App Usability Questionnaire in a cohort of patients with cancer: translation and validation study. JMIR Hum Factors. Nov 1, 2023;10:e51090. [CrossRef] [Medline]
  18. Podda J, Grange E, Susini A, et al. Italian version of the mHealth App Usability Questionnaire (Ita-MAUQ): translation and validation study in people with multiple sclerosis. JMIR Hum Factors. Sep 30, 2024;11:e58079. [CrossRef] [Medline]
  19. Quifer-Rada P, Aguilar-Camprubí L, Gómez-Sebastià I, Padró-Arocas A, Mena-Tudela D. Spanish version of the mHealth App Usability Questionnaire (MAUQ) and adaptation to breastfeeding support apps. Int J Med Inform. Jun 2023;174:105062. [CrossRef] [Medline]
  20. Sousa VD, Rojjanasrirat W. Translation, adaptation and validation of instruments or scales for use in cross-cultural health care research: a clear and user-friendly guideline. J Eval Clin Pract. Apr 2011;17(2):268-274. [CrossRef] [Medline]
  21. Polit DF, Beck CT, Owen SV. Is the CVI an acceptable indicator of content validity? Appraisal and recommendations. Res Nurs Health. Aug 2007;30(4):459-467. [CrossRef] [Medline]
  22. Polit DF, Beck CT. The content validity index: are you sure you know what’s being reported? Critique and recommendations. Res Nurs Health. Oct 2006;29(5):489-497. [CrossRef] [Medline]
  23. Zamanzadeh V, Ghahramanian A, Rassouli M, Abbaszadeh A, Alavi-Majd H, Nikanfar AR. Design and implementation content validity study: development of an instrument for measuring patient-centered communication. J Caring Sci. Jun 2015;4(2):165-178. [CrossRef] [Medline]
  24. Fleiss JL. Statistical Methods for Rates and Proportions. 2nd ed. John Wiley; 1981. ISBN: 0471064289
  25. Cohen J. A coefficient of agreement for nominal scales. Educ Psychol Meas. Apr 1960;20(1):37-46. [CrossRef]
  26. Wynd CA, Schmidt B, Schaefer MA. Two quantitative approaches for estimating content validity. West J Nurs Res. Aug 2003;25(5):508-518. [CrossRef] [Medline]
  27. Cicchetti DV, Sparrow SA. Developing criteria for establishing interrater reliability of specific items: applications to assessment of adaptive behavior. Am J Ment Defic. Sep 1981;86(2):127-137. [Medline]
  28. Arifin WN, Unit of Biostatistics and Research Methodology, School of Medical Sciences, Universiti Sains Malaysia, Kelantan, MALAYSIA. A web-based sample size calculator for reliability studies. EIMJ. 2018;10(3):67-76. URL: http://eduimed.usm.my/index.php/eimj/issue/view/29 [CrossRef]
  29. Ferrara G, Kim J, Lin S, Hua J, Seto E. A focused review of smartphone diet-tracking apps: usability, functionality, coherence with behavior change theory, and comparative validity of nutrient intake and energy estimates. JMIR Mhealth Uhealth. May 17, 2019;7(5):e9232. [CrossRef] [Medline]
  30. Fakih El Khoury C, Karavetian M, Halfens RJG, Crutzen R, Khoja L, Schols J. The effects of dietary mobile apps on nutritional outcomes in adults with chronic diseases: a systematic review and meta-analysis. J Acad Nutr Diet. Apr 2019;119(4):626-651. [CrossRef] [Medline]
  31. Allen P, Bennett K, Heritage B. SPSS Statistics Version 22: A Practical Guide. Cengage Learning Australia; 2014. ISBN: 9780170348973
  32. Onwuegbuzie AJ, Daniel LG. A framework for reporting and interpreting internal consistency reliability estimates. Meas Eval Couns Dev. Jul 1, 2002;35(2):89-103. [CrossRef]
  33. Shan Y, Ji M, Xie W, et al. Chinese version of the Mobile Health App Usability Questionnaire: translation, adaptation, and validation study. JMIR Form Res. Jul 6, 2022;6(7):e37933. [CrossRef] [Medline]
  34. Moulaei K, Dinari F, Sheikhtaheri A, Bahaadinbeigy K, Hajesmaeel-Gohari S. The Persian version of mhealth App Usability Questionnaire (MAUQ) for patients: a psychometric assessment study. BMC Med Inform Decis Mak. Jan 23, 2025;25(1):38. [CrossRef] [Medline]


I-CVI: item-level content validity index
I-FVI: item-level face validity index
MARS: Mobile App Rating Scale
MAUQ: mHealth App Usability Questionnaire
mHealth: mobile health
Pc: probability of chance occurrence
S-CVI: scale content validity index
S-FVI: scale face validity index
S-MAUQ: Spanish version of the MAUQ
uMARS: MARS user version


Edited by Lorraine Buis; submitted 16.Aug.2024; peer-reviewed by Aziah Daud, Florence Carrouel; final revised version received 20.Dec.2025; accepted 11.Feb.2026; published 20.Apr.2026.

Copyright

© Antonio L Carrillo-León, Juan Falgueras-Cano, Esther Díaz-Mohedo. Originally published in JMIR mHealth and uHealth (https://mhealth.jmir.org), 20.Apr.2026.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mHealth and uHealth, is properly cited. The complete bibliographic information, a link to the original publication on https://mhealth.jmir.org/, as well as this copyright and license information must be included.