Published on in Vol 12 (2024)

Preprints (earlier versions) of this paper are available at https://preprints.jmir.org/preprint/51592, first published .
Patient-Centered Chronic Wound Care Mobile Apps: Systematic Identification, Analysis, and Assessment

Patient-Centered Chronic Wound Care Mobile Apps: Systematic Identification, Analysis, and Assessment

Patient-Centered Chronic Wound Care Mobile Apps: Systematic Identification, Analysis, and Assessment

1Department of Dermatology, Venereology and Allergology, University Hospital Würzburg, , Würzburg, , Germany

2Department of Computer Science, University of Würzburg, , Würzburg, , Germany

*these authors contributed equally

Corresponding Author:

Tassilo Dege, MSc, MD


Background: The prevalence of chronic wounds is predicted to increase within the aging populations in industrialized countries. Patients experience significant distress due to pain, wound secretions, and the resulting immobilization. As the number of wounds continues to rise, their adequate care becomes increasingly costly in terms of health care resources worldwide. eHealth support systems are being increasingly integrated into patient care. However, to date, no systematic analysis of such apps for chronic wounds has been published.

Objective: The aims of this study were to systematically identify and subjectively assess publicly available German- or English-language mobile apps for patients with chronic wounds, with quality assessments performed by both patients and physicians.

Methods: Two reviewers independently conducted a systematic search and assessment of German- or English-language mobile apps for patients with chronic wounds that were available in the Google Play Store and Apple App Store from April 2022 to May 2022. In total, 3 apps met the inclusion and exclusion criteria and were reviewed independently by 10 physicians using the German Mobile App Rating Scale (MARS) and the System Usability Scale (SUS). The app with the highest mean MARS score was subsequently reviewed by 11 patients with chronic wounds using the German user version of the MARS (uMARS) and the SUS. Additionally, Affinity for Technology Interaction (ATI) scale scores were collected from both patients and physicians.

Results: This study assessed mobile apps for patients with chronic wounds that were selected from a pool of 118 identified apps. Of the 73 apps available in both app stores, 10 were patient oriented. After excluding apps with advertisements or costs, 3 apps were evaluated by 10 physicians. Mean MARS scores ranged from 2.64 (SD 0.65) to 3.88 (SD 0.65) out of 5, and mean SUS scores ranged from 50.75 (SD 27) to 80.5 (SD 17.7) out of 100. WUND APP received the highest mean MARS score (mean 3.88, SD 0.65 out of 5) among physicians. Hence, it was subsequently assessed by 11 patients and achieved a similar rating (uMARS score: mean 3.89, SD 0.4 out of 5). Technical affinity, as measured with the ATI scale, was slightly lower in patients (score: mean 3.62, SD 1.35 out of 6) compared to physicians (score: mean 3.88, SD 1.03 out 6).

Conclusions: The quality ratings from physicians and patients were comparable and indicated mediocre app quality. Technical affinity, as assessed by using the ATI scale, was slightly lower for patients. Adequate apps for patients with chronic wounds remain limited, emphasizing the need for improved app development to meet patient needs. The ATI scale proved valuable for assessing technical affinity among different user groups.

JMIR Mhealth Uhealth 2024;12:e51592

doi:10.2196/51592

Keywords



Background

Aging societies in industrialized nations are experiencing an increasing prevalence of chronic wounds, resulting in growing challenges in patient care. The high costs of therapy and personnel often limit proper treatment. Chronic wounds significantly impact patients’ quality of life, requiring intensive therapy multiple times per week, along with regular medical checkups. Moreover, they have substantial economic implications, including hospitalizations, personnel costs (eg, wound care nurses, home health care services, physicians, wound managers, day clinics, etc), material costs, and management (eg, medical transportation). Purwins et al [1] identified hospitalizations, nursing staff, and material costs as the main contributors to the overall expenses.

Mobile health (mHealth) apps hold promise for bridging gaps in health care. However, a lack of evidence exists for the effectiveness of available mHealth apps [2], and high-quality trials are needed to examine their effects [3]. So far, only reviews of the use of mHealth apps without systematic searches and evaluations have been performed [4]. To address this gap in evidence for patients with chronic wounds, a systematic analysis is necessary to gather reliable data in this area.

A study by Svendsen et al [5] demonstrated that patient-centered smartphone apps can significantly improve treatment adherence in cohorts of patients with psoriasis and patients with rheumatic diseases. Participants expressed favorable views toward medical apps, indicating a willingness to use such apps if available.

To subjectively assess app quality, the Mobile App Rating Scale (MARS) [6] was developed, which evaluates engagement, functionality, aesthetics, and information. Additionally, the System Usability Scale (SUS) is a 10-item questionnaire for assessing the usability of a system. It has been effectively applied to evaluate mobile apps related to dementia, depression, pediatric obesity, and smoking cessation [7]. Moreover, the Affinity for Technology Interaction (ATI) scale [8] provides an easy and reliable means to quantify an individual’s technology affinity. An aim of this study was to conduct a focused analysis on the core functionalities and core features of the included apps; therefore, we excluded apps with advertisements in order to prevent disruptions in users’ experiences and advertisements’ effects on usability [9]. Advertisements can be confounding variables and make it difficult to compare apps’ performance or users’ experiences. Advertisements can frustrate or annoy users and reduce the clarity and intuitiveness of an app’s interface [10]. We also excluded paid apps to prevent bias toward users with financial means, maintain analysis inclusivity, and ensure that preferences align with intrinsic app quality rather than financial considerations, thereby enhancing the analysis’ fairness and validity.

Aim of This Study

The primary objectives of this study were to identify and evaluate publicly available smartphone apps designed for patients with chronic wounds. The assessment aimed to provide subjective quality ratings for these apps, while also collecting data on the technical affinity of this specific patient group. To date, a systematic review and assessment of smartphone apps tailored for patients with chronic wounds has not been conducted.


Ethical Considerations

This study was conducted in accordance with the Declaration of Helsinki, and ethical approval was waived by the local ethics committee of the University of Würzburg.

App Selection

A systematic search of the German Google Play Store and the Apple App Store was performed from April 2022 to May 2022. The search terms used were “wound,” “pressure ulcer,” “ulcus,” “Wunde,” “Wunddokumentation,” and “wound documentation.” Two independent reviewers searched each app store. The inclusion criteria were that apps had to (1) be available in both app stores, (2) be available in English or German, and (3) be specifically designed for patients. Apps that were not available free of charge and apps that contained advertisements were excluded.

The following information, when available in the app stores and on the associated app websites, was collected: app name, target group (eg, patients and medical personnel), cost, platform, advertisements, features, and search term used to identify the app.

Evaluation of App Quality

The MARS [11] was developed for professional raters to evaluate mobile apps, and it is a validated and reliable scale. The user version of the MARS (uMARS) [12] was designed for users to evaluate the quality of mHealth apps. Both scales are based on a 5-point Likert scale for the following four sections: “Engagement,” “Aesthetics,” “Functionality,” and “Information.” Additionally, there is a “Subjective” section. Studies using the MARS have already been performed for several chronic diseases and apps related to breast cancer. The uMARS [12] has been broadly applied to evaluate apps for rheumatic diseases, weight loss, nutrition tracking, and menstrual tracking. The SUS is a simple, 10-item attitude Likert scale that gives an overall view of subjective assessments of usability [13].

Prior to the evaluation, suitable apps were selected based on the inclusion and exclusion criteria, resulting in a total of 3 apps. The quality of these three apps was then evaluated by 10 physicians using the MARS and SUS. Before the assessment, the physicians first watched a short training video that explained the MARS and were then asked to use the apps for more than 10 minutes.

Finally, the best app—the one with the highest mean MARS score among the physicians—was evaluated by 11 patients with chronic wounds using the uMARS and SUS.

Evaluation of Technical Affinity

Many studies that evaluate the quality of mobile apps lack information on cohorts’ technical affinity, which is necessary to assess and interpret the results. Therefore, the ATI scale [8] was used to gather information on physicians’ and patients’ technical affinity. The ATI scale was designed to quantify a tendency to actively engage in intensive technology interaction or a tendency to avoid technology interaction. In both the patient group and the physician group, ATI scale scores were collected. A Pearson correlation was used to correlate the ages of patients and ATI scale scores.

Comparative Analysis of Patients’ and Physicians’ Data

The MARS results represented the physicians’ evaluations, as stated in the Evaluation of App Quality section, whereas uMARS results corresponded to the patients’ evaluations. As a next step, after normalization, the Mann-Whitney U test was used to analyze whether there was a significant difference between MARS and uMARS scores and between patients’ and physicians’ SUS and ATI scale median scores. A comparison of the five subcategories within the MARS was performed. P values of <.05 were considered significant. The data analysis was performed with SPSS 23 (IBM Corp).


App Selection

A total of 118 apps were identified—95 in the Apple App Store and 96 in the Google Play Store—of which 73 were available in both app stores (Figure 1). Of these 73 apps, 10 were specifically designed for patients. Of these 10 apps, 1 contained advertising, 4 were not free of charge, and 3 met both exclusion criteria, resulting in a total of 7 apps that were excluded from further analysis.

Figure 1. Flowchart illustrating the screening process for identifying suitable mobile apps.

Evaluation of App Quality

A total of 3 apps—WoundEducation, APD Skin Monitoring, and WUND APP—met all inclusion and exclusion criteria and were evaluated by 10 physicians. WoundEducation provides an overview of different wound types and treatment options. APD Skin Monitoring includes an automated area calculation function to quantify wound area. WUND APP offers advice and information on chronic wounds and contains a diary function to track patient-related outcomes, such as pain or the level of wound secretion.

The mean MARS scores (and SDs) of the physicians, including subcategory scores and SUS scores, are provided in Table 1.

Table 1. Evaluation of WUND APP, APD Skin Monitoring, and WoundEducation by 10 physicians using the Mobile App Rating Scale (MARS) and System Usability Scale (SUS).
Mobile app nameMARS score, mean (SD)Engagement score, mean (SD)Functionality score, mean (SD)Aesthetics score, mean (SD)Information score, mean (SD)Psychotherapy score, mean (SD)SUS score (%), mean (SD)
WUND APP3.88 (0.65)3.36 (0.89)4.38 (0.66)4.13 (0.76)3.67 (0.56)2.98 (0.36)80.5 (17.7)
WoundEducation3.01 (0.5)2.14 (0.44)3.93 (0.99)2.73 (0.75)3.25 (0.58)2.57 (0.46)72.75 (18.5)
APD Skin Monitoring2.64 (0.65)2.36 (0.56)2.65 (0.83)2.77 (0.69)2.79 (0.82)2.47 (0.63)50.75 (27)

WUND APP had the highest mean MARS score (mean 3.88, SD 0.65 out of 5) and was subsequently analyzed by 11 patients with chronic wounds. WUND APP had a similar mean uMARS score (mean 3.89, SD 0.4 out of 5) when analyzed by the patients.

Evaluation of Technical Affinity

The ATI scale scores ranged from 1.56 to 6 out of 6 for patients and from 2.11 to 5.67 out of 6 for physicians (Figure 2). An ATI scale score of >3 indicates average technology affinity, and a score of >4 indicates high technology affinity [8]. Mean ATI scale scores were slightly lower for patients (mean 3.62, SD 1.35 out of 6) than those for physicians (mean 3.88, SD 1.03 out of 6), but the difference was not statistically significant (P=.43). Patients ranged in age from 28 to 70 years; 7 were male, and 4 were female. Physicians ranged in age from 26 to 43 years; 4 were male, and 6 were female. The Pearson correlation coefficient between ATI scale scores and the ages of patients was –0.706 (P=.02), indicating a strong negative correlation.

Figure 2. Scatter plot depicting ATI scale scores in patients and physicians. ATI: Affinity for Technology Interaction.

Comparative Analysis of Patients’ and Physicians’ Data

MARS and uMARS scores were almost identical. The subcategory “Information” was rated worse by physicians, while the subcategory “Functionality” was rated worse by patients (Figure 3). For the subcategories “Aesthetics” and “Engagement,” no differences between physicians and patients were found. No statistically significant difference (Mann-Whitney U test) was found between MARS and uMARS scores (P=.76) or between patients’ and physicians’ SUS scores (P=.39; Multimedia Appendix 1).

Figure 3. Comparison of MARS and uMARS subscale scores (“Engagement,” “Functionality,” “Aesthetics,” and “Information” scores). Physicians (professional raters) used the MARS, and patients used the uMARS. MARS: Mobile App Rating Scale; uMARS: user version of the Mobile App Rating Scale.

Principal Findings

This study represents the first systematic examination that aimed to identify and assess smartphone apps specifically designed for patients with chronic wounds. The quality of these apps was evaluated by independent professional reviewers (physicians) and patients using well-validated scoring systems—the MARS, uMARS, and SUS. Using subjective measures in the assessment of eHealth is relevant and essential, and it is crucial to recognize this fact. Agreement among the assessors, even for subjective measures, is a good sign of the reliability of the assessment [14].

The overall findings indicate that the apps received moderate ratings. Among all apps evaluated, WUND APP achieved the highest mean MARS score (mean 3.88, SD 0.65) when assessed by physicians. Similarly, it received a mean uMARS score of 3.89 (SD 0.4) when evaluated by patients. Given that the preselection was exclusively conducted by physicians, patients’ assessments might have differed.

WUND APP includes a diary function for documenting wound photos and patient-related outcomes, such as pain or wound secretion. Furthermore, it has a reminder function for necessities such as physician appointments and dressing changes. Its user interface is clear, and the app is user-friendly. Consequently, physicians assigned high scores for the subcategories “Engagement” (mean 3.36, SD 0.89), “Functionality” (mean 4.38, SD 0.66), and “Aesthetics” (mean 4.13, SD 0.76). Additionally, the app provides information on the causes, diagnostics, and treatments of the different wound types, resulting in a mean MARS score of 3.67 (SD 0.56) for the subcategory “Information.”

WoundEducation presents information in a straightforward structure that resembles an article with embedded links, delivers medical information, and illustrates various wound types through examples. The app received high ratings for overall functionality (mean 3.93, SD 0.99) and information provision (mean 3.25, SD 0.58). However, it demonstrated limited interactivity, as reflected by having the lowest score for the “Engagement” subcategory (mean 2.14, SD 0.44). Additionally, its aesthetics (mean 2.73, SD 0.75) were rated the lowest among the three apps.

APD Skin Monitoring [15] uses a coin as a reference for calculating the area of a wound, proving particularly useful in assessing wound progression and healing status regardless of whether the wounds are irregularly shaped. In addition, the coloration of wounds can be analyzed and tracked over time. However, the app received the lowest ratings for the subcategories “Functionality” (mean 2.65, SD 0.83) and “Engagement” (mean 2.36, SD 0.56).

Our findings revealed that a limited number of wound apps were specifically tailored for patients, accounting for only 14% (10/73) of all wound apps that were available in both app stores. Overall, both physicians and patients rated WUND APP similarly; however, although not statistically significant, it is noteworthy that differences emerged in 2 subcategories (“Information”: P=.97; “Functionality”: P=.56). Physicians rated the information content lower, which could be attributed to their expert medical knowledge. As professionals, they may have had higher expectations regarding the app’s information content and may have found certain aspects less informative. In a recent study, it was shown that patients trusted recommendations and reviews from medical organizations and health care professionals when selecting apps [16]; their motivation to continue using apps was driven by features that supported goal setting and tracking, data sharing, decision-making, and empowerment.

On the other hand, patients rated the functionality of WUND APP lower, which was possibly due to their limited involvement in the app’s development process. A systemic review reported that health care professionals were engaged in the development process for only 35% of the 7 analyzed apps that were specifically designed for patients with rheumatoid arthritis [17]. Another systematic review revealed that patients were only engaged in the development process for 15% of the 32 analyzed apps that were designed for individuals with rheumatic and musculoskeletal diseases [18]. The inclusion of patients in the development process of future wound apps could help to ensure that the apps meet their specific needs and preferences. Our study highlights the scarcity of wound apps designed explicitly for patients and the importance of involving patients in the app development process. Tailoring apps to meet patients’ specific requirements and involving them in the design process would likely result in improved app functionality and overall user satisfaction.

To the best of our knowledge, this study represents the first systematic study to collect data on the technical affinity of patients with wounds. Surprisingly, no statistically significant difference in technical affinity was observed between wound care patients and physicians, even with the inclusion of patients aged up to 70 years (P=.43). Nevertheless, future studies and app development projects should aim to include older patients and comprehensively assess and address their specific needs. The low adoption and use of mHealth apps among older patients are frequently attributed to inadequate designs [19].

In 2019, Germany introduced a digital health app (DiGA) directory that includes scientifically validated apps. Physicians can prescribe DiGAs in a manner similar to how they prescribe medications [20]. However, to date, no DiGAs specifically tailored for patients with chronic wounds are available in the directory. It is worth noting that physicians with a higher technical affinity and those who are female hold significantly more positive attitudes toward DiGAs [21].

The absence of wound care apps highlights the unmet potential for innovative digital solutions to address the needs of patients with chronic wounds. The inclusion of validated and effective wound care apps in the DiGA directory could significantly improve patient outcomes and health care management in this specific area. It is imperative for future app development initiatives to focus on developing and validating apps that cater to the unique requirements of patients with chronic wounds, to provide them with accessible and effective digital health care resources.

Multiple attempts have been made to integrate mHealth apps into wound care [22], including apps for wound care measurements [23], wound care dressing decision support systems [24], and home-based self-management systems [25]. An Australian study assessed an artificial intelligence app for wound assessment, involving 166 patients in the standard group and 124 in the intervention group. The intervention group demonstrated significantly improved wound documentation, along with positive outcomes such as enhanced patient adherence, efficient digital care provision, and substantial reductions in wound size [26].

A cultural shift toward greater technology affinity has been accelerated by the COVID-19 pandemic [27]. Patients are increasingly using mobile apps when they perceive clear benefits to their use, such as reducing social contact during the COVID-19 pandemic. For patients with chronic wounds, these benefits may include using an app as a diary, while for others, apps may offer the advantage of saving time and transportation costs through telemedicine services. Telemedical approaches could significantly alter and improve the current wound care landscape, bearing the potential to improve the efficiency, accuracy, and accessibility of both diagnoses and treatments. Telemedicine in chronic wound management was shown to be noninferior to conventional standard care in a systematic review and meta-analysis [28]. The earlier diagnosis of complications, such as wound infections, can reduce the use of antibiotics and lower health care costs by preventing hospital stays.

Limitations

Due to the limited number of apps that met our strict inclusion and exclusion criteria, only a small subset could be analyzed in this study. The selection of apps was conducted within a relatively short time frame, which might have further restricted the available options for evaluation.

The nonsignificant differences between physicians’ and patients’ ratings in this study may be attributed to the small sample of only 11 patients and 10 physicians, which resulted in limited statistical power. Therefore, we emphasize the importance of cautious interpretation and the consideration of larger sample sizes for future research.

Another factor that influenced this study’s results was the inclusion of patients who agreed to evaluate WUND APP. This approach potentially introduced selection bias, as our patient sample may not represent the true technical affinity of all patients with chronic wounds. The overall technical affinity of a larger and more diverse patient population might be lower than what was observed in this study. Future studies with larger patient cohorts might provide a more comprehensive understanding of the technical affinity and usability experience of patients with chronic wounds who use mHealth apps. Further, the assessment relied on a qualitative survey—a method that may be susceptible to various biases. However, to truly ascertain the effectiveness of WUND APP, a randomized controlled trial is needed.

Conclusions

Patient involvement is crucial in app development. By involving all stakeholders, including physicians, wound care experts, and patients, throughout the development process, apps can be tailored to meet the specific needs and preferences of the end users, resulting in increased user satisfaction and improved health outcomes.

The validated ATI scale proved to be a valuable tool for evaluating an individual’s technical affinity. In future studies and app evaluations, technical affinity should be determined to generalize outcomes to specific patient and consumer cohorts.

Acknowledgments

TD is supported by the Clinician Scientist Program (project number: Z-2/CSP-25) of the Interdisciplinary Center of Clinical Research (IZKF) of the Medical Faculty, University of Würzburg.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Patient and physician characteristics (age and sex) and the Affinity for Technology Interaction scale, System Usability Scale, Mobile App Rating Scale (MARS), and user version of the MARS scores (including subcategory scores) of all 11 patients and 10 physicians who participated.

DOCX File, 19 KB

  1. Purwins S, Herberger K, Debus ES, et al. Cost-of-illness of chronic leg ulcers in Germany. Int Wound J. Apr 2010;7(2):97-102. [CrossRef] [Medline]
  2. Buijink AWG, Visser BJ, Marshall L. Medical apps for smartphones: lack of evidence undermines quality and safety. Evid Based Med. Jun 2013;18(3):90-92. [CrossRef] [Medline]
  3. Zhao J, Freeman B, Li M. Can mobile phone apps influence people's health behavior change? an evidence review. J Med Internet Res. Oct 31, 2016;18(11):e287. [CrossRef] [Medline]
  4. Koepp J, Baron MV, Martins PRH, et al. The quality of mobile apps used for the identification of pressure ulcers in adults: systematic survey and review of apps in app stores. JMIR Mhealth Uhealth. Jun 16, 2020;8(6):e14266. [CrossRef] [Medline]
  5. Svendsen MT, Andersen F, Andersen KH, et al. A smartphone application supporting patients with psoriasis improves adherence to topical treatment: a randomized controlled trial. Br J Dermatol. Nov 2018;179(5):1062-1071. [CrossRef] [Medline]
  6. Messner EM, Terhorst Y, Barke A, et al. The German version of the Mobile App Rating Scale (MARS-G): development and validation study. JMIR Mhealth Uhealth. Mar 27, 2020;8(3):e14479. [CrossRef] [Medline]
  7. Hyzy M, Bond R, Mulvenna M, et al. System Usability Scale benchmarking for digital health apps: meta-analysis. JMIR Mhealth Uhealth. Aug 18, 2022;10(8):e37290. [CrossRef] [Medline]
  8. Franke T, Attig C, Wessel D. A personal resource for technology interaction: development and validation of the Affinity for Technology Interaction (ATI) scale. Int J Hum Comput Interact. 2019;35(6):456-467. [CrossRef]
  9. Haggag O, Grundy J, Abdelrazek M, Haggag S. A large scale analysis of mHealth app user reviews. Empir Softw Eng. 2022;27(7):196. [CrossRef] [Medline]
  10. Lobo EH, Abdelrazek M, Frølich A, et al. Detecting user experience issues from mHealth apps that support stroke caregiver needs: an analysis of user reviews. Front Public Health. May 25, 2023;11:1027667. [CrossRef] [Medline]
  11. Stoyanov SR, Hides L, Kavanagh DJ, Zelenko O, Tjondronegoro D, Mani M. Mobile App Rating Scale: a new tool for assessing the quality of health mobile apps. JMIR Mhealth Uhealth. Mar 11, 2015;3(1):e27. [CrossRef] [Medline]
  12. Stoyanov SR, Hides L, Kavanagh DJ, Wilson H. Development and validation of the user version of the Mobile Application Rating Scale (uMARS). JMIR Mhealth Uhealth. Jun 10, 2016;4(2):e72. [CrossRef] [Medline]
  13. Brooke J. SUS: a quick and dirty usability scale. In: Jordan PW, Thomas, B, McClelland IL, Weerdmeester B, editors. Usability Evaluation In Industry. CRC Press; 1996. [CrossRef]
  14. Powell R, Brown K, Davies M, et al. The value of the modern vascular surgeon to the health care system: a report from the Society for Vascular Surgery Valuation Work Group. J Vasc Surg. Feb 2021;73(2):359-371.e3. [CrossRef] [Medline]
  15. Wu W, Yong KYW, Federico MAJ, Gan SKE. The APD Skin Monitoring app for wound monitoring: image processing, area plot, and colour histogram. Scientific Phone Apps and Mobile Devices. May 28, 2019. [CrossRef]
  16. Jezrawi R, Balakumar S, Masud R, et al. Patient and physician perspectives on the use and outcome measures of mHealth apps: exploratory survey and focus group study. Digit Health. May 24, 2022;8:20552076221102773. [CrossRef] [Medline]
  17. Luo D, Wang P, Lu F, Elias J, Sparks JA, Lee YC. Mobile apps for individuals with rheumatoid arthritis: a systematic review. J Clin Rheumatol. Apr 2019;25(3):133-141. [CrossRef] [Medline]
  18. Najm A, Gossec L, Weill C, Benoist D, Berenbaum F, Nikiphorou E. Mobile health apps for self-management of rheumatic and musculoskeletal diseases: systematic literature review. JMIR Mhealth Uhealth. Nov 26, 2019;7(11):e14730. [CrossRef] [Medline]
  19. Liu N, Yin J, Tan SSL, Ngiam KY, Teo HH. Mobile health applications for older adults: a systematic review of interface and persuasive feature design. J Am Med Inform Assoc. Oct 12, 2021;28(11):2483-2501. [CrossRef] [Medline]
  20. Sauermann S, Herzberg J, Burkert S, Habetha S. DiGA - a chance for the German healthcare system. J Eur CME. Dec 23, 2021;11(1):2014047. [CrossRef] [Medline]
  21. Dahlhausen F, Zinner M, Bieske L, Ehlers JP, Boehme P, Fehring L. Physicians’ attitudes toward prescribable mHealth apps and implications for adoption in Germany: mixed methods study. JMIR Mhealth Uhealth. Nov 23, 2021;9(11):e33012. [CrossRef] [Medline]
  22. Friesen MR, Hamel C, McLeod RD. A mHealth application for chronic wound care: findings of a user trial. Int J Environ Res Public Health. Nov 19, 2013;10(11):6199-6214. [CrossRef] [Medline]
  23. Biagioni RB, Carvalho BV, Manzioni R, Matielo MF, Neto FCB, Sacilotto R. Smartphone application for wound area measurement in clinical practice. J Vasc Surg Cases Innov Tech. Feb 26, 2021;7(2):258-261. [CrossRef] [Medline]
  24. Jordan S, McSwiggan J, Parker J, Halas GA, Friesen M. An mHealth app for decision-making support in wound dressing selection (wounds): protocol for a user-centered feasibility study. JMIR Res Protoc. Apr 24, 2018;7(4):e108. [CrossRef] [Medline]
  25. Zhang J, Mihai C, Tüshaus L, Scebba G, Distler O, Karlen W. Wound image quality from a mobile health tool for home-based chronic wound management with real-time quality feedback: randomized feasibility study. JMIR Mhealth Uhealth. Jul 30, 2021;9(7):e26149. [CrossRef] [Medline]
  26. Barakat-Johnson M, Jones A, Burger M, et al. Reshaping wound care: evaluation of an artificial intelligence app to improve wound assessment and management amid the COVID-19 pandemic. Int Wound J. Oct 2022;19(6):1561-1577. [CrossRef] [Medline]
  27. Rettinger L, Klupper C, Werner F, Putz P. Changing attitudes towards teletherapy in Austrian therapists during the COVID-19 pandemic. J Telemed Telecare. Jun 2023;29(5):406-414. [CrossRef] [Medline]
  28. Chen L, Cheng L, Gao W, Chen D, Wang C, Ran X. Telemedicine in chronic wound management: systematic review and meta-analysis. JMIR Mhealth Uhealth. Jun 25, 2020;8(6):e15574. [CrossRef] [Medline]


ATI: Affinity for Technology Interaction
DiGA: digital health app
MARS: Mobile App Rating Scale
mHealth: mobile health
SUS: System Usability Scale
uMARS: user version of the Mobile App Rating Scale


Edited by Lorraine Buis; submitted 04.08.23; peer-reviewed by Christine Jacob, Mo Zhou; final revised version received 17.12.23; accepted 25.01.24; published 21.03.24.

Copyright

© Tassilo Dege, Bernadette Glatzel, Vanessa Borst, Franziska Grän, Simon Goller, Caroline Glatzel, Matthias Goebeler, Astrid Schmieder. Originally published in JMIR mHealth and uHealth (https://mhealth.jmir.org), 21.3.2024.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mHealth and uHealth, is properly cited. The complete bibliographic information, a link to the original publication on https://mhealth.jmir.org/, as well as this copyright and license information must be included.