Published on in Vol 9, No 8 (2021): August

Preprints (earlier versions) of this paper are available at, first published .
Evaluating the Validity and Utility of Wearable Technology for Continuously Monitoring Patients in a Hospital Setting: Systematic Review

Evaluating the Validity and Utility of Wearable Technology for Continuously Monitoring Patients in a Hospital Setting: Systematic Review

Evaluating the Validity and Utility of Wearable Technology for Continuously Monitoring Patients in a Hospital Setting: Systematic Review


1Faculty of Medicine, University of Toronto, Toronto, ON, Canada

2Library and Information Services, University Health Network, Toronto, ON, Canada

3Division of General Internal Medicine, University Health Network, Toronto, ON, Canada

Corresponding Author:

Vikas Patel, BMSc, MD

Faculty of Medicine

University of Toronto

1 King's College Cir

Toronto, ON, M5S 1A8


Phone: 1 4169756585


Background: The term posthospital syndrome has been used to describe the condition in which older patients are transiently frail after hospitalization and have a high chance of readmission. Since low activity and poor sleep during hospital stay may contribute to posthospital syndrome, the continuous monitoring of such parameters by using affordable wearables may help to reduce the prevalence of this syndrome. Although there have been systematic reviews of wearables for physical activity monitoring in hospital settings, there are limited data on the use of wearables for measuring other health variables in hospitalized patients.

Objective: This systematic review aimed to evaluate the validity and utility of wearable devices for monitoring hospitalized patients.

Methods: This review involved a comprehensive search of 7 databases and included articles that met the following criteria: inpatients must be aged >18 years, the wearable devices studied in the articles must be used to continuously monitor patients, and wearables should monitor biomarkers other than solely physical activity (ie, heart rate, respiratory rate, blood pressure, etc). Only English-language studies were included. From each study, we extracted basic demographic information along with the characteristics of the intervention. We assessed the risk of bias for studies that validated their wearable readings by using a modification of the Consensus-Based Standards for the Selection of Health Status Measurement Instruments.

Results: Of the 2012 articles that were screened, 14 studies met the selection criteria. All included articles were observational in design. In total, 9 different commercial wearables for various body locations were examined in this review. The devices collectively measured 7 different health parameters across all studies (heart rate, sleep duration, respiratory rate, oxygen saturation, skin temperature, blood pressure, and fall risk). Only 6 studies validated their results against a reference device or standard. There was a considerable risk of bias in these studies due to the low number of patients in most of the studies (4/6, 67%). Many studies that validated their results found that certain variables were inaccurate and had wide limits of agreement. Heart rate and sleep were the parameters with the most evidence for being valid for in-hospital monitoring. Overall, the mean patient completion rate across all 14 studies was >90%.

Conclusions: The included studies suggested that wearable devices show promise for monitoring the heart rate and sleep of patients in hospitals. Many devices were not validated in inpatient settings, and the readings from most of the devices that were validated in such settings had wide limits of agreement when compared to gold standards. Even some medical-grade devices were found to perform poorly in inpatient settings. Further research is needed to determine the accuracy of hospitalized patients’ digital biomarker readings and eventually determine whether these wearable devices improve health outcomes.

JMIR Mhealth Uhealth 2021;9(8):e17411




Most physiologic parameters, such as vital signs or activity, are routinely monitored a few times each day in hospital ward settings [1]. Some parameters, such as sleep, are not routinely monitored at all [2,3]. More frequent monitoring could allow for the timely identification of the deteriorating health of patients and spur efforts for improving patients’ overall health through increased sleep and activity. Since subtle changes in vital signs are often present 8 to 24 hours before a life-threatening event, such as intensive care unit admission or cardiac arrest, vital sign surveillance has the potential to detect clinical deterioration at an earlier phase, thereby permitting clinicians to make corrective interventions [4-7]. This includes identifying patients with poorly controlled pain and recognizing arrhythmias. The term posthospital syndrome has been used to denote the deleterious effects of acute illnesses that are compounded with poor sleep and low activity and occur during hospital stay [8]. Measuring sleep and activity could improve the recognition of such issues and encourage health providers to introduce interventions that improve patients’ experiences in hospitals by encouraging mobilization and to identify targets for sleep-promoting interventions [9-11]. In addition, access to other digital biomarkers (eg, heart rate, blood pressure, oxygen saturation, etc) would allow clinicians to determine underlying etiologies and make tailored interventions.

The rapid uptake of affordable wearables, such as fitness bands, may provide a method for continuously measuring sleep; activity; and vital signs, such as heart rate [12-15]. However, existing literature that describes wearable devices is mostly limited to ambulatory settings and focuses on the management of chronic diseases [16,17]. More inpatient data are needed on both the validity of wearables and patient adherence. Although wearable testing has been conducted with healthy volunteers, it will be important to validate these signals in inpatient settings, where algorithms for processing sensor data into digital signals, such as those for sleep, heart rate, and activity, may be less accurate [18]. Despite the proposed benefit of intensive monitoring, many wearable studies have found issues with patient adherence [18-20]. Adherence is a crucial barrier to acquiring data and can be influenced by device convenience, the comfort of use, and interaction requirements [19]. Studies of wearable devices worn by hospitalized inpatients have been limited by large dropout rates [20].

Although there have been systematic reviews of the monitoring of patients’ physical activity in hospitals [21-23], there are no reviews of the use of wearables that can reliably measure other health parameters. Therefore, in this review, we aimed to expand our search by including articles that used wearables to assess parameters other than physical activity and to assess the adherence of patients in inpatient settings.


For the purposes of this review, a wearable was considered to be any electronic device that has at least 1 sensor and can be worn on the body [24]. Wearables were examined for their ability to measure digital biomarkers, which are defined as digitally collected physiological and behavioral measures (eg, heart rate, average sleep duration, and daily step count) that explain, influence, or predict health-related outcomes [18]. Consistent with previous research, patient adherence was objectively assessed by reporting the mean proportion of patients who completed a given study [25]. The primary objectives of this review were to determine patients’ adherence to using wearable devices in hospitals and to examine the validity of wearable-derived biomarker readings.

Identification and Selection of Studies

A comprehensive search strategy was developed to identify articles on the three main concepts of our question—wearables, monitoring, and inpatients. The initial search strategy was developed for Ovid MEDLINE by using a combination of database-specific subject headings and text words (Multimedia Appendix 1). Additional key words were generated based on input from the subject specialists on the team, and the revised search strategy was customized for each database.

Searches of the following databases were executed on August 16, 2018: Ovid MEDLINE, Ovid MEDLINE Epub Ahead of Print and In-Process & Other Non-Indexed Citations, Cochrane Database of Systematic Reviews, Cochrane Central Register of Controlled Trials, Health Technology Assessment database (Ovid), and CINAHL with Full Text. The search in Ovid Embase was not executed until September 5, 2018, due to issues with the vendor’s August database reload. Additional search methods included reviewing the cited references of eligible studies via Web of Science (May 6, 2019) and the reference lists of eligible studies. There were no restrictions on publication period. Limits were imposed to ensure that only English-language studies and those with adult populations were included in this review. No other limits were applied to the literature search.

Article Selection and Exclusion Criteria

Records were screened by two reviewers (VP and RW) independently. For selected studies, full-text articles were obtained and evaluated for eligibility [26]. The eligibility criteria for inclusion in this review were as follows:

  • Medical or surgical inpatients aged >18 years
  • Device studied in the article must be a wearable (such as a watch, vest, pendant, jewelry, headset, and wristband)
  • Articles must describe an element of continuous monitoring for at least 24 hours or greater
  • Articles must describe the measurement of 1 or more digital biomarkers other than just physical activity or standard hospital telemetry for heart rate recording.

We excluded articles that were not considered original research, such as letters to the editor, comments, and reviews. We also excluded articles that monitored less than 3 patients, described the monitoring of a very specialized system in the body (eg, insole devices, ventricular assistive devices, and cochlear implants), involved the monitoring of patients in rehabilitation hospitals, or used wearables as tools for therapy (eg, insulin delivery).

Data Extraction

Two reviewers (RW and VP) independently extracted the data and resolved any disagreements by discussing the findings and making a collective decision. The data extracted for each article included the year of publication, study setting and design, number of participants, gender ratio, mean age of participants, digital biomarkers measured in the study, average and maximum duration that the wearable was worn by participants in each study, and patient completion rate (the proportion of patients that wore the wearable for the minimum monitoring duration that was set by the study authors). For studies that used a reference standard, any participants who were missing data from the wearable or the standard were determined to be incomplete measurement pairs and were omitted from the final count of patients who completed the study. Furthermore, we extracted the types of wearables that were worn by the participants in each study along with the placement sites on the body. Devices were classified as medical grade (approved or cleared by the US Food and Drug Administration), research grade (typically used in research settings only), and consumer grade (used by general consumers).

Validation data were also collected for each article by assessing whether the authors compared the accuracy of their digital readings to a reference standard. To determine the validity of measures that were compared to a reference standard, correlation coefficients, mean differences, and limits of agreement were extracted from each study.

Risk of Bias Assessment

All articles that assessed for validated readings were independently assessed for their risk of bias by two independent reviewers (VP and RW) using a modification of the validation subscale from a checklist for assessing the methodological quality of studies on the measurement properties of health status measurement instruments (Consensus-Based Standards for the Selection of Health Status Measurement Instruments [COSMIN]) [27] (Table 1). All discrepancies were resolved by discussion and consensus. The quality evaluation included 5 study design and methodology components (the percentage of missing data, missing data management, adequate sample size, acceptable criterion comparison, and design or methodological flaws) and 1 analysis component (acceptable accuracy analyses). We rated the quality of each dimension as excellent, good, fair, or poor based on a priori modifications to the COSMIN validation subscale for scoring criteria that are appropriate for accuracy studies (Multimedia Appendix 2) [28].

Table 1. Risk of bias assessment for studies that validated their wearable readings.
StudyAssessment criterion

Mean or % differenceCorrelationLOAaPercentage of missing dataMissing data managementAdequate sample size (patients)Adequate sample size (measurements)Acceptable reference comparisonOther methodological flawsAcceptable accuracy analyses
Bloch et al [29]NoNoNoExcellentExcellentPoorPoorExcellentNoPoor
Breteler et al [30]YesNoYesExcellentExcellentPoorExcellentExcellentNoExcellent
Gallo and Lee [13]NoYesNoExcellentExcellentFairFairFairNoExcellent
Kroll et al [11,31]YesYesYesExcellentExcellentGoodExcellentExcellentNoExcellent
Steinhubl et al [32]NoYesNoExcellentExcellentPoorExcellentExcellentNoExcellent
Weenk et al [4]YesNoYesExcellentExcellentPoorGoodExcellentNoExcellent

aLOA: limits of agreement.

Characteristics of Included Studies

Our literature search identified 2754 article citations. After excluding duplicate records, 2012 records were deemed eligible for screening. A total of 83 studies were selected based on abstracts and underwent full-text review. After applying our inclusion and exclusion criteria, 15 articles that described 14 studies were selected for this review (Figure 1).

Figure 1. PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) flow diagram of included and excluded studies.
View this figure

All of the articles included were prospective cohort studies (Table 2) [4,11,13,14,20,29-38]. Overall, 9 different types of commercial wearables were described across the 14 studies, and 7 different health variables were assessed collectively by the 14 studies (Table 3). The wearable devices that were described by the studies came in various different forms and were attached to a range of sites on the body (Figure 2). A total of 13 articles included both men and women as the study participants; the other two papers assessed sleep changes in postpartum women [13,34]. Kroll et al [11,31] published two articles from the same study. Both articles analyzed different aspects of the continuous monitoring of inpatients (ie, they used the same cohort of patients) but were included as the same study entry in this review (Table 2).

Collectively, the mean patient completion rate across all 14 studies was over 90%. Of the 8 articles that included a qualitative analysis as a part of their methodology, 7 reported that wearables were well received by either or both patients and clinicians.

Of the 14 studies, 6 validated wearable measurements against another standard device or measure (Table 3). The studies conducted by Bloch et al [29], Gallo and Lee [13], and Steinhubl et al [32] used intermittent measurements (nurse or questionnaires) for their reference standard. Further, Breteler et al [30] used a continuous reference (continuous electrocardiography and impedance pneumography) to compare the wearable readings for heart rates and respiratory rates [30]. Weenk et al [4] and Kroll et al [11,31] validated their wearable readings against both intermittent and continuous reference measurements. Of the 9 wearables included in the studies, 6 were cleared or approved by the US Food and Drug Administration as medical devices (ViSi Mobile [Sotera Wireless], Hidalgo EQ02 [Equivital], wrist actigraphy [Ambulatory Monitoring Inc; Actigraph LLC], LifeTouch [Isansys Lifecare], Zephyr Biopatch [Medtronic], and HealthPatch [VitalConnect]).

Table 2. Summary of included studies.
StudyYear publishedSetting (ward)MethodologyPatients, NMale %: Female % ratio, mean age (years)Variables measuredNumber of days device was worn, average (maximum)Patient completion rate, %
Lee and Lee [34]2007ObstetricProspective cohort21Females only, 32Sleep2a100
Gallo and Lee [13]2008ObstetricProspective cohort39Females only, 29Sleep2 (2)100
Bloch et al [29]2011GeriatricProspective cohort10Males and femalesb, 83Falls21a90
Chiu et al [33]2013NeurosurgeryProspective cohort6065:35, 35Sleep7a87
Watkins et al [37]2015Medicine and surgicalProspective cohort236Males and femalesb,cHRd, RRe, SpO2f, and BPg3 (3)100

Jeffs et al [20]2016MedicineProspective cohort20872:28cHR. RR, SpO2, temperature, and accelerometry(14)h32
Steinhubl et al [32]2016MedicineProspective cohort2665:35, 33HR, RR, and temperature3 (3)100
Razjouyan et al [35]2017Hematology and
Prospective cohort3545:55, 55HR and fall risk1a94
Weenk et al [4]2017General internal medicine and surgicalProspective cohort2065:35, 50HR, RR, BP, SpO2, and temperature2.5 (3)100
Kroll et al [11,31]2017Intensive care unitProspective cohort5052:48, 64HR, sleep1a96
Weller et al [36]2017Neurology and
Prospective cohort73654:46cHR, RR, SpO2, and BP1.7 (9)100

Breteler et al [30]2018SurgicalProspective cohort3372:28, 63HR and RR2.6 (3)76
Yang et al [14]2018OncologyProspective cohort1164:36cSleep16a91
Duus et al [38]2018General
Prospective cohort5058:42, 71HR, RR, and SpO23.1 (4)100

aThe maximum number of days was not reported in the study.

bThe study included both male and female participants but did not report a ratio.

cMean age was not reported in the study.

dHR: heart rate.

eRR: respiratory rate.

fSpO2: oxygen saturation

gBP: blood pressure.

hThe average number of days was not reported in the study.

Table 3. Distribution of the health variables that were assessed for accuracy in each study.
StudyDevice characteristicsDigital biomarkers

Device, manufacturerFDAa clearance or approvalHeart rateSleepRespiratory rateSpO2bSkin temperatureBlood pressureFall risk
Gallo and Lee [13]Wrist Actigraph, Ambulatory Monitoring IncYescR=0.53
Lee and Lee [34]Mini-MotionloggerActigraphy, Ambulatory Monitoring IncYesNot validated
Chiu et al [33]ActiGraph GT1M, Actigraph LLCYesNot validated
Yang et al [14]Actigraph GT3X+ watch, Actigraph LLCYesNot validated
Kroll et al [11,31]Fitbit Charge HR, Fitbit IncLoAd (sinus): 23.9 to 21.9 beats per minuteR=0.33
Breteler et al [30]HealthPatch, VitalConnectYesLoA: −8.8 to 6.5 beats per minuteLoA: −15.8 to 11.2 breaths per minuteNot validatedNot validated
Jeffs et al [20]Hidalgo EQ02, EquivitalYesNot validatedNot validatedNot validatedNot validated
Duus et al [38]LifeTouch, Isansys LifecareYesNot validatedNot validatedNot validated
Steinhubl et al [32]eMultiSense patch, Rhythm Diagnostic SystemsR=0.75R=0.83R=0.99
Bloch et al [29]Vigi’Fall, Vigilio TelemedicalSensitivity: 37.5%
Weenk et al [4]ViSi Mobile, Sotera WirelessYesLoA: −11.1 to 10.7 beats per minute−5.5 to 7.9 breaths per minute−3.1% to 3.3%Not validatedSBPf: −23 to 24 mm Hg; DBPg: 27.5 to 11.5 mm Hg
Weenk et al [4]HealthPatch, VitalConnectYes−12.6 to 9.5 beats per minute−10.3 to 9.0 breaths per minuteNot validatedNot validatedNot validated
Weller et al [36]ViSi Mobile, Sotera WirelessYesNot validatedNot validatedNot validatedNot validatedNot validated
Watkins et al [37]ViSi Mobile, Sotera WirelessYesNot validatedNot validatedNot validatedNot validated
Razjouyan et al [35]Zephyr BioPatch, MedtronicYesNot validatedNot validated

aFDA: US Food and Drug Administration.

bSpO2: oxygen saturation.

cNot available.

dLOA: limits of agreement.

eSteinhubl et al [32] did not report limits of agreement.

fSBP: systolic blood pressure.

gDBP: diastolic blood pressure.

Figure 2. Illustration of the types of and body locations for used wearable devices.
View this figure

Risk of Bias

Of the 6 studies in the risk of bias assessment, 4 were ranked as poor due to a small sample size (participants: N<30). The study conducted by Gallo and Lee [13] used sleep questionnaires as a reference measure and therefore received a fair rating for the “acceptability of reference” criterion, whereas the other five studies were ranked as excellent (ie, they used intermittent nurse readings or other validated methodologies). Further, in terms of assessing the accuracy analyses, only the study conducted by Bloch et al [29] did not report mean differences, correlations, and limits of agreement.

Validation by Digital Biomarker

Heart Rate

A total of 5 studies assessed heart rate accuracy. Breteler et al [30] found that the bias and 95% limits of agreement for heart rate were −1.1 beats per minute (BPM) and −8.8 to 6.5 BPM, respectively, for 55,565 heart rate pairs [30]. Specifically, the wearable sensor accurately detected tachycardia with a sensitivity of 90% and a specificity of 97% [30]. In a cohort of intensive care unit patients, Kroll et al [11,31] found that the Fitbit (Fitbit Inc)-derived heart rate values were slightly lower than those derived from continuous electrocardiography monitoring but that 73% of the readings were within 5 BPM of the electrocardiogram value (average bias: −1.14 BPM; R=0.74; P<.001; heart rate pairs: n=12,358) [11]. Overall, the limit of agreement for the Fitbit device was 24 BPM, but its performance was significantly better in patients in sinus rhythm than in those who were not in sinus rhythm (average bias: −0.99 BPM vs −5.02 BPM, respectively; P=.02; limits of agreement: 22.9 BPM vs 46.4 BMP, respectively; P=.049) [11]. Kroll et al [11,31] also found that the Fitbit was very specific when it detected tachycardia (sensitivity=70%; specificity=99%) [31]. Steinhubl et al [32] demonstrated that manual and automated heart rate readings correlated well (R=0.75; measurements: n=111), but limits of agreement were not reported [32]. Weenk et al [4] reported that heart rate readings were generally consistent when compared to the nurse recordings; the limits of agreement for the ViSi Mobile and the HealthPatch were −11.1 to 10.7 BPM and −12.6 to 9.5 BPM, respectively (86 measurements).


A total of 6 studies used wearables to assess sleep, of which 2 assessed whether wearable readings were reliable. Gallo and Lee [13] found that self-reported sleep correlated with the actigraphy-recorded number of awakenings (R=0.53; P=.01) [13]. Kroll et al [11,31] found that there was a moderate correlation between wearable-derived sleep duration and questionnaire-derived sleep quality (R=0.33; P=.03) [31].

Respiratory Rate

Of the 8 articles that used different wearables to measure the respiratory rate of patients, 3 assessed the wearables’ accuracy. Breteler et al [30] found that for respiratory rate, the bias was −2.3 breaths per minute, and wide limits of agreement were reported (−15.8 to 11.2 breaths per minute; measurement pairs: n=56,674) [30]. Steinhubl et al [32] reported that there was a strong correlation between wearable and manual respiratory rate readings (R=0.83; P<.001; measurements: n=111), but limits of agreement were not reported [32]. Weenk et al [4] described wide limits of agreement for respiratory rate based on 86 measurements (ViSi Mobile limits of agreement: −5.5 to 7.9 breaths per minute; HealthPatch limits of agreement: −10.3 to 9.0 breaths per minute) [4].

Other Measures

Only 1 study, which was conducted by Weenk et al [4], assessed the accuracy of oxygen saturation and blood pressure readings from ViSi Mobile by comparing them to HealthPatch readings as well as intermittent nurse measurements. From 86 measurements, they found that the automated readings for the systolic blood pressure, diastolic blood pressure, and oxygen saturation had wide limits of agreement (systolic blood pressure: −23.1 to 24.0 mm Hg; diastolic blood pressure: −27.5 to 11.5 mm Hg; oxygen saturation: −3.1% to 3.3%) [4]. Of the 6 articles that used wearables that measured skin temperature, only Steinhubl et al [32] validated the results against a reference standard to conclude that the automated readings were reliable (R=0.99; n=112), but bias and limits of agreement were not reported [32]. Of the 3 articles in this review that detected falls by using wearables, only Bloch and colleagues [29] assessed accuracy and found that the Vigi’Fall system had a low sensitivity (37.5%) to fall risk [29].

Principal Findings

We conducted a systematic review that evaluated the utility of wearable technology in continuously monitoring hospitalized patients for a wide variety of health parameters. Our review focused on the breadth of devices used and the signals measured in hospitalized patients and included consumer, research, and medical-grade devices. There was evidence to support the use of Fitbit, ViSi Mobile, and the HealthPatch to measure heart rate [4,11,31], since the readings were validated against both intermittent and continuous reference standards. This review demonstrated that the validity of the data did not necessarily correlate with the classification of the device because even some medical-grade devices did not perform well and yielded data with wide limits of agreement. We found that only 6 studies validated the accuracy of wearable-derived health data from hospitalized patients by comparing the readings against a reference standard. Overall, the quality of most of these studies was excellent in terms of the reporting of missing data (6/6, 100%) and the use of acceptable accuracy evaluations (5/6, 83%). However, there was a considerable risk of bias in these studies due to the low number of participants in most of the studies (4/6, 67%). Many studies reported wide limits of agreement for other digital biomarkers, such as respiratory rate and blood pressure. Of note, we also found that the majority of studies (8/14, 57%) did not validate the studied device or parameter measured.

Of the various health parameters, the best evidence of validity was in the monitoring of heart rate in hospitalized patients. We also found that, in hospital settings, limits of agreement for medical-grade devices ranged from 16.4 to 21.8 BPM, whereas the limit for a Fitbit consumer device that uses photoplethysmography signals was 24 BPM. Further, during Fitbit-based continuous electrocardiogram monitoring, 73% of the readings were within 5 BPM of electrocardiogram readings. In a systematic review of 158 studies that measured heart rate by using consumer wearable devices, 71% and 51% of Apple Watch (Apple Inc) readings (used in 49 studies) and Fitbit readings (used in 71 studies), respectively, were within 3% of electrocardiogram readings in controlled settings [39]. Moreover, in 3 free-living studies, the wrist-worn Fitbit Charge had a mean absolute error percentage of 10% [39]. A systematic review of wrist-worn devices that measure heart rate via plethysmography found limits of agreement of 8.4 BPM at rest, 30.1 BPM while on a treadmill, and 41.5 BPM while cycling [40]. Overall, our findings found large limits of agreement for all devices, and inpatient results were consistent with the wide limits of agreement found in free-living environments or with activity.

We found that sleep only had a moderate correlation with sleep survey results from inpatient settings the use research and consumer devices. A recent systematic review of Fitbit-based sleep assessments found that readings from more recently developed devices correlated well with polysomnography readings for assessing sleep episodes [41]. It is unclear whether the lower correlation that we found was due to inpatient settings with high nighttime interruptions, patient factors that were perhaps associated with acute illness, or issues with sleep surveys (or a combination of these three factors) [3]. With respect to respiratory rate, 2 studies of 2 medical-grade devices provided limits of agreement. Wider limits of agreement were found in the study that had over 50,000 measurement pairs and used a gold standard (27 breaths per minute) compared to those in the study that had less than 100 measurement pairs and used clinician-reported vitals (13.4-19.3 breaths per minute) [30,32]. Additionally, previous studies found that medical-grade devices were only accurate under laboratory conditions or at-home conditions [42,43]. There was a limited number of studies on oxygen saturation, temperature, blood pressure, and fall risk.

Limitations and Future Research

There are a few limitations that should be noted for our systematic review. There is a considerable risk of bias, as the number of participants in the studies was low. Further, the studies included were observational in design and had a high degree of heterogeneity in terms of the objectives, populations, and outcomes reported. Thus, the data analysis methods were limited to broad categorization and the extraction of the common themes and trends that emerged from the results. Reports of wearable monitoring from individual studies should be viewed based on their methodological limitations. Although patient adherence has been found to correlate well with patients’ acceptability of wearables devices in inpatient settings, we realize that studying factors such as data loss, the duration of data gaps, and qualitative feedback from nurses and patients would further strengthen the generalizability of the results. Finally, it is important to note that wearable studies are being increasingly performed, and more relevant articles will become increasingly available.

This review also identifies gaps in knowledge that still exist within literature and provides information about what is required for further research. Specifically, the further validation of digital biomarkers by using gold standard comparators, such as polysomnography for assessing sleep and continuous electrocardiogram monitoring for assessing heart rate, is required. Ideally, large participant sample sizes and large numbers of measurement pairs within a population of interest should be used to assess parameters such as vital signs. The use of 2 reference standards to validate each health parameter, such as a heart rate, has also been recommended [44]. Moreover, data that are derived under real life conditions are still needed to better understand the factors that may contribute to between-patient heterogeneity when comparing the accuracy of wearable readings, such as those for patient activity, posture, gait type and velocity, locations of wearables, and patients’ diagnoses (eg, seizures). Future studies can aim to further qualify the process of retrieving data by using wearables to explore other barriers and avenues that might hinder the collection of reliable health information (ie, a weak Bluetooth connectivity, a lack of patient digital health literacy, the added burden that the process of taking wearable readings has on clinicians, the learning curve required to operate a wearable, etc) Finally, while we found that some digital biomarkers appeared to be valid for the monitoring of inpatients via wearables, we were unable to find any studies that supported the use of wearables in inpatient settings to improve clinical outcomes.


Overall, the assessment of studies in this review suggested that wearable devices show promise for monitoring the heart rate and sleep of patients in hospitals. The results show that many devices were not validated in inpatient settings, and the readings from most of the devices that were validated in such settings had wide limits of agreement. Further research is needed to determine the accuracy of the digital biomarker readings of hospitalized patients and to eventually determine whether wearable devices improve the health outcomes of hospitalized patients.


RW is supported by an award from the Mak Pak Chiu and Mak-Soo Lai Hing Chair of General Internal Medicine, University of Toronto. Funding for this study was kindly provided by the University Health Network Foundation Complex Care Fund.

Authors' Contributions

VP, AOC, and RW designed and planned the review. AOC conducted the search strategy. VP and RW screened the articles and conducted the data analysis. VP, AOC, and RW wrote and revised the manuscript.

Conflicts of Interest

None declared.

Multimedia Appendix 1

Search methods.

DOCX File , 64 KB

Multimedia Appendix 2

Modified Consensus-Based Standards for the Selection of Health Status Measurement Instruments (COSMIN) criteria used for the risk of bias assessment.

DOCX File , 30 KB

  1. Mariani P, Saeed MU, Potti A, Hebert B, Sholes K, Lewis MJ, et al. Ineffectiveness of the measurement of 'routine' vital signs for adult inpatients with community-acquired pneumonia. Int J Nurs Pract 2006 Apr;12(2):105-109. [CrossRef] [Medline]
  2. Kessler R, Knutson KL, Mokhlesi B, Anderson SL, Shah M, Meltzer DO, et al. Sleep and activity patterns in older patients discharged from the hospital. Sleep 2019 Oct 21;42(11):zsz153 [FREE Full text] [CrossRef] [Medline]
  3. Stewart NH, Arora VM. Sleep in hospitalized older adults. Sleep Med Clin 2018 Mar;13(1):127-135 [FREE Full text] [CrossRef] [Medline]
  4. Weenk M, van Goor H, Frietman B, Engelen LJ, van Laarhoven CJ, Smit J, et al. Continuous monitoring of vital signs using wearable devices on the general ward: Pilot study. JMIR Mhealth Uhealth 2017 Jul 05;5(7):e91 [FREE Full text] [CrossRef] [Medline]
  5. Mapp ID, Davis LL, Krowchuk H. Prevention of unplanned intensive care unit admissions and hospital mortality by early warning systems. Dimens Crit Care Nurs 2013;32(6):300-309. [CrossRef] [Medline]
  6. Churpek MM, Yuen TC, Edelson DP. Predicting clinical deterioration in the hospital: the impact of outcome selection. Resuscitation 2013 May;84(5):564-568 [FREE Full text] [CrossRef] [Medline]
  7. Cretikos M, Chen J, Hillman K, Bellomo R, Finfer S, Flabouris A. The objective medical emergency team activation criteria: a case-control study. Resuscitation 2007 Apr;73(1):62-72. [CrossRef] [Medline]
  8. Krumholz HM. Post-hospital syndrome--an acquired, transient condition of generalized risk. N Engl J Med 2013 Jan 10;368(2):100-102 [FREE Full text] [CrossRef] [Medline]
  9. Alharbi M, Straiton N, Smith S, Neubeck L, Gallagher R. Data management and wearables in older adults: A systematic review. Maturitas 2019 Jun;124:100-110. [CrossRef] [Medline]
  10. Whyman J. Post-hospital syndrome: Tips to keep yourself or a loved one healthy after hospitalization. Harvard Health Publishing.   URL: https:/​/www.​​blog/​post-hospital-syndrome-tips-to-keep-yourself-or-a-loved-one-healthy-after-hospitalization-2019012315830 [accessed 2021-07-29]
  11. Kroll RR, Boyd JG, Maslove DM. Accuracy of a wrist-worn wearable device for monitoring heart rates in hospital inpatients: A prospective observational study. J Med Internet Res 2016 Sep 20;18(9):e253 [FREE Full text] [CrossRef] [Medline]
  12. Welch J, Kanter B, Skora B, McCombie S, Henry I, McCombie D, et al. Multi-parameter vital sign database to assist in alarm optimization for general care units. J Clin Monit Comput 2016 Dec;30(6):895-900 [FREE Full text] [CrossRef] [Medline]
  13. Gallo A, Lee KA. Sleep characteristics in hospitalized antepartum patients. J Obstet Gynecol Neonatal Nurs 2008;37(6):715-721 [FREE Full text] [CrossRef] [Medline]
  14. Yang CJ, Aibel K, Meyerhoff R, Wang F, Harpole D, Abernethy AP, et al. Actigraphy assessment of sleep quality among patients with acute myeloid leukaemia during induction chemotherapy. BMJ Support Palliat Care 2018 Sep;8(3):274-277. [CrossRef] [Medline]
  15. Mercer K, Li M, Giangregorio L, Burns C, Grindrod K. Behavior change techniques present in wearable activity trackers: A critical analysis. JMIR Mhealth Uhealth 2016;4(2):e40 [FREE Full text] [CrossRef] [Medline]
  16. Wieringa FP, Broers NJH, Kooman JP, Van Der Sande FM, Van Hoof C. Wearable sensors: can they benefit patients with chronic kidney disease? Expert Rev Med Devices 2017 Jul;14(7):505-519. [CrossRef] [Medline]
  17. Kirk MA, Amiri M, Pirbaglou M, Ritvo P. Wearable technology and physical activity behavior change in adults with chronic cardiometabolic disease: A systematic review and meta-analysis. Am J Health Promot 2019 Jun;33(5):778-791. [CrossRef] [Medline]
  18. Sim I. Mobile devices and health. N Engl J Med 2019 Sep 05;381(10):956-968. [CrossRef] [Medline]
  19. Wright SP, Brown TSH, Collier SR, Sandberg K. How consumer physical activity monitors could transform human physiology research. Am J Physiol Regul Integr Comp Physiol 2017 Mar 01;312(3):R358-R367. [CrossRef] [Medline]
  20. Jeffs E, Vollam S, Young JD, Horsington L, Lynch B, Watkinson PJ. Wearable monitors for patients following discharge from an intensive care unit: practical lessons learnt from an observational study. J Adv Nurs 2016 Aug;72(8):1851-1862. [CrossRef] [Medline]
  21. Baldwin C, van Kessel G, Phillips A, Johnston K. Accelerometry shows inpatients with acute medical or surgical conditions spend little time upright and are highly sedentary: Systematic review. Phys Ther 2017 Nov 01;97(11):1044-1065. [CrossRef] [Medline]
  22. Lim SER, Ibrahim K, Sayer AA, Roberts HC. Assessment of physical activity of hospitalised older adults: A systematic review. J Nutr Health Aging 2018;22(3):377-386. [CrossRef] [Medline]
  23. Anderson JL, Green AJ, Yoward LS, Hall HK. Validity and reliability of accelerometry in identification of lying, sitting, standing or purposeful activity in adult hospital inpatients recovering from acute or critical illness: a systematic review. Clin Rehabil 2018 Feb;32(2):233-242. [CrossRef] [Medline]
  24. Yetisen AK, Martinez-Hurtado JL, Ünal B, Khademhosseini A, Butt H. Wearables in medicine. Adv Mater 2018 Jun 11:e1706910 [FREE Full text] [CrossRef] [Medline]
  25. Pavic M, Klaas V, Theile G, Kraft J, Tröster G, Guckenberger M. Feasibility and usability aspects of continuous remote monitoring of health status in palliative cancer patients using wearables. Oncology 2020;98(6):386-395 [FREE Full text] [CrossRef] [Medline]
  26. Moher D, Liberati A, Tetzlaff J, Altman DG. Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med 2009 Jul 21;6(7):e1000097. [CrossRef] [Medline]
  27. Mokkink LB, Terwee CB, Patrick DL, Alonso J, Stratford PW, Knol DL, et al. The COSMIN checklist for assessing the methodological quality of studies on measurement properties of health status measurement instruments: an international Delphi study. Qual Life Res 2010 May;19(4):539-549 [FREE Full text] [CrossRef] [Medline]
  28. Terwee CB, Mokkink LB, Knol DL, Ostelo RWJG, Bouter LM, de Vet HCW. Rating the methodological quality in systematic reviews of studies on measurement properties: a scoring system for the COSMIN checklist. Qual Life Res 2012 May;21(4):651-657 [FREE Full text] [CrossRef] [Medline]
  29. Bloch F, Gautier V, Noury N, Lundy J, Poujaud J, Claessens Y, et al. Evaluation under real-life conditions of a stand-alone fall detector for the elderly subjects. Ann Phys Rehabil Med 2011 Sep;54(6):391-398 [FREE Full text] [CrossRef] [Medline]
  30. Breteler MJM, Huizinga E, van Loon K, Leenen LPH, Dohmen DAJ, Kalkman CJ, et al. Reliability of wireless monitoring using a wearable patch sensor in high-risk surgical patients at a step-down unit in the Netherlands: a clinical validation study. BMJ Open 2018 Dec 27;8(2):e020162 [FREE Full text] [CrossRef] [Medline]
  31. Kroll RR, McKenzie ED, Boyd JG, Sheth P, Howes D, Wood M, WEARable Information Technology for hospital INpatients (WEARIT-IN) study group. Use of wearable devices for post-discharge monitoring of ICU patients: a feasibility study. J Intensive Care 2017;5:64 [FREE Full text] [CrossRef] [Medline]
  32. Steinhubl SR, Feye D, Levine AC, Conkright C, Wegerich SW, Conkright G. Validation of a portable, deployable system for continuous vital sign monitoring using a multiparametric wearable sensor and personalised analytics in an Ebola treatment centre. BMJ Glob Health 2016 Jul 05;1(1):e000070 [FREE Full text] [CrossRef] [Medline]
  33. Chiu H, Lo W, Chiang Y, Tsai P. The effects of sleep on the relationship between brain injury severity and recovery of cognitive function: a prospective study. Int J Nurs Stud 2014 Jun;51(6):892-899. [CrossRef] [Medline]
  34. Lee SY, Lee KA. Early postpartum sleep and fatigue for mothers after cesarean delivery compared with vaginal delivery: an exploratory study. J Perinat Neonatal Nurs 2007;21(2):109-113. [CrossRef]
  35. Razjouyan J, Grewal GS, Rishel C, Parthasarathy S, Mohler J, Najafi B. Activity monitoring and heart rate variability as indicators of fall risk: Proof-of-concept for application of wearable sensors in the acute care setting. J Gerontol Nurs 2017 Jul 01;43(7):53-62. [CrossRef] [Medline]
  36. Weller RS, Foard KL, Harwood TN. Evaluation of a wireless, portable, wearable multi-parameter vital signs monitor in hospitalized neurological and neurosurgical patients. J Clin Monit Comput 2018 Oct;32(5):945-951. [CrossRef] [Medline]
  37. Watkins T, Whisman L, Booker P. Nursing assessment of continuous vital sign surveillance to improve patient safety on the medical/surgical unit. J Clin Nurs 2016 Jan;25(1-2):278-281 [FREE Full text] [CrossRef] [Medline]
  38. Duus CL, Aasvang EK, Olsen RM, Sørensen HBD, Jørgensen LN, Achiam MP, et al. Continuous vital sign monitoring after major abdominal surgery-Quantification of micro events. Acta Anaesthesiol Scand 2018 Oct;62(9):1200-1208. [CrossRef] [Medline]
  39. Fuller D, Colwell E, Low J, Orychock K, Tobin MA, Simango B, et al. Reliability and validity of commercially available wearable devices for measuring steps, energy expenditure, and heart rate: Systematic review. JMIR Mhealth Uhealth 2020 Sep 08;8(9):e18694 [FREE Full text] [CrossRef] [Medline]
  40. Zhang Y, Weaver RG, Armstrong B, Burkart S, Zhang S, Beets MW. Validity of wrist-worn photoplethysmography devices to measure heart rate: A systematic review and meta-analysis. J Sports Sci 2020 Sep;38(17):2021-2034. [CrossRef] [Medline]
  41. Haghayegh S, Khoshnevis S, Smolensky MH, Diller KR, Castriotta RJ. Accuracy of wristband Fitbit models in assessing sleep: Systematic review and meta-analysis. J Med Internet Res 2019 Nov 28;21(11):e16273 [FREE Full text] [CrossRef] [Medline]
  42. Chan AM, Selvaraj N, Ferdosi N, Narasimhan R. Wireless patch sensor for remote monitoring of heart rate, respiration, activity, and falls. 2013 Presented at: 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); July 3-7, 2013; Osaka, Japan. [CrossRef]
  43. Chan AM, Ferdosi N, Narasimhan R. Ambulatory respiratory rate detection using ECG and a triaxial accelerometer. 2013 Presented at: 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC); July 3-7, 2013; Osaka, Japan. [CrossRef]
  44. Sartor F, Papini G, Cox LGE, Cleland J. Methodological shortcomings of wrist-worn heart rate monitors validations. J Med Internet Res 2018 Dec 02;20(7):e10108 [FREE Full text] [CrossRef] [Medline]

BPM: beats per minute
COSMIN: Consensus-Based Standards for the Selection of Health Status Measurement Instruments

Edited by L Buis; submitted 12.12.19; peer-reviewed by M Breteler, D Atsma, L Li, K Ng, R Eckhoff; comments to author 28.08.20; revised version received 21.10.20; accepted 15.07.21; published 18.08.21


©Vikas Patel, Ani Orchanian-Cheff, Robert Wu. Originally published in JMIR mHealth and uHealth (, 18.08.2021.

This is an open-access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mHealth and uHealth, is properly cited. The complete bibliographic information, a link to the original publication on, as well as this copyright and license information must be included.