Abstract
Background: Digital health interventions incorporating self-management strategies are increasingly used to support individuals in managing depression. These interventions often leverage self-monitoring and passive sensor-based data collection to provide personalized feedback, guiding behavioral change. With the proliferation of smartphones and wearable devices, there is growing potential to continuously collect behavioral and physiological data. However, a major limitation in the field is the lack of consolidated evidence on which specific parameters are most useful for monitoring and predicting depression-related outcomes.
Objective: This scoping review aims to identify and synthesize common digital phenotyping and self-report parameters for monitoring and predicting depression. Specifically, it addresses the methodological and knowledge gap concerning which types of sensor-based and self-reported data are most frequently used and which demonstrate predictive value in tracking changes in depressive symptoms across digital platforms.
Methods: A literature search was conducted across 4 databases, including PubMed, Embase, Cochrane Library, and the Web of Science Core Collection. Articles published between January 1, 2021, and November 26, 2025, were included. Eligible studies included adults (≥18 years) with depression confirmed through validated clinical measures and using digital approaches that collected passive sensor data, self-reports, or both. Studies focusing on comorbid disorders, nondigital interventions, or not reporting depression-related outcomes were excluded. The PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines were followed, and a 5-stage methodological framework for scoping reviews was implemented. Quality assessment was performed using the Downs and Black Instrument and the Mixed Methods Appraisal Tool (MMAT).
Results: Nineteen studies were included, comprising a total of 85,193 participants. Most studies used smartphone- or wearable-based tools, with passive sensing as the predominant data source and Patient Health Questionnaire-9 (PHQ-9) as the most commonly used depression measure. Five overarching parameter categories were identified: (1) physical activity and location, (2) behavioral patterns, (3) physiological signals, (4) sleep indicators, and (5) sociability and self-reported assessments. Within these categories, 11 metrics, including step count, heart rate variability, sleep duration and mood self-ratings, were most frequently reported. Most studies used a multimodal digital phenotyping approach, integrating passive sensor-derived data with active user-reported input, enabling more individualized symptom monitoring over time.
Conclusions: This scoping review provides a novel synthesis of common digital parameters used across diverse tools for monitoring and predicting depression, moving beyond tool- or modality-specific perspectives adopted in prior reviews. Unlike existing reviews focusing on individual sensing modalities, prediction methods, or intervention effectiveness, this review maps shared parameters across observational, predictive, and interventional studies. By identifying convergent digital markers, the review supports comparability across studies and informs future model development. These findings have practical implications for the design of scalable digital mental health tools and for translating digital phenotyping into real-world clinical and self-management contexts.
doi:10.2196/70840
Keywords
Introduction
Advancements in personal digital devices and sensor technologies have enabled continuous and unobtrusive data collection from individuals in real-world settings, expanding the possibilities in health care data management []. This has given rise to digital phenotyping, defined as the moment-by-moment quantification of individual-level behavior and aspects of physiology using data from personal smartphones and wearable devices of users [].
The sensors commonly embedded in these devices can be broadly categorized into (1) motion sensors, (2) location sensors, and (3) environment sensors []. More specifically, smartphone-based sensors include linear acceleration, rotation vector, humidity, orientation, gravity, magnetic field, accelerometer, ambient temperature, gyroscope, light, pressure, microphone, proximity, and others []. Each device uses multiple sensors working together to provide optimal output for the user. For instance, environmental sensors, including humidity, illumination intensity, surrounding pressure, and temperature sensors enable automatic smartphone screen brightness adjustment or calculation of the dew point for the day []. Similarly, the geomagnetic sensor and accelerometer show user-relative location to the North Pole, providing basic data for location sensors. Motion sensors assess acceleration, speed of revolution, gravity, curl vector values, and drift, allowing tracking of multiple motions of the user, such as car direction (eg, navigation application), steps walked (eg, fitness or exercise mobile applications), and speed of movement (eg, user car speed) [,].
Unobtrusive tracking of health data relying on physiological parameters such as heart rate, step count, sleep, and others commonly occurs by using sensor fusion (ie, putting multiple signals together) []. For instance, biosensors can track physical parameters such as heart rate [], while location sensors can disclose patient activity and potentially help infer how often outdoor activities occur, which is particularly important for individuals living with mood disorders such as depression []. Social patterns can be monitored through the aggregated amount of mobile phone calls and textual messages exchanged with others, pointing to social activity or potential social withdrawal, often serving as an indicator of a depressive episode []. Finally, sensor fusion can be used by combining multiple sensors such as accelerometer, ambient light, battery level, microphone, phone lock status, phone unlock status, and screen status for the physiological parameter of sleep that has a significant role in indicating depressive episodes [].
Some of the earlier works that explored the field of digital phenotyping in mental health include a systematic review [] of smartphone-based passive sensing, summarizing common data sources such as GPS, accelerometry, and phone usage metrics. Several research papers have outlined the promise of personal sensing approaches in psychiatry while also emphasizing challenges related to data interpretation and real-world implementation [,]. Early empirical support for the clinical relevance of smartphone-derived data has been demonstrated through significant correlations between sensor-based measures and depressive symptom severity [].
Recent reviews have demonstrated growing interest in the use of passively collected data for depression monitoring, reflecting the rapid advancement of digital phenotyping approaches. For example, several studies have focused on specific technologies, such as smartphone-based sensing alone, or emphasized algorithm development for symptom detection rather than broader clinical integration [-]. Others have targeted a narrower set of parameters [], which limits their scope for cross-study synthesis. In contrast, the present review aims to provide a comprehensive overview of the parameters most commonly used in digital interventions for depression, regardless of delivery platform, by synthesizing data across smartphone- and wearable-based sensing studies. This broader focus is intended to support standardization efforts and inform the development of more effective and clinically relevant digital mental health tools.
There is a considerable variability in the types of parameters used, the degree of personalization, and the delivery modes, making it difficult to compare studies or determine which features are most effective [-]. This lack of consistency makes it difficult to develop reliable digital health tools. Moreover, health practitioners and end users often lack clear guidance on which digital tools offer meaningful and reliable support.
To address such gaps, this review aims to map the landscape of passively collected parameters used in digital interventions for depression, with the goal of informing future research, standardization, and clinical application.
This review is guided by the following research questions:
- What passively collected parameters are commonly used in digital interventions for depression to support longitudinal monitoring of depressive symptoms?
- Which types of passively collected data are used?
- Which parameters are most frequently used?
- Which parameters have demonstrated predictive value for changes in symptom severity?
We define a parameter as any sensor-derived measure used to monitor or predict depression-related outcomes.
Methods
Study Design and Framework
This scoping review was conducted in accordance with the 5-stage methodological framework for scoping reviews, including (1) identifying the research question, (2) identifying relevant studies, (3) study selection, (4) data charting and synthesis, and (5) collating, summarizing, and reporting results []. The data were systematically collected and reported by following the PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews) guidelines () [].
This scoping review was not registered in PROSPERO (International Prospective Register of Systematic Reviews), as PROSPERO does not accept registrations for scoping review protocols at the time of this writing.
Data Sources and Search Strategy
We conducted a comprehensive search of 4 electronic databases, including PubMed, Embase, Cochrane Library, and Web of Science Core Collection. Study registries were searched, including ClinicalTrials.gov and the World Health Organization International Clinical Trials Registry Platform (WHO ICTRP). The findings from these registries are categorized under “Others” in the PRISMA flow chart. No online or print sources were purposefully searched or browsed outside of the 4 electronic databases. No table of contents, conference proceedings, organizational websites, or print materials were hand-searched for this review. However, reference lists of included articles were also manually screened to identify additional eligible studies, which are categorized under “Other” in the PRISMA flow chart. No databases were searched simultaneously on a single platform, and no additional studies or data were sought by contacting study authors, experts, manufacturers, or others.
Studies published between January 1, 2021, and November 26, 2025, were included to reflect the recent advancements in digital phenotyping and mobile sensor technology in mental health. Our searches applied limits for publication date and language. All searches were last conducted on November 26, 2025 across all included databases, restricted to publications in English or German. These limits were used to ensure that the review captured the most current developments in digital interventions and digital phenotyping, as the technological landscape in this field evolves rapidly. No additional limits (eg, study design filters) were applied. The searches were updated manually by rerunning all database queries prior to final analysis. No automated alerts, citation tracking tools, or continuous updating services were used. The search strategy combined controlled vocabulary where applicable (eg, medical subject headings [MeSH] in PubMed) and free-text terms related to depression, digital health interventions, sensor data, and digital phenotyping. No published search filters were used for this review, and no search strategies from previous literature reviews were adapted or reused. All search strategies were developed manually using combinations of keywords and medical subject headings relevant to depression, digital phenotyping, and digital health interventions. The search strategy did not use any existing validated filters (eg, clinical study filters or methodological filters). No formal search peer review process was conducted for this review.
A summary of keyword combinations is provided in . For detailed search strategy see PRISMA-ScR containing a detailed search strategy across databases.
| Database | Keywords combined | Date range |
| PubMed | “Depression” AND “digital health intervention” “Depression” AND “digital phenotyping” “Self-management” OR “Depression” AND “digital phenotyping” “Depression” AND “digital phenotyping” AND “passive intervention” “Depression” AND “sensor data” | January 1, 2021-November 26, 2025 |
| EMBASE | “Depression,” “digital health intervention” “Depression,” “digital phenotyping” “Self-management” / “Depression,” “digital phenotyping” “Depression,” “digital phenotyping,” “passive intervention” “Depression,” “sensor data” | January 1, 2021-November 26, 2025 |
| Cochrane Library | “Depression” AND “digital health intervention” “Depression” AND “digital phenotyping” “Self-management” OR “Depression” AND “digital phenotyping” “Depression” AND “digital phenotyping” AND “passive intervention” “Depression” AND “sensor data” | January 1, 2021-November 26, 2025 |
| Web of Science Core Collection | “Depression,” “digital health intervention” “Depression,” “digital phenotyping” “Self-management” / “Depression,” “digital phenotyping” “Depression,” “digital phenotyping,” “passive intervention” “Depression,” “sensor data” | January 1, 2021-November 26, 2025 |
A total of 3243 records were identified during the study selection process, screened (n=1406), and reviewed for eligibility (n=134). After applying inclusion and exclusion criteria, a total of 19 studies were included in the final analysis. Reasons for exclusion included studies without digital interventions, nonclinical populations, and studies focused on comorbid conditions or unrelated mental health disorders.
The total number of records identified within each database and other information sources was 3243 (PubMed n=518; Cochrane Library n=1256; EMBASE n=791; Web of Science n=276; and Others n=402). Relevant records were identified from 4 databases (n=518) and clinical trial registers (n=2725). After duplicate removal, 1406 records were screened. Following full-text screening and exclusions, 19 studies were included in the final review. Common reasons for exclusion included lack of intervention focus (n=27), nonclinical populations (n=13), or nondigital approaches (n=9). Records retrieved from all databases and clinical trial registries were exported into Microsoft Excel and manually reviewed to identify and remove duplicates. Deduplication was performed by comparing study titles, authors, publication years, and digital object identifiers. No automated or software-assisted deduplication tools (eg, Covidence [Veritas Health Innovation] or EndNote [Clarivate]) were used.
Eligibility Criteria
Studies including randomized controlled trials, observational studies, and mixed method primary studies were considered. Eligible studies include digital health interventions for adults (≥18 years) concretely addressing depression that is assessed primarily with clinical measures such as the Patient Health Questionnaire-9 (PHQ-9) or similar validated instruments. The population focus is mainly on participants that, at the time of study participation, were experiencing depression, noted with a clinical measurement tool and digital phenotyping methodology. Study reports on outcomes related to the assessment of digital phenotyping for depression or dedicate a substantial amount of work to such reporting. The timeframe for published studies was limited from January 1, 2021 to November 26, 2025 to capture the most recent advances in digital phenotyping and mobile sensor monitoring in mental health. This period marks a notable shift from early feasibility and pilot work to more sophisticated methods, including multisite validation, machine learning–based prediction models, and integration of wearables and smartphone sensors []. By focusing on this timeframe, we ensure that the review remains aligned with current technological capabilities and clinically relevant research directions.
Study Selection
Title and abstract preselection was performed by 2 authors LB and MP. Screening questions included:
- Does the study use a digital intervention for depression delivered via a digital tool (eg, mobile app, wearable, web-based platform)?
- Does the study report on depression outcomes assessed through clinical instruments and/or digital phenotyping?
Relevant studies were retrieved and screened in full text by one author LB. After removing duplicates, LB, MP, and RA cross-checked the full texts against the inclusion and exclusion criteria. UH served as a supervisory reviewer. In an online meeting, final study selections were presented by LB and MP, who provided justification for inclusion decisions to UH.
Inclusion Criteria
Studies were included if they met the following criteria:
- Examined a digital intervention aimed at monitoring or improving depressive symptoms
- Delivered via smartphone apps, wrist-worn devices, web-based platforms, offline software, or automated interactive voice response systems
- Included adult participants (≥18 years) with depression, confirmed through validated clinical instruments (eg, PHQ-9)
- Reported on outcomes related to monitoring and/or symptom improvement using digital phenotyping (passive sensor data and/or active user input)
- Provided data on device use (eg, number of interventions delivered, duration of engagement)
- Published in English or German, and accessible via open access or institutional access.
Exclusion Criteria
Studies were excluded if they:
- Focused on depression comorbid with other conditions (eg, cancer, substance use, pregnancy, or other mental health disorders)
- Focused on depression in the context of a viral pandemic (eg, COVID-19)
- Studies use or report on depression based on digital interventions aimed at supporting and improving physical activity [].
Analysis Approach
Data Extraction, Charting, and Synthesis
A structured Microsoft Excel form was developed by the author LB to extract key information from each included study. Extracted variables included:
- Study characteristics (author, year, and country)
- Participant demographics and sample size
- Type of digital intervention and delivery platform
- Sensor types and data collection methods
- Depression assessment tools (baseline and follow-up)
- Frequency and duration of interventions
- Outcomes related to depression monitoring or prediction
Once data were initially charted, it was presented in an online meeting to co-author MP and cross-checked among author LB and co-author MP. For more detailed procedures for selection of the studies, extraction, and exclusion, please see .
Due to heterogeneity in study design and reporting, a narrative synthesis approach was used. Data were charted and grouped into thematic categories of parameters (see ), including physical activity and location, behavioral patterns, physiological signals, sleep indicators, and sociability and self-reported input.
| Study ID | Reference identifier | Sample size (depression) | Data type | Digital tool type | Depression measure | Main findings | Predictive value reported |
| 1 | [] | 313 | Passive + Active | Smartphone app, wearable | PHQ-9 | Multiple behavioral and phone use features significantly correlated with PHQ-9 | Yes |
| 2 | [] | 400 | Active | Computer | PHQ-9 | Behavioral test metrics significantly predicted depression | Yes |
| 3 | [] | 290 | Passive | Fitbit | PHQ-9 | Variations in heart rate and circadian rhythm linked to depression | Yes |
| 4 | [] | 92 | Passive | Smartphone app | PHQ-9 | User engagement is essential for data reliability | Unclear |
| 5 | [] | 81277 | Active | Smartphone app | PHQ-9 | Minimal average usage; thought diary and tests most used | No |
| 6 | [] | 934 | Passive | Smartphone app | PHQ-2 | Weak correlation between keystroke timing and symptoms | No |
| 7 | [] | 65 | Passive | Smartphone app | CES-D | Subject-specific standardization improved forecasting | Yes |
| 8 | [] | 60 | Passive + Active | Smartphone app, Oura Ring | DASS-21 | Best prediction from combined smartphone + wearable + mood | Yes |
| 9 | [] | 28 | Active | Computer | PHQ-8 | Digital social rhythm therapy improved depression | Yes |
| 10 | [] | 614 | Passive | Smartphone app, Fitbit | PHQ-8 | Lower engagement correlated with higher depression | Unclear |
| 11 | [] | 1 | Passive | Smartphone app, wearable | Not stated | Demonstrated potential of ambient light and usage data | Unclear |
| 12 | [] | 18 | Passive | Video assessment | MADRS, MINI, CSSRS | Facial and speech features correlated with MADRS | Yes |
| 13 | [] | 20 | Passive + Active | Smartphone app, wearable | MADRS, HDRS | Self-report had higher predictive correlation than passive sensors | Yes |
| 14 | [] | 623 | Passive | Smartphone, Fitbit | PHQ-8 | Sleep timing features correlated with depression severity | Yes |
| 15 | [] | 107 | Active | Smartphone app | PHQ-9 | Web-based support linked to symptom improvement | Yes |
| 16 | [] | 43 | Active | Smartphone app | BDI | Smartphone mood ratings correlated with standard assessments | Yes |
| 17 | [] | 48 | Passive + Active | Smartphone app, wearable | PHQ-9 | Higher engagement led to greater symptom improvement | Yes |
| 18 | [] | 136 | Active | Smartphone app | BDI-II | Significant treatment-related symptom changes | Yes |
| 19 | [] | 124 | Active | Computer | DASS-21 | Personalized feedback led to greater symptom reduction | Yes |
aPHQ-9: Patient Health Questionnaire-9.
bPHQ-2: Patient Health Questionnaire-2.
cCES-D: Center for Epidemiological Studies Depression Scale.
dDASS-21: Depression Anxiety Stress Scale Short Form 21.
ePHQ-8: Patient Health Questionnaire-8.
fMADRS: Structured Interview Guide for the Montgomery-Åsberg Depression Rating Scale.
gMINI: Mini-International Neuropsychiatric Interview,
hCSSRS: Columbia Suicide Severity Rating Scale.
iHDRS: Hamilton Depression Rating Scale.
jBDI: Beck Depression Inventory.
kBDI-II: Beck Depression Inventory, Second Edition.
The frequency of each parameter category and its reported predictive value were also noted.
Quality Appraisal
To assess the quality of the included studies, we used two tools, including (1) the Downs and Black checklist [] used for randomized controlled trials and quantitative studies and (2) the Mixed Methods Appraisal Tool (MMAT) [] used for mixed method and nonrandomized studies.
These complementary tools allowed a consistent evaluation of methodological rigor across diverse study designs. Detailed quality assessments are included in -.
Results
Summary of Included Studies
A total of 19 studies [,,-] with 85,193 participants across studies were included, covering a diverse range of digital intervention types and data collection methods. The study selection process is shown in .

Most digital tools or platforms (n=15) used smartphone apps, either alone or in combination with wearable devices (eg, Fitbit [Fitbit, Inc], Oura Ring [Oura Health Oy]). Three studies (16%) used computer-based platforms, and one used a computer-based platform for video-based digital phenotyping. The included studies varied greatly in design and measurement approaches (see ). The majority were observational studies (n=14), including cross-sectional, longitudinal cohort, retrospective, interformat validation, and feasibility designs, while 5 studies delivered a digital mental health intervention [,,-]. Three studies were randomized controlled trials [,,], and 2 used a retrospective observational approach [,]. Several observational studies used repeated-measures designs to monitor depressive symptoms over time, although study duration and assessment frequency were heterogeneous. One study collected single-participant feasibility data [], while another followed participants longitudinally for 43 weeks [].
| Study design and subtype (if applicable) | N | Reference number | |
| Randomized controlled trials (RCTs) | 3 | [,,] | |
| Nonrandomized interventional pilot studies | |||
| 2 | [,] | ||
| Observational studies | |||
| 5 | [,,,,] | ||
| 5 | [,,,,] | ||
| 2 | [,] | ||
| 1 | [] | ||
| 1 | [] | ||
Data Collection Approaches
Three data collection types were noted, including passive data collection, active input, self-reports, and multimodal approaches, which combined both passive and active inputs.
Passive data collection (eg, from GPS, accelerometer, and heart rate sensors) was used in 84% of studies (n=16).
Active input and self-reports (eg, mood ratings, diary entries) were used in 58% of studies (n=11).
Multimodal approaches combining both passive and active inputs showed the highest predictive potential and were used in 47 % of studies (n=9).
Clinical Measures Used Across Studies
Validated tools used across studies included Patient Health Questionnaire versions 9, 8 and 2 (PHQ-9, PHQ-8, and PHQ-2), the Center for Epidemiologic Studies Depression Scale (CES-D), the Depression Anxiety and Stress Scale (DASS-21), the Beck Depression Inventory original and revised version (BDI-I and -II), the Montgomery-Åsbergg Depression Rating Scale (MADRS), the Hamilton Depression Rating Scale (HDRS), and the Mini-International Neuropsychiatric Interview (MINI), reflecting a broad clinical foundation across studies (see ).
The PHQ-9 was the most used depression measure, used in 9 studies. The second most used depression measure included PHQ-8 (studies n=3) [,,] while PHQ-2 was used in only one study []. Other often used measures were DASS (studies n=2) [,], MADRS (studies n=2) [,], MINI. (studies n=2) [,] and HDRS [] used in only one study.
Predictive Value
Of the 19 studies, 15 (79%) [,,-,-,-] reported predictive or diagnostic associations between digital parameters and depression outcomes. Predictive value was strongest in studies that:
- Incorporated multisource data fusion [,,]
- Applied machine learning or modeling techniques for forecasting [,].
Sensor Types and Platforms Breakdown
Across the reviewed studies, most digital interventions assessed or observed relied on smartphones as the primary platform, often enhanced with wearable devices (eg, Fitbit, Oura Ring) for additional collection of physiological data.
The most frequently used sensors included accelerometers, gyroscopes, GPS, light sensors, and microphone sensors.
Wearables primarily contributed heart rate and sleep cycle data, while smartphones captured mobility, screen interaction, ambient context, and social communication metrics.
Only a few studies [,] used computer-based platforms or video-based assessments, which focused on cognitive tasks, facial expressions, and speech expression as digital biomarkers.
Across retrieved studies the most common device used for digital data collection was a smartphone (n=13), while wearable devices such as sensors and other wrist-worn devices (eg, Fitbit) were used in 8 studies. A detailed list of devices with the main findings reported can be seen in .
Key Findings
The key findings were grouped into five major categories of passively collected parameters: (1) physical activity and location, (2) behavioral patterns, (3) physiological data, (4) sleep, and (5) sociability and self-reported assessments (see and ). summarizes the most commonly used passively collected parameters in digital interventions for depression. Sleep-related parameters such as total sleep time, sleep onset, and heart rate during sleep were particularly prevalent, often inferred through sensor fusion.
Similarly, physical activity metrics like step count and accelerometer data, along with location tracking (eg, distance from home), were commonly used to assess daily functioning and routine stability.
Sociability parameters, including phone call frequency, text message frequency, and app usage, were often used to approximate social withdrawal.
Behavioral data like reaction time or user engagement further contributed to characterizing depressive symptomatology. illustrates how these domains interconnect and highlights their contribution to monitoring and predicting depression.
This multidomain structure, used in 47% of studies, supports a more concrete understanding of symptom trajectories and lays a foundation for developing more reliable personalized digital interventions that incorporate digital parameters.
- Physical Activity and location parameters—such as step count, phone use duration, and location-based data—were reported in 8 studies and captured through various digital interventions (see ).
- Behavioral patterns included parameters like user engagement, often assessed via response time to survey notifications. These were also reported in 8 studies.
- Physiological data encompassed diverse indicators such as heart rate, circadian rhythms, and facial expressivity. These were observed in 8 studies and often overlapped conceptually with parameters categorized under “Sleep”, such as total sleep time, sleep onset, and nighttime heart rate. Sleep-related parameters were collected in 7 studies.
- Sociability and self-reported assessments involved features inferred from communication behaviors, including the number and duration of phone calls, text message counts, and emoticon use. These were reported in 3 studies. Self-reported mood or symptom assessments, often gathered through in-app or EMA-style prompts, were present in nearly all included studies and frequently used alongside sensor data.

| Study ID | Sleep | Physical activity | Sociability | Location | Physiological | Other | Total parameters measured |
| 1 | Yes | Yes | Yes | No | Yes | No | 34 |
| 2 | No | No | No | No | No | No | — |
| 3 | Yes | Yes | Yes | No | Yes | No | 126 |
| 4 | No | Yes | No | Yes | No | No | — |
| 5 | No | No | Yes | Yes | No | No | — |
| 6 | Yes | No | No | No | No | Yes | 54 |
| 7 | Yes | No | No | Yes | No | Yes | 19 |
| 8 | Yes | Yes | No | Yes | Yes | Yes | — |
| 9 | No | No | No | No | No | Yes | — |
| 10 | No | No | No | No | No | No | — |
| 11 | No | Yes | No | Yes | No | Yes | 5 |
| 12 | No | Yes | No | No | Yes | Yes | — |
| 13 | No | No | No | No | No | No | — |
| 14 | Yes | No | No | Yes | No | Yes | 21 |
| 15 | No | No | No | No | No | Yes | — |
| 16 | Yes | Yes | No | Yes | No | Yes | — |
| 17 | No | Yes | No | Yes | Yes | Yes | — |
| 18 | No | No | No | No | No | Yes | — |
| 19 | No | No | No | No | No | No | — |
aNot applicable.
Across the studies, the most robust prediction of depressive symptoms emerged from models that combined smartphone and wearable sensor data with self-reported assessments. Although there was considerable heterogeneity in the parameters used, ranging from location variability and facial expressivity to sleep duration and keystroke timing, some overlap was noted. For instance, nighttime heart rate could be classified under either physiological data or sleep, depending on the study context. In such cases, we categorized the parameter according to how it was used in the original study.
Elaboration of Key Findings
Physical Activity and Location
The majority of studies that gathered passive data continually also supplied data about patient’s locations. Additionally, these studies tracked patients’ physical activity via accelerometer sensor. Results regarding physical activity and location parameters were homogenous. One study described a negative association between variability of locations visited and symptoms of depression (r=−0.21) [], showing that the more active the participant is (ie, the more locations visited), the lower the symptoms of depression appear to be, which can also infer social context, as more activity could also signify more social interactions. Similarly, another study described a negative correlation between higher MADRS scores and a lower average distance from home (r=−0.25) [], supporting the argument that greater depression severity is associated with reduced physical activity.
Behavioral Patterns
Studies categorized in behavioral patterns predominantly focused on user engagement and behavior of the patient. One of the included studies revealed that lower user engagement results in higher depression scores []. Although, the engagement in this case was defined as the time it took the user to respond to survey notifications. The longer the time users took to respond, the less engaged they were and the more severe their depressive symptoms were shown to be. In line with this, another study noted a correlation between higher engagement and an increase in symptom improvement, described as a decrease in at least one of the depressive symptoms and/or a clinically significant improvement []. User engagement can be used to assess the severity of depressive symptoms, and it is also important for gathering data continually, especially for gathering relevant data such as location and accelerometer; hence, active user engagement was found to be essential [] also in supporting the physical activity and location category.
Apart from engagement rates, other behavioral markers differed in relation to depressive symptoms as well. A prospective cohort study, included in this review, exploring factors associated with the onset and course of mental health conditions during adolescence reported a weak negative association between most of the 54 analyzed keystroke timing features and mental health symptoms []. Machine learning models using keystroke features alone were shown to be unable to predict mental health symptoms, while speech patterns and other associated voice-related markers, such as words per minute, were found to be negatively correlated (r=−0.09) with PHQ-9 [] indicating that higher depression scores are associated with decreased speech and expression, which potentially has value for the (5) sociability and self-reported assessments category as well.
Physiological Data
Common parameters in this category were difficult to compare, as they vary in nature. Multiple physiological markers were shown to correlate with the severity of depressive symptoms across studies. For instance, one study reported significant increases in multiple digital markers, especially in facial expressivity such as fear and anger []. These markers were shown to decrease often in frequency, correlating to changes in the MADRS scores. Concretely, head movement and amount of speech correlated with changes in the MADRS scores [], while other physiological parameters were associated (r=0.70) with greater severity of depressive symptom, such as greater variation of nighttime heart rate []. Based on the daily number of steps, depressive symptoms were associated (r=0.62) with lower regularity of weekday circadian rhythms []. In fact, this particular study also reported overall a limited ability of digital biomarkers to detect depression in the whole sample []. However, stronger correlations (r=0.8‐1.0) were described among pairs of variables in alpha power and gamma power spectra in electroencephalography with MADRS [], indicating that some pairs of variables in electroencephalographic streams can be measured when depression scores are increasing as well, being potentially useful as a digital biomarker in the prediction of depression.
Sleep
Multiple studies reported sleep-related data, consisting of passively gathered data or sleep-related questionnaires such as the Pittsburgh Sleep Quality Index, the Sleep Hygiene Index, or the Epworth Sleepiness Scale. Results were comparatively homogenous. One study reported a significant positive association between the total sleep time and depression (r=0.15), also waking after the sleep onset and depression (r=0.20), as well as heart rate variability and anxiety (r=0.12) []. Similarly, another study described a negative correlation (r=−0.12) between reported sleep duration and PHQ-9 []. Eventually, a stronger correlation of sleep-onset time with PHQ-8 was found cross-sectionally, and a higher correlation (r=0.24) of wakefulness after sleep-onset with PHQ-8 was also observed longitudinally [].
Sociability and Self-Reported Assessments
In this category, studies included parameters such as self-reported mood or parameters pertaining to social behavior and sociability (eg, phone calls, text messages, and other communication apps). Since this is a rather broad range, comparison is difficult. In accordance with earlier described results, one study reported a correlation of PHQ-9 with voice-related markers such as Voice Diary words per minute, Voice Diary sentiment, and Voice Diary duration (Mood Study App, Verily Life Sciences LLC). Markers such as call ringing until missed minutes (r=0.10) showed a correlation, and phone call count (r=−0.06) showed a negative correlation as well, as did emoticon count in text messages (r=−0.06), battery percentage (r=−0.7), and ambient audio level (r=0.10) [].
Another study described a negative correlation of high MADRS scores with lower entropy of the usage time of communication apps and lower total count of communication app usage (r=−0.42), indicating a decreasing sociability as depression scores were increasing. A strong positive correlation (r=0.9) was noted between self-reported outcomes and total MADRS score []. User behavior was also analyzed in a different study that described a typical person using the mental health app 3 times a day for a total of 12 minutes over the course of 90 days. The most frequently used tools in this study app were the Thought Diary (MoodTools, Inquiry Health LLC) and the self-assessment tests []. Significant treatment and time interactions for the BDI, GAD-7, Recovery Assessment Scale (RAS), Rosenberg Self-Esteem Scale (RSES), and Sheehan Disability Scale (SDS) were described []. Another study reported that participation in Cognitive Behavior Immersion (CBI) sessions via app reduced symptoms of anxiety and depression in patients (r=−1.5) [].
A large portion of findings within this category relies on behavior data and as such is closely related to the behavioral patterns category. However, apart from user behavior, social patterns such as social support are found to be influencing the severity of depressive symptoms and the prediction thereof. Web-based social support was found to predict the improvement of depressive symptoms [] while social-rhythm-focused digital interventions were shown to benefit patients significantly. In fact, a more rapid rate of improvement according to the PHQ-8 was shown compared to nonpersonalized interventions []. Additionally, one study reported a greater decline in DASS-21 scores using personalized feedback derived from the experience sampling method (ESM) compared to using nonpersonalized feedback [].
The majority of studies apply several categories with a range of parameters used. Commonly, at least 4 categories were used for monitoring, assessing, or tracking depression across digital tools of the studies included (see ). The type of device used to deliver the support was not shown as a relevant factor in effectiveness, and high comparability between smartphone-based and nonsmartphone-based questionnaire scores was reported []. Nevertheless, it must be noted that categories are interdependent, and all studies used parameters existing across categories, focusing on multiple aspects.
Discussion
Principal Findings
By focusing on parameters rather than platforms or algorithms, this review responds directly to calls for greater comparability in digital phenotyping research. This scoping review identified five major categories of passively collected parameters commonly used in digital interventions for depression: (1) physical activity and location, (2) behavioral patterns, (3) physiological data, (4) sleep, and (5) sociability and self-reported assessments. Most studies used a combination of smartphone-based and wearable sensors alongside self-reports. The strongest predictive value for depression outcomes was observed in studies that integrated multiple data sources, especially when combined with machine learning or subject-specific modeling approaches. Despite promising findings, the field remains methodologically fragmented and lacks standardization in both sensor use and clinical integration.
Five overall categories distinguished in this review contain relevant parameters for monitoring, assessing, and potentially predicting depression by combining self-assessment and digital phenotyping in digital interventions. The existing findings point that the categories of (1) physical activity and location, (2) behavioral patterns, (3) physiological data, (4) sleep, and (5) sociability and self-reported assessments appear across literature as interdependent categories, used across studies to monitor and predict depression by using various passively gathered data from parameters such as accelerometer sensor, location clusters, average distance from home, step count, phone usage, number of phone calls, heart rate, sleep time, and voice-related markers, among others. Moreover, additional parameters retrieved also include 11 frequently used clinical and self-assessment measurements such as PHQ-9, PHQ-2, CES-D, DASS-21, PHQ-8, MINI, MADRS, CSSRS, HDRS, BDI, and BDI-II (see ).
This synthesis aligns with the World Health Organization guideline recommendations [] for strengthening universal health coverage, where the role of digital health interventions is addressed particularly in line with the insufficient quality parameters that can assure optimal health care support []. While they emphasize the need for improved quality parameters in digital health, our findings indicate that such parameters are arguably not missing but are rather scattered across literature without comparative analysis. It might be argued that current work is one of the early efforts in the field to collect or summarize such parameters for digital health interventions overall in a more structured manner.
Previous research has often examined digital phenotyping and depression in isolation [,,] or focused on either digital interventions [,,-] or productive models for symptom progression [,] without comprehensively linking all aspects. This review aims to contribute to bridging that gap by grouping relevant parameters into more standard categories.
Retrieved parameters in this review can be used as a comprehensive guideline in determining relevant aspects to be implemented into digital health interventions for depression to combine optimal sensors and self-assessment-based tools for monitoring and predicting depression outcomes. It can be argued that the interdependence of distinguished categories provides a more solid understanding for developing optimal digital health intervention in terms of monitoring and self-management support for depression. Combining these categories by implementing described parameters can support more beneficial outcomes for the end user of digital self-management tools for depression.
Synthesis of result sections of the included studies indicated that predicting depressive symptoms by combining clinical assessment and digital phenotyping is a promising approach for further improvement of digital interventions’ reliability and efficiency for individuals living with depressive symptoms and major depressive disorder. In a study using a transfer learning approach, an improvement of 12% for one-day-ahead forecasting was noted []. Moreover, using the MADRS score as a baseline for depression severity, the highest correlation of prediction was reported with user self-reported assessment []. Among other such assessments were the PHQ-2 and a model consisting of the Bond and Lader lethargic-energetic visual analog scale and the Bond and Lader interested-bored visual analog scale plus the N-Back working memory task [].
The scores obtained in a smartphone-based intervention correlated with the ratings of nonsmartphone-based HDRS. Concretely, the smartphone-based single-item self-rating of mood correlated negatively (r=−0.54) with the BDI sum of scores as well [], indicating that the lower the perceived mood of the user was, the higher the depression score. Therefore, participant self-reported assessments can potentially be valuable in predicting depression severity when combined with digital phenotyping for higher accuracy and reliability. In line with that, the findings support prior work showing that self-reported assessments when combined with sensor data, can meaningfully predict depressive symptom severity [,,].
Reflecting on the device delivering the intervention with a combined model of smartphone features, wearable features, and self-reported mood items is also noted in one of the retrieved studies’ approaches [].
However, this review also reveals substantial variability and methodological inconsistencies across studies. Among the retrieved studies there was a notable lack of standardization in contrast with the World Health Organization recommendations guidelines [] addressing the quality parameters. Sample sizes, study and intervention duration, sampling frequencies, and types of intervention differed considerably. This was particularly evident when efforts for comparison during the charting procedures were made, mirroring concerns in literature about standardization and reproducibility [,,]. Missing data was frequent, and most studies did not report a follow-up or provide justification for demographic admissions, therefore leading to the inclusion of studies assessing depression severity with validated measures with a notably limited amount of measurement scales used across studies. Taking into consideration the chronic nature of depression, it must be noted that short-term observations, less than 6 months, might also not be sufficient to successfully record meaningful changes in symptom trajectories or assess the sustained effectiveness of digital interventions for depression.
Across the included studies, it is important to distinguish between studies that primarily examined associations and those that validated predictive models based on digital data. Only a subset of studies (6/19) developed or evaluated models to predict depression status or symptom severity from digital signals, including work on behavioral and clinical modeling, cognitive task-based digital biomarkers, wearable-derived screening models, and sensor-plus-EMA-based forecasting [,-,,]. The remaining studies (13/19) were observational, feasibility, validation, or intervention trials. These studies primarily characterized associations between digital parameters and depressive symptoms, evaluated feasibility and engagement, validated self-report formats, or tested the clinical effects of digital interventions without undertaking formal predictive model validation [,-,-,-]. This distinction underlines that much of the evidence base still addresses which digital features relate to depression, whereas a smaller subset of studies evaluates how well these features can support robust and generalizable prediction.
Considering the rapidly developing nature of digital interventions and digital phenotyping in the mental health field, multiple challenges in conducting this review were noted. While digital phenotyping has been increasingly used to monitor symptoms and to gather vast amounts of passive data, discovering studies that combine digital phenotyping with digitally delivered interventions proved challenging. Although many studies address digital tools or passive monitoring separately, only a small percentage integrate both, further showing the fragmentation within the field. Additionally, the narrow focus of this review on depression and depressive symptoms exclusively led to discarding similar studies focused on other mental illnesses, narrowing our evidence pool. Nevertheless, the assessment of studies in our review was rather challenging due to the enormous number of articles addressing either digital interventions and depression or digital phenotyping and depression but rarely combining all these areas of interest together. This observation further aligns with literature where similar limitations have been noted previously in an umbrella review of smartphone sensors for monitoring depression []. In this review, studies often focused on generalized categories rather than specific parameters, restricting reproducibility and cross-study comparison.
Strengths and Limitations
This review provides a structured synthesis of passively gathered parameters used in digital interventions for depression, supplementing existing reviews that focus exclusively on either digital phenotyping [,] or digital interventions [,,,,]. By incorporating both, this review offers a clearer framework for understanding how sensor-derived data and self-reported assessments can be integrated in monitoring depression.
Although many high-quality systematic reviews and meta-analyses on digital phenotyping or digital interventions have been published, there is rather limited systematic literature synthesis known to the authors of this review that includes and combines all relevant terms such as “digital intervention,” “depression,” and “digital phenotyping.”
The review allows a structured overview of all relevant categories and parameters used in digital interventions that address depression with clinical measures and digital phenotyping methodology. An important strength of this review is the categorization of parameters. It allows a comparison across 5 interdependent categories, supporting efforts in the mental health field to determine and assure reliability in parameter use for future research and intervention designs, as well as clinical use, aligning also with World Health Organization recommendations [].
Nonetheless, several limitations of this review need to be acknowledged. First, the review’s narrow scope, focusing exclusively on studies that used both validated depression assessments and digital phenotyping, may have excluded relevant work on related conditions or interventions that did not incorporate passive data collection. While this decision was made to enhance clarity and focus, it inevitably limited the breadth of the review, as noted in other recent scoping and mixed methods reviews as well [,]. Second, although passive data gathering enables the collection of large volumes of data over relatively short periods, a few studies reported that only a small proportion of the collected data was directly relevant to their hypotheses, raising concerns about potential publication bias or overinterpretation of limited findings.
Included studies mostly described individual interventions in detail, although many did not report the individual parameters measured but rather focused on larger terms, which we define in our study as categories. Even though general categories across studies were comparable, the lack of information on the exact data gathered through digital phenotyping leads to limited reproducibility of these studies as well as limited comparability with similar studies, thereby raising questions about the implications of such parameters for future studies. Missing data handling was an issue encountered in numerous studies, varying from a lack of explanation regarding the drop-out rates over user-related problems or confounders to basic demographic information. Similar methodological weaknesses have been described as well in prior meta-analyses of mental health data for depression [,,].
Third, technical variability also limited comparison. Concretely, retrieved studies used different sensor types and different software, which could possibly influence data reliability. Technological difficulties were reported in some studies, though their influence on the reliability of data was rarely discussed. In line with that, few studies discussed the influence of signals being generated due to reasons outside of illness context as potential confounders (eg, the influence of weather conditions on the amount of time spent at home limiting both physical and social activity). Previous work has highlighted the need to better account for such nonclinical sources of signal variability [,].
Finally, this review confronts challenges comparable to those of many trials researching technology or digital devices, including limited transferability of results due to the nature of study populations. The users that participated in the included studies were for the most part younger people who are commonly more familiar and comfortable with the technologies used, apart from a few studies, and had only a few issues using the interventions as intended. However, this might not be the case with an older population, thus indicating that results may not necessarily be representative, an issue that has also been raised in other studies [,].
Although this review identifies promising directions for standardizing and optimizing digital interventions for depression, it also underlines the need for improved methodological consistency, demographic inclusivity, and clearer reporting of sensor-derived parameters in the future.
Conclusion
This scoping review offers an innovative contribution by systematically identifying and synthesizing common digital parameters used across heterogeneous digital tools for monitoring and predicting depression. In contrast to existing reviews that predominantly focus on specific sensing modalities, predictive algorithms, or the effectiveness of digital interventions, this review adopts a parameter-centric perspective, enabling comparison across observational, predictive, and interventional study designs. By mapping convergent parameters irrespective of device type or analytic approach, the review provides a clearer understanding of which digital features are consistently associated with depressive symptoms and outcomes. This synthesis advances the field by supporting standardization efforts, guiding future predictive model development, and reducing fragmentation in digital mental health research. In real-world settings, these findings can inform the design of interoperable digital tools, facilitate integration into clinical monitoring workflows, and support personalized, data-driven depression management beyond controlled research environments.
Acknowledgments
We would like to sincerely thank the editor and anonymous reviewers for their constructive feedback and thoughtful suggestions which helped improve the clarity and quality of this manuscript. AI tools (ChatGPT, OpenAI) were used for editing/formatting assistance in the first revision round for refining the existing paragraphs and removing redundancy; authors critically reviewed and approved all content.
Funding
This research did not receive any specific grant from funding agencies in the public or commercial sectors.
Data Availability
The data sets generated and analyzed during this study are available from the corresponding author on reasonable request.
Authors' Contributions
LB and UH conceptualized the study. LB, MP, and RA performed data extraction. LB and MP conducted the data analysis, while UH supervised the analysis process and contributed to clinical interpretation which then also supported the discussion. All authors contributed to the drafting of the manuscript and approved the final version.
Conflicts of Interest
None declared.
Multimedia Appendix 2
Downs and Black Instrument assessment performed for each retrieved paper.
DOCX File, 469 KBMultimedia Appendix 3
Mixed Methods Appraisal Tool, version 2018. Assessment performed for each retrieved paper.
DOCX File, 413 KBReferences
- Chib A, Lin SH. Theoretical advancements in mHealth: a systematic review of mobile apps. J Health Commun. 2018;23(10-11):909-955. [CrossRef] [Medline]
- Torous J, Kiang MV, Lorme J, Onnela JP. New tools for new research in psychiatry: a scalable and customizable platform to empower data driven smartphone research. JMIR Ment Health. May 5, 2016;3(2):e16. [CrossRef] [Medline]
- What kinds of sensors are embedded in smartphones? [Internet]. Samsung SDS. 2024. URL: https://www.samsungsds.com/en/insights/what-kinds-of-sensors-are-embedded-in-smartphones.html [Accessed 2026-02-04]
- Perera C, Zaslavsky A, Christen P, Salehi A, Georgakopoulos D. Capturing sensor data from mobile phones using global sensor network middleware. Presented at: 2012 IEEE 23rd International Symposium on Personal, Indoor and Mobile Radio Communications - (PIMRC 2012); Sep 9-12, 2012:24-29; Sydney, NSW. URL: https://ieeexplore.org/document/6362778 [CrossRef]
- Pires IM, Garcia NM, Pombo N, Flórez-Revuelta F, Spinsante S. Approach for the development of a framework for the identification of activities of daily living using sensors in mobile devices. Sensors (Basel). Feb 21, 2018;18(2):640. [CrossRef] [Medline]
- Celik Y, Godfrey A. Bringing it all together: wearable data fusion. NPJ Digit Med. Aug 17, 2023;6(1):149. [CrossRef] [Medline]
- Draghici AE, Taylor JA. The physiological basis and measurement of heart rate variability in humans. J Physiol Anthropol. Sep 28, 2016;35(1):22. [CrossRef] [Medline]
- De Angel V, Lewis S, White K, et al. Digital health tools for the passive monitoring of depression: a systematic review of methods. NPJ Digit Med. Jan 11, 2022;5(1):3. [CrossRef] [Medline]
- Teo AR, Nelson S, Strange W, et al. Social withdrawal in major depressive disorder: a case-control study of hikikomori in japan. J Affect Disord. Sep 1, 2020;274:1142-1146. [CrossRef] [Medline]
- Leimhofer J, Petrovic M, Dominik A, Heider D, Hegerl U. Cross-platform availability of smartphone sensors for depression indication systems: mixed-methods umbrella review. Interact J Med Res. Aug 7, 2025;14:e69686. [CrossRef] [Medline]
- Cornet VP, Holden RJ. Systematic review of smartphone-based passive sensing for health and wellbeing. J Biomed Inform. Jan 2018;77:120-132. [CrossRef] [Medline]
- Mohr DC, Lyon AR, Lattie EG, Reddy M, Schueller SM. Accelerating digital mental health research from early design and creation to successful implementation and sustainment. J Med Internet Res. May 10, 2017;19(5):e153. [CrossRef] [Medline]
- Moshe I, Terhorst Y, Opoku Asare K, et al. Predicting symptoms of depression and anxiety using smartphone and wearable data. Front Psychiatry. 2021;12:625247. [CrossRef] [Medline]
- Saeb S, Zhang M, Karr CJ, et al. Mobile phone sensor correlates of depressive symptom severity in daily-life behavior: an exploratory study. J Med Internet Res. Jul 15, 2015;17(7):e175. [CrossRef] [Medline]
- Rohani DA, Faurholt-Jepsen M, Kessing LV, Bardram JE. Correlations between objective behavioral features collected from mobile and wearable devices and depressive mood symptoms in patients with affective disorders: systematic review. JMIR Mhealth Uhealth. Aug 13, 2018;6(8):e165. [CrossRef] [Medline]
- Choi A, Ooi A, Lottridge D. Digital phenotyping for stress, anxiety, and mild depression: systematic literature review. JMIR Mhealth Uhealth. May 23, 2024;12:e40689. [CrossRef] [Medline]
- Shen S, Qi W, Zeng J, et al. Passive sensing for mental health monitoring using machine learning with wearables and smartphones: scoping review. J Med Internet Res. Aug 14, 2025;27:e77066. [CrossRef] [Medline]
- Leaning IE, Ikani N, Savage HS, et al. From smartphone data to clinically relevant predictions: a systematic review of digital phenotyping methods in depression. medRxiv. Health Informatics. Preprint posted online on 2023. [CrossRef]
- Ettore E, Müller P, Hinze J, et al. Digital phenotyping for differential diagnosis of major depressive episode: narrative review. JMIR Ment Health. Jan 23, 2023;10:e37225. [CrossRef] [Medline]
- Ezawa ID, Hollon SD, Robinson N. Examining predictors of depression and anxiety symptom change in cognitive behavioral immersion: observational study. JMIR Ment Health. Jul 14, 2023;10:e42377. [CrossRef] [Medline]
- WHO guideline recommendations on digital interventions for health system strengthening. World Health Organization. 2019. URL: http://www.ncbi.nlm.nih.gov/books/NBK541902 [Accessed 2026-02-04]
- Roberts LW, Chan S, Torous J. New tests, new tools: mobile and connected technologies in advancing psychiatric diagnosis. NPJ Digit Med. 2018;1:20176. [CrossRef] [Medline]
- Arksey H, O’Malley L. Scoping studies: towards a methodological framework. Int J Soc Res Methodol. Feb 2005;8(1):19-32. [CrossRef]
- Tricco AC, Lillie E, Zarin W, et al. PRISMA Extension for Scoping Reviews (PRISMA-ScR): checklist and explanation. Ann Intern Med. Oct 2, 2018;169(7):467-473. [CrossRef] [Medline]
- Torous J, Linardon J, Goldberg SB, et al. The evolving field of digital mental health: current evidence and implementation issues for smartphone apps, generative artificial intelligence, and virtual reality. World Psychiatry. Jun 2025;24(2):156-174. [CrossRef] [Medline]
- Pearce M, Garcia L, Abbas A, et al. Association between physical activity and risk of depression: a systematic review and meta-analysis. JAMA Psychiatry. Jun 1, 2022;79(6):550-559. [CrossRef] [Medline]
- Nickels S, Edwards MD, Poole SF, et al. Toward a mobile platform for real-world digital measurement of depression: user-centered design, data quality, and behavioral and clinical modeling. JMIR Ment Health. Aug 10, 2021;8(8):e27589. [CrossRef] [Medline]
- Mandryk RL, Birk MV, Vedress S, et al. Remote assessment of depression using digital biomarkers from cognitive tasks. Front Psychol. 2021;12:767507. [CrossRef] [Medline]
- Rykov Y, Thach TQ, Bojic I, Christopoulos G, Car J. Digital biomarkers for depression screening with wearable devices: cross-sectional study with machine learning modeling. JMIR Mhealth Uhealth. Oct 25, 2021;9(10):e24872. [CrossRef] [Medline]
- Currey D, Torous J. Increasing the value of digital phenotyping through reducing missingness: a retrospective review and analysis of prior studies. BMJ Ment Health. Feb 2023;26(1):e300718. [CrossRef] [Medline]
- Su L, Anderson PL. User behavior of a publicly available, free-to-use, self-guided mHealth app for depression: observational study in a global sample. JMIR Form Res. Oct 25, 2022;6(10):e35538. [CrossRef] [Medline]
- Braund TA, O’Dea B, Bal D, et al. Associations between smartphone keystroke metadata and mental health symptoms in adolescents: findings from the future proofing study. JMIR Ment Health. May 15, 2023;10:e44986. [CrossRef] [Medline]
- Kathan A, Harrer M, Küster L, et al. Personalised depression forecasting using mobile sensor data and ecological momentary assessment. Front Digit Health. 2022;4:964582. [CrossRef] [Medline]
- Frank E, Wallace ML, Matthews MJ, et al. Personalized digital intervention for depression based on social rhythm principles adds significantly to outpatient treatment. Front Digit Health. 2022;4:870522. [CrossRef] [Medline]
- Zhang Y, Pratap A, Folarin AA, et al. Long-term participant retention and engagement patterns in an app and wearable-based multinational remote digital depression study. NPJ Digit Med. Feb 17, 2023;6(1):25. [CrossRef] [Medline]
- Langholm C, Kowatsch T, Bucci S, Cipriani A, Torous J. Exploring the potential of Apple SensorKit and digital phenotyping data as new digital biomarkers for mental health research. Digit Biomark. 2023;7(1):104-114. [CrossRef] [Medline]
- Abbas A, Sauder C, Yadav V, et al. Remote digital measurement of facial and vocal markers of major depressive disorder severity and treatment response: a pilot study. Front Digit Health. 2021;3:610006. [CrossRef] [Medline]
- Sverdlov O, Curcic J, Hannesdottir K, et al. A study of novel exploratory tools, digital technologies, and central nervous system biomarkers to characterize unipolar depression. Front Psychiatry. 2021;12:640741. [CrossRef] [Medline]
- Sun S, Folarin AA, Zhang Y, et al. Challenges in using mHealth data from smartphones and wearable devices to predict depression symptom severity: retrospective analysis. J Med Internet Res. Aug 14, 2023;25:e45233. [CrossRef] [Medline]
- Goltermann J, Emden D, Leehr EJ, et al. Smartphone-based self-reports of depressive symptoms using the Remote Monitoring Application in Psychiatry (ReMAP): interformat validation study. JMIR Ment Health. Jan 12, 2021;8(1):e24333. [CrossRef] [Medline]
- Tsirmpas C, Andrikopoulos D, Fatouros P, et al. Feasibility, engagement, and preliminary clinical outcomes of a digital biodata-driven intervention for anxiety and depression. Front Digit Health. 2022;4:868970. [CrossRef] [Medline]
- Ben-Zeev D, Chander A, Tauscher J, et al. A smartphone intervention for people with serious mental illness: fully remote randomized controlled trial of CORE. J Med Internet Res. Nov 12, 2021;23(11):e29201. [CrossRef] [Medline]
- Suen YN, Yau JY, Wong PS, et al. Effect of brief, personalized feedback derived from momentary data on the mental health of women with risk of common mental disorders in Hong Kong: a randomized clinical trial. Psychiatry Res. Nov 2022;317:114880. [CrossRef] [Medline]
- Downs and black checklist (DBC) [internet]. Reach Out and Read. 2024. URL: https://reachoutandread.org/metric/dbc/ [Accessed 2026-02-04]
- Hong QN, Fàbregues S, Bartlett G, et al. The mixed methods appraisal tool (MMAT) version 2018 for information professionals and researchers. EFI. 2018;34(4):285-291. [CrossRef]
- Martin-Key NA, Spadaro B, Funnell E, et al. The current state and validity of digital assessment tools for psychiatry: systematic review. JMIR Ment Health. Mar 30, 2022;9(3):e32824. [CrossRef] [Medline]
- Kang E, Kim S, Rhee YE, Lee J, Yun YH. Self-management strategies and comorbidities in chronic disease patients: associations with quality of life and depression. Psychol Health Med. Sep 2021;26(8):1031-1043. [CrossRef] [Medline]
- Currey D, Torous J. Digital phenotyping data to predict symptom improvement and mental health app personalization in college students: prospective validation of a predictive model. J Med Internet Res. Feb 9, 2023;25:e39258. [CrossRef] [Medline]
- Laban G, Morrison V, Kappas A, S Cross E. Coping with emotional distress via self-disclosure to robots: an intervention with caregivers. Int J Soc Robot. 2025;17(9):1837-1870. [CrossRef] [Medline]
Abbreviations
| BDI: Beck Depression Inventory |
| BDI-II: Beck Depression Inventory, Second Edition |
| CES-D: Center for Epidemiological Studies Depression Scale |
| CSSRS: Columbia Suicide Severity Rating Scale |
| DASS-21: Depression Anxiety Stress Scale Short Form 21 |
| HDRS: Hamilton Depression Rating Scale |
| ICTRP: International Clinical Trials Registry Platform |
| MADRS: Montgomery-Åsberg Depression Rating Scale |
| MINI: Mini-International Neuropsychiatric Interview |
| MMAT: Mixed Methods Appraisal Tool |
| PHQ-2: Patient Health Questionnaire-2 |
| PHQ-8: Patient Health Questionnaire-8 |
| PHQ-9: Patient Health Questionnaire-9 |
| PRISMA-ScR: Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews |
| PROSPERO: International Prospective Register of Systematic Reviews |
| RAS: Recovery Assessment Scale |
| RSES: Rosenberg Self-Esteem Scale |
| SDS: Sheehan Disability Scale |
| WHO: World Health Organization |
Edited by Stefano Brini; submitted 04.Jan.2025; peer-reviewed by Acacia Parks, Alireza Sameh; accepted 21.Dec.2025; published 02.Mar.2026.
Copyright© Lisa Busshart, Milica Petrovic, Rebeka Amin, Ulrich Hegerl. Originally published in JMIR mHealth and uHealth (https://mhealth.jmir.org), 2.Mar.2026.
This is an open-access article distributed under the terms of the Creative Commons Attribution License (https://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work, first published in JMIR mHealth and uHealth, is properly cited. The complete bibliographic information, a link to the original publication on https://mhealth.jmir.org/, as well as this copyright and license information must be included.

