Elmer et al. (2025)
A Social Support Just-in-Time Adaptive Intervention for Individuals With Depressive Symptoms
Elmer T, Wolf M, Snippe E, Scholz U · JMIR Mental Health, 12:e74103 · 2025
A smartphone-delivered social support JITAI proved technically feasible and well-tolerated among psychotherapy-seeking individuals with depressive symptoms. Crucially, interventions triggered by participants’ own expressed need for support were rated as significantly more timely, helpful, and effective than those triggered by fixed or personalized distress thresholds — suggesting that subjective receptivity matters more than inferred vulnerability.
Research Question
Is a social support JITAI — one that encourages depressed individuals to activate their social networks during moments of distress — technically feasible, acceptable to users, and capable of prompting support-seeking behavior? And which triggering strategy (fixed cutoff, personalized Shewhart control chart, or self-reported support need) works best?
The Gap This Study Addresses
Existing mental health JITAIs focus primarily on mindfulness, self-monitoring, or CBT elements. None had specifically targeted social support activation — despite robust evidence that social support is a key protective factor against depression. Individuals waiting for psychotherapy (sometimes months) receive little structured support during that critical period.
The Intervention Design
When triggered, the JITAI walked participants through three steps via the m-Path smartphone app: (1) reflect on what type of support would be helpful right now, (2) view a list of identified personal support figures to broaden awareness of available resources, and (3) receive one of six evidence-based support-seeking strategies (e.g., articulating needs clearly, expressing gratitude, diversifying providers). Delivery was randomized across four triggering conditions using a microrandomized trial design.
Main Results at a Glance
High compliance (85.37% of EMA surveys completed; 80% of participants met the ≥70% threshold) and negligible careless responding (1.5%). Participants adopted the support-seeking behavior in roughly one-third of triggered instances. The “support need” triggering condition outperformed distress-based triggers on timing appropriateness, helpfulness, and behavior adoption — with Cohen d effect sizes averaging −0.69. Effects on distress reduction were small (d = 0.06–0.14) and non-significant, as expected in a feasibility study.
Triggering Condition Comparison
| Condition | Trigger Logic | N Triggered | Appropriate Timing (mean) | Behavior Adoption |
|---|---|---|---|---|
| 1 — Fixed Cutoff | Any distress item ≥ 5 on 7-pt scale | 221 (58.6%) | 2.98 / 7 | 23% |
| 2 — Personalized SPC | Any distress item exceeds individual UCL | 110 (29.2%) | 2.82 / 7 | 22% |
| 3 — Support Need | Participant answers “yes” to needing support | 46 (12.2%) | 4.41 / 7 ✓ | 43% ✓ |
| 4 — Control | No JITAI delivered (condition met but omitted) | — | — | — |
■ JITAI Concepts ■ Research Methods ■ Clinical Context ■ Outcomes & Theory
JITAI Concepts
Research Methods
Clinical Context
Outcomes & Theoretical Frameworks
Study Design at a Glance
Preregistered feasibility study embedding a microrandomized trial (MRT). Over 21 days, participants received 6 daily EMA surveys via smartphone. For the final 18 days, a JITAI was randomly assigned to one of four triggering conditions at each EMA time point. Outcome data were collected via post-EMA questions, evening surveys, and weekly surveys. Qualitative interviews preceded and followed the EMA phase.
Participants
- N = 25 completing participants (of 130 screened; 36 eligible)
- Seeking outpatient psychotherapy; BDI-II score > 9 (minimal to severe depressive symptoms)
- Exclusions: suicidal ideation (BDI-II item 9 > 2), manic symptoms, therapy session within 4 weeks, night shift workers, age < 18 or > 70
- 88% women; mean age 35.1 years (SD 11.4); 84% self-identified as White
- Depression distribution: 12% minimal, 48% mild, 28% moderate, 12% severe
EMA Protocol
- 126 surveys over 21 days (6/day, randomly prompted in 2-hour windows: 8 AM–10 PM)
- Distress items: negative affect, stress, loneliness, rumination — all on 7-point Likert scales
- Support need: single-item yes/no/no worries question
- Additional evening survey (8–9 PM daily) and weekly surveys (days 7, 14, 21)
- Days 1–3: baseline EMA only (no JITAI); SCC algorithm “learns” from days 1–7
The Four Triggering Conditions (Microrandomized)
- Condition 1 — Fixed Cutoff: Any distress item ≥ 5 (disjunctive rule). Simple, group-level threshold.
- Condition 2 — Personalized SPC: Any distress item exceeds individual UCL = phase-1 mean + 2σ̂. Personalized to each participant’s natural variability.
- Condition 3 — Support Need: Participant responds “yes” to needing support right now. Direct self-report trigger.
- Condition 4 — Control: No JITAI delivered, even if triggering criteria are met (enables causal comparison).
- Cap: maximum 2 JITAIs per day per participant (76 JITAIs were withheld due to this cap).
Intervention Content (When Triggered)
- Step 1: Participant specifies what type of social support would be helpful
- Step 2: App shows list of identified personal support figures (broadens awareness of available resources)
- Step 3: Participant receives one of six evidence-based support-seeking strategies (articulate needs, reframe situation, express gratitude, foster reciprocation, diversify providers)
- Channel: participant’s own choice — call, text, video chat, or meet in person
Feasibility Outcomes Measured
- Appropriate timing of JITAI (post-EMA Likert)
- Helpfulness (evening survey Likert)
- Behavior adoption — “Did you seek support because of the app?” (yes/no, post-EMA)
- Intervention engagement (weekly Likert)
- Burden (weekly Likert — “I don’t mind doing another week”)
- Technical functioning (weekly Likert)
- Negative effects of study participation and of seeking support (weekly Likert)
- EMA compliance rate and attention-check careless responding
Engagement & Compliance Metrics
Nahum-Shani et al. (2018) — The JITAI Framework
Foundational Theory JITAI DesignKlasnja et al. (2015) — Microrandomized Trials
Study Design MethodsSchat et al. (2023) — SPC Methods for EMA Data
Statistical Method PersonalizationCohen S. (2004) — Social Support and Health
Social Support TheoryRafaeli & Gleason (2009) — Skilled Support
Intervention TheoryCutrona & Russell (1990) — Optimal Matching Theory
Theoryvan Genugten et al. (2025) — JITAIs in Mental Health: Systematic Review
Systematic Review Field OverviewMontgomery, D.C. (2009) — Introduction to Statistical Quality Control
Statistical Methods SCC FoundationsA smartphone JITAI designed to activate social support networks is technically feasible and acceptable among depressed individuals awaiting psychotherapy. However, when individuals are asked directly whether they want support — rather than inferred from distress signals — the intervention is more timely, more helpful, and more likely to produce behavior change. Subjective need and receptivity appear to matter more than objectively detected vulnerability.
-
Technical Feasibility Is Demonstrated
85.4% EMA compliance, 80% of participants exceeding the ≥70% threshold, and only 1.5% careless responses confirm that intensive data collection via smartphone is achievable in this clinical population. Burden ratings were low and stable across all three study weeks — suggesting the JITAI could be integrated into daily life without undue strain.
-
Self-Reported Need Outperforms Distress-Based Triggers
Across all user-rated outcomes — timing appropriateness, helpfulness, and behavior adoption — interventions triggered by the participant’s own expressed need for support significantly outperformed both fixed cutoff and personalized SPC triggers. Average Cohen d was −0.69 across these outcomes, a meaningful effect for a feasibility study.
-
Distress Detection and Self-Perceived Need May Be Complementary
Despite lower user ratings, distress-based triggers produced a numerically larger (though non-significant) effect on actual support-seeking behavior — suggesting they may capture moments of implicit need that individuals don’t recognize or articulate. Future JITAIs may benefit from combining both: detect distress, but also assess readiness to act.
-
Receptivity Is an Underexplored JITAI Variable
The authors raise a critical design question: should decision rules focus on vulnerability (is the person distressed?) or receptivity (is the person ready and able to act?) — or both? Most existing JITAIs only address vulnerability. Social support interventions in particular may depend on contextual readiness: having time, perceiving support providers as available, and feeling willing to reach out.
-
Behavior Adoption Remains the Key Challenge
Overall adoption was just 29%: participants sought social support in fewer than one-third of instances post-JITAI. Qualitative data pointed to time constraints and perceived unavailability of support providers as key barriers. Future iterations should integrate real-time availability signals (does the participant have time? are support figures likely available?) into the triggering logic.
-
A Low-Threshold Bridge for the Therapy Waitlist Period
The study directly targets a genuine system-level gap: individuals waiting weeks or months for psychotherapy receive little structured support. A social support JITAI represents a scalable, low-entry-barrier approach that could complement care — not replace it. Its theoretical grounding in skilled support and optimal matching frameworks gives it conceptual coherence; confirming effectiveness at scale remains the next step.
What the study actually did — and what it actually found. A plain-language guide to the design rationale, measurement instruments, triggering logic, and an honest reading of the results.
The core idea
Most mental health JITAIs focus on mindfulness or cognitive behavioral therapy elements delivered through a smartphone. This study did something different: it tried to use a JITAI to nudge depressed individuals — specifically those waiting for psychotherapy — to reach out to people in their social networks during difficult moments. The underlying logic is well-supported: social support is a genuine protective factor against depression, and the waiting period before therapy is a critical window where people receive little structured help. The question was whether a smartphone prompt could bridge that gap.
How the triggering worked — and why it matters
The study tested four different “alarm systems” for deciding when to send the prompt. Understanding the difference between them is essential to understanding the results.
Conditions 1 and 2 are the distress-based triggers. They watch the participant’s EMA survey answers and try to infer that the person needs support without asking them directly. To understand what that actually means, it is worth looking at what the EMA survey consisted of in practice.
Six times a day, at a random moment within a two-hour window, the participant’s phone buzzed with a notification from the m-Path app. When opened, it presented a short series of single questions. The four distress questions were each answered on an explicitly 7-point Likert scale:
“To what extent do you experience negative emotions at the moment?”
1 (I do not experience negative emotions) — 7 (very negative)
“I feel stressed at the moment.”
1 (disagree completely) — 7 (agree completely)
“I feel lonely at the moment.”
1 (disagree completely) — 7 (agree completely)
“I realize that I am thinking the same negative thoughts over and over again.”
1 (disagree completely) — 7 (agree completely)
“Would it help you right now to talk to someone about your worries or negative feelings?”
Not a sliding scale — three options only: Yes · No, it would not help me · I have no worries or negative feelings
That was essentially the entire survey. Four slider questions and one three-option question, delivered six times a day — sometimes while the participant was walking, eating, or half-distracted. Each question answered in seconds. And yet those numbers were being fed directly into algorithms that decided whether to send a mental health intervention.
This is worth sitting with. The fixed cutoff in Condition 1 set its threshold at ≥ 5 on that 7-point scale — which is actually the upper third of the scale, not the midpoint. A participant had to be reporting fairly high distress before it fired. That makes the poor timing appropriateness ratings for Condition 1 even more interesting: participants were scoring 5 or above on distress and still reporting that the intervention arrived at the wrong moment. High distress and readiness to seek support, it turns out, are not the same thing.
Condition 2 uses personalized Shewhart control charts — one per distress variable, per person. During the first seven days, the algorithm learns each participant’s personal baseline and natural variability for each of the four distress variables independently. It then calculates a personal upper control limit (UCL) for each one. From day 4 onward, the JITAI fires whenever any distress score exceeds that person’s own UCL — not a group-level threshold, but a deviation from their own normal. A person who routinely scores 6 on stress will not be triggered by a 6; a person who normally scores 3 will be. The formulas underpinning this personalization come from Montgomery’s foundational statistical quality control textbook, adapted for psychological EMA data by Schat et al. (2023).
Condition 3 takes a completely different approach. Rather than watching scores and inferring need, the EMA survey simply asks the support need question described above. If the answer is yes, the JITAI fires. The participant themselves is the trigger.
Condition 4 is the control: no JITAI is sent, even if the criteria for one of the other three conditions would have been met. This is what makes causal comparison possible.
The randomization logic — and why it was necessary
At every single one of the 6 daily EMA surveys, the system randomly assigned that moment to one of the four conditions with equal probability — 25% each. This happened regardless of how distressed the participant was or what time of day it was.
This design choice is the methodological backbone of the study. If the researchers had simply let each condition fire whenever its criterion was met, the conditions would never be fairly comparable. Condition 1 fired 221 times; Condition 3 fired only 46 times. Without randomization, any difference in outcomes between them could simply reflect the fact that Condition 3 moments were rarer and therefore more personally meaningful — not that the triggering strategy itself was better.
By randomizing the condition assignment first, the researchers broke that link. And Condition 4 in particular is what enables causal inference: because the control condition is randomly assigned, you occasionally get moments where a participant is highly distressed and would have met the criteria for an active condition — but no JITAI was sent. Those moments become the counterfactual: what would have happened if nothing had been done? Without random assignment you can never answer that question cleanly.
The baseline learning problem
Condition 2 required a learning phase before it could function properly — seven days of EMA data to estimate each participant’s upper control limit. But the researchers started the intervention phase on day 4, not day 8, because asking depressed individuals awaiting therapy to go two full weeks without any intervention felt ethically uncomfortable and would likely have hurt compliance. This means the Shewhart control chart UCLs were still being estimated during days 4 through 7 and only stabilized after day 7. It was a deliberate tradeoff: personalization accuracy sacrificed for participant welfare and retention.
Why six surveys a day — and what that frequency costs
The paper presents six EMA surveys per day as a given rather than a decision requiring justification, but there is an established methodological rationale behind it. Six is a commonly used frequency in EMA research on mood and daily life, and the specific implementation here matters: surveys were not sent at fixed times but at one random moment within each of six consecutive two-hour windows spanning 8 AM to 10 PM. This stratified random sampling approach ensures coverage across the whole waking day while preventing participants from anticipating the next survey — which matters because predictable timing would allow participants to prepare answers mentally in advance, undermining the “momentary” quality of the assessment.
The frequency of six is a tradeoff between three competing pressures. More surveys per day gives finer-grained emotional data and more decision points for the JITAI — especially important for Condition 2, whose algorithm needs sufficient observations to estimate personal baselines reliably. But more surveys also increase burden, fatigue, and the risk of careless responding or dropout. Six has become a pragmatic consensus in the field: enough to capture meaningful within-day emotional variation without overwhelming participants.
What the paper does not address is whether six was the right number for this specific population. Individuals with depressive symptoms may experience survey fatigue differently from healthy participants, and 126 surveys over 21 days is not a trivial ask. The fact that 80% of participants met the ≥70% compliance threshold is reassuring — but 20% did not, and one participant dropped out specifically citing lifestyle incompatibility with the survey schedule. There is also a subtler risk: with six decision points per day over three weeks, participants may gradually learn to recognize the triggering logic — noticing patterns in when the app responds and when it does not. That kind of learning could alter how honestly they answered distress questions, which would undermine the entire measurement system the JITAI depends on. The study does not resolve this tension.
What the support need condition revealed
The support need condition (Condition 3) outperformed both distress-based conditions on every user-rated measure: timing appropriateness (4.41 vs. 2.82–2.98 out of 7), helpfulness, and behavior adoption (43% vs. 22–23%). The average effect size across these outcomes was Cohen d = −0.69, which is meaningful.
The reason appears to be that Condition 3 captured not just vulnerability (being distressed) but receptivity — whether the participant was actually ready and willing to act. A participant can be stressed or lonely and have no interest in calling anyone: they might be in a meeting, the source of their stress might be a relationship conflict, or they simply might not feel ready. Distress and readiness to seek support are not the same thing, and the distress-based triggers were conflating the two.
However, there is an important counterpoint. When looking at actual subsequent support-seeking behavior rather than user ratings, Condition 1 (fixed cutoff) showed the largest numerical effect — suggesting that distress detection may catch moments of implicit need that the participant has not consciously recognized or articulated. So the two approaches may not be competing so much as complementary: one captures readiness, the other captures need the person has not yet named.
Who was in the study — and why it matters
The study was conducted at the University of Zurich in Switzerland. Of the 25 participants, 88% self-identified as women, 12% as men, and 84% as White. Mean age was 35.1 years. Participants were recruited through psychotherapist referrals and social media posts.
The paper acknowledges the gender skew as a limitation, noting that gender differences in social support preferences and help-seeking behavior may affect engagement and effectiveness. However it does not address the ethnic homogeneity as a limitation at all — which is itself notable.
For a feasibility study conducted in Zurich, the sample is arguably representative of the population the intervention was designed for. Switzerland has specific structural features relevant to the study’s premise: long therapy waiting times, high smartphone penetration, and a particular cultural context around help-seeking. Within that context the demographics are not surprising.
But the skew becomes a problem the moment anyone considers scaling or generalizing the intervention. Social support norms vary enormously across cultures — who you turn to, how you ask, whether seeking help carries stigma, and what kinds of support are culturally legible. An intervention built on the assumption that participants have an accessible, willing social network and feel comfortable activating it may work very differently in populations where those assumptions do not hold.
The honest bottom line
On the outcome that actually matters clinically — whether participants felt better — the JITAI produced no detectable effect. The effect sizes on distress reduction are worth examining directly against conventional benchmarks:
These are not numbers that failed to reach significance because the sample was small. They are genuinely tiny. The paper defends this by noting it was never designed to test efficacy — which is true. But the 29% behavior adoption rate gives reason to question the mechanism itself. Only one in three participants actually sought support after receiving the JITAI, even under the most favorable possible study conditions: motivated participants, financial compensation, researcher contact, and qualitative interviews framing the whole experience. If the core chain — prompt leads to support-seeking, support-seeking leads to feeling better — only activated partially even here, scaling this up to a real-world deployment raises serious questions.
The app worked technically, participants tolerated it reasonably well, and the support need condition showed genuine promise as a triggering strategy. But there is currently no evidence that it makes depressed people feel meaningfully better — and the conditions under which it was tested were about as favorable as they could realistically be. Those questions are sharpened by the demographic profile: if 29% adoption is the ceiling under optimal conditions with an optimal population, the intervention has a significant burden of proof to meet before it can claim broader relevance.
