Effectiveness of the Minder Mobile Mental Health and Substance Use Intervention for University Students: Randomized Controlled Trial


IntroductionBackground

University attendance is a transitional period in which many students experience novel stressors related to moving away from home, navigating new social environments, and managing increased educational and financial demands in the absence of their traditional support systems [,]. The transition to attending university also coincides with the peak period of onset of many mental disorders, including mood, anxiety, and substance use disorders [,]. Studies have documented the high rates of mental health and substance use problems experienced by university students [], with research also indicating that students with preexisting mental health problems can experience a worsening of their conditions following the transition to attending university [].

Despite this high need, most students experiencing mental health problems do not receive treatment []. Research investigating help seeking among university students indicates that, compared to structural barriers, attitudinal barriers are the most important reasons for not seeking help []. The most commonly cited reason students give for not seeking help is a preference for handling things on their own [,]. One way of adapting interventions to align with this preference is to provide students with tools that are self-guided and allow them autonomy over how and when to use the tools provided. e-Interventions can be accessed by users at any time and have been demonstrated to be effective in improving various mental health [] and substance use outcomes among university students []. These interventions have also been identified as key components in proposed models of care for universities []. While much of the literature on e-interventions has been focused on web-based tools, mobile apps have been identified as a promising means of delivering mental health interventions due to not only the increase in smartphone use but also the wide range of interventions that can be delivered through mobile platforms [,].

Given the range of challenges faced by university students, including the high rates of disorder-level and subclinical mental health and substance use problems, transdiagnostic approaches to early intervention and prevention may be beneficial for this population []. Developing this type of intervention requires the use of a holistic, student-centered design approach to identify evidence-based condition-specific and cross-cutting opportunities for intervention. Furthermore, the intervention needs to be aligned with the perceived needs and preferences of students to ensure meaningful engagement []. On the basis of these requirements, we codeveloped a mental health and substance use mobile app called Minder for Canadian university students. This participatory codevelopment process involved significant input from students through the creation of a Student Advisory Committee, usability testing via a virtual boot camp (ie, individual user-testing combined with a web-based survey), focus groups, and a pilot feasibility study [].

Objectives

The objective of this study was to test the effectiveness of the Minder mobile app in improving mental health and substance use outcomes in a general population of university students.


MethodsTrial Design

This study was based on a 2-arm, parallel-assignment, single-blinded (the statistician was blinded), 30-day randomized controlled trial with 1 intervention group and 1 waitlist control group. The study was registered at ClinicalTrials.gov (NCT05606601), and a full study protocol has been published []. No significant changes to the trial protocol or intervention content were made during the trial period; however, several minor adjustments, along with a description of minor technical issues, can be found in .

Ethical Considerations

Ethics approval was obtained from the University of British Columbia (UBC) Behavioural Research Ethics Board on January 6, 2022 (ethics ID: H21-03248). Informed consent was obtained through a web-based self-assessment questionnaire at the beginning of the study. Participants were informed of their ability to opt out at any point within the study by emailing the research team. Identifiable data were stored in data files within the app backend, which were separate from all deidentified app use and survey data. This information can only be linked using a unique study ID number. Participants received a CAD $10 (US $7.4) gift card for completion of the baseline survey and an additional CAD $10 (US $7.4) gift card for completion of the 30-day follow-up survey.

Participants

The study was conducted at the UBC Point Grey (Vancouver, British Columbia, Canada) campus. Participants needed to confirm their eligibility using a web-based self-assessment questionnaire before registering and consenting to the study. The inclusion criteria were as follows: students currently enrolled at the UBC Vancouver campus, aged ≥17 years, having access to and being able to use a smartphone with Wi-Fi or cellular data, and speaking English. The only exclusion criterion was based on a single screening question assessing suicidality risk (“We want to make sure that this app is appropriate for you at this time. Do you have a current suicidal plan [i.e., a plan to end your life]?”). Anyone endorsing a current suicidal plan (ie, answering “yes”) was prevented from registering and was instead provided with a list of local crisis resources. The eligibility criterion of being a current UBC student was confirmed using a unique student log-in checkpoint as part of the registration process. This process also ensured that each student could only enroll once.

Recruitment and Consent

Given the large sample size needed for this study, many different recruitment methods were used. Online recruitment occurred through various social media platforms and a linked ongoing Student E-Mental Health trend study []. Recruitment also occurred through in-person and on-campus engagements, such as setting up informational booths at the university, displaying posters about the study, visiting in-person and online classes, having professors share study information with their classes, and contacting student groups to share information with their members. Paid bus and bus stop advertisements at the university were also used. A more detailed description of the recruitment methods can be found in the study protocol for this trial [].

Participants’ consent was obtained using Qualtrics (Qualtrics International Inc), a web-based form, after completing the eligibility screening. The consent form indicated that participants would either gain access to the full app immediately or in 30 days following completion of the final survey. Individual accounts were created for each participant and sent to them with a link to download the app. Upon downloading the app and completing the baseline survey, participants were randomly assigned through the app to the intervention group, which received full access to the Minder app, or to the control group, which only had access to a restricted version of the app that included a short introduction video and the baseline and follow-up surveys. Participants received a CAD $10 (US $7.40) gift card for completion of the baseline survey and an additional CAD $10 (US $7.40) gift card for completion of the 30-day follow-up survey; however, the use of the app itself was not remunerated.

Randomization and Intervention

Participants were randomized using a custom-developed automated process incorporated directly into the mobile app following completion of the baseline survey. The system assigned participants to either the intervention or control group using a predetermined block randomization list (1:1 randomization in blocks of 10) stratified for past drug use (any lifetime use of opioids or nonmedical stimulants). Stratification by past drug use was used to account for the low number of students using these substances and the need to ensure that they were evenly distributed across the study groups. The randomization lists (1 for each stratification group) were generated using the web-based stratified block randomization list creator in the clinical trial software Sealed Envelope (Sealed Envelope Ltd) []. The intervention and control groups completed the main assessments of the primary and secondary outcomes at baseline and the 30-day follow-up. The intervention group was also prompted to complete a short survey at 2 weeks that consisted of a limited set of questions on anxiety and depression symptoms.

The Minder mobile app was codeveloped with university students and professionals with the goal of creating an engaging self-directed tool for students to improve their mental health and manage substance use. The intervention is designed for a general population of students and, thus, addresses a wide range of challenges related to postsecondary student life, including managing emotions, relationships, well-being, and university life. The self-directed nature of the app also allows students to access features when needed. The codevelopment process consisted of ongoing student input through student staff members and volunteers along with several phases of purposeful student engagement and feedback. Further details on the codevelopment process can be found in the study by Vereschagin et al [].

Participants who were randomized to the intervention group were given full access to the Minder app and instructed to use it as they wanted. They were also presented with a tutorial video outlining the different features of the app. The Minder intervention consists of 4 main components: Chatbot Activities, Services, Community, and Peer Coaching. The chatbot activities consist of an automated preprogrammed chatbot that delivers evidence-based messages and videos. The content is based primarily on cognitive behavioral therapy and psychoeducation; however, there is also content adapted from dialectical behavioral therapy, mindfulness, metacognitive training, and motivational interviewing. The content sections are presented on a home page map with several islands: University Life, Wellbeing, Relationships, Sadness, Stress & Anxiety, and Substance Use (). There is also an Explore Chat located on the home page map that guides participants to select an activity that may be relevant to their current needs. A full list of the content included can be found in []. Most of the chat activities also contain a summary page that is unlocked upon completion of the chat activity and allows participants to review content they learned at a later time. In addition, several chat activities contain specific practice components that also become unlocked after the main activity is complete.

The Peer Coaching component consists of trained volunteers assigned to each participant. These peer coaches reach out to participants at the beginning of the trial and midway through. They can provide support in navigating the app or nonclinical peer support based on active listening and problem-solving. Peer coaches can communicate with participants through an in-app chat asynchronously or synchronously through scheduled appointments delivered over in-app chat message or audio or video call. Before engaging with peer coaches, participants must provide a phone number that can be used to contact them in a crisis situation and affirm that they are not currently at risk of self-harm or having suicidal thoughts, are not under the influence of substances, and understand the circumstances in which confidentiality would need to be broken (ie, crisis situations or abuse of a minor; A).

Figure 1. Minder home screen and content maps. The home screen contains 6 topic islands that users can select from. Each island leads to a separate map with chatbot activities and summaries. Additional features include the SOS button and the Explore Chat that guides users to activities. Figure 2. Peer coach, services, and community components. (A) The peer coach component allows users to connect with trained student volunteers, (B) the services component matches users with resources, and (C) the community component lists student groups by interest.

The Services component consists of a 10-question survey tool that provides participants with recommendations for resources based on their current needs and preferences (B). The survey tool and recommendations were adapted from a previously developed tool for university students [] and can be completed multiple times to receive new recommendations. Recommendations are provided for 6 areas related to student well-being: mental health and relationships, substance use, abuse, sexual wellness, housing, and education and activities. An additional safety component was added so that participants who based on the services survey were considered to be at high risk of suicidality (ie, plan for suicide or recent attempt and thoughts of hurting others) were asked to consent to provide their contact information and receive an expedited appointment with the university counseling services.

The Community component consists of a searchable directory of student groups or clubs at the university that are sorted by interest (eg, volunteering, arts, and advocacy; C).

The Minder app also contains several other general features. An SOS button appears in the top corner of the home page and provides a list of crisis resources if needed (). The settings page allows participants to update their username and password as well as change their avatar. Several types of push notifications were delivered through the app. General notifications were sent on days 4, 18, and 24. Reminders to complete the 2-week and 30-day follow-up surveys were sent as push notifications and via automated email reminders on those dates. Additional email reminders were sent on days 35 and 41 to remind participants to complete the follow-up survey if they had not done so already.

Participants who were randomized to the control group had access to a locked version of the app that only allowed them to complete the baseline survey and view a short introduction video that appears before the log-in screen. Following completion of the baseline survey, they received a pop-up message telling them that they would be notified when it was time to complete the next survey. The app was then locked so that control participants were not able to access any other areas of the app. At 30 days, participants were notified that it was time to complete the 30-day follow-up survey within the app, and this survey became unlocked. The app provided push notifications and automated email reminders to complete the follow-up survey on day 30 as well as additional email reminders on days 35 and 41 if the survey had not yet been completed.

Individuals who consented to participate in the study and received an invitation email to download the app but did not complete the baseline survey received additional email reminders at 7 days, at approximately 20 days, and several months later following the creation of their account.

The participants’ use of the app components was recorded through the app back-end system. This included starting each of the chatbot activities, completing the services survey and receiving recommendations, viewing community groups, and communicating with a peer coach.

Measures

The main assessments of the primary and secondary outcomes were collected through self-assessment within the Minder app at baseline and at the 30-day follow-up. The 30-day follow-up survey had to be completed within 44 days of the beginning of the baseline survey (30 days plus 2 weeks to accommodate the use of reminders) for the participants to be included in the analysis.

Primary Outcomes

The primary outcomes assessed in this study were changes in general anxiety symptomology, depressive symptomology, and alcohol consumption risk from baseline to follow-up at 30 days. All outcomes were assessed using self-report questionnaires completed directly in the mobile app.

Anxiety symptoms were assessed using the 7-item General Anxiety Disorder scale (GAD-7) assessment—a commonly used self-report scale that assesses symptoms of generalized anxiety []. Each GAD-7 question is scored from 0 (not at all) to 3 (nearly every day), with total scores ranging from 0 to 21 and higher scores indicating a worse outcome (ie, greater frequency of anxiety symptoms).

Depressive symptoms were assessed using the 9-item Patient Health Questionnaire (PHQ-9) self-report scale []. Each of the 9 questions is scored from 0 (not at all) to 3 (nearly every day). The total scores range from 0 to 29, with higher scores indicating a worse outcome (ie, a greater frequency of depressive symptoms).

Alcohol consumption risk was assessed using the US Alcohol Use Disorders Identification Test–Consumption Scale (USAUDIT-C) []. The USAUDIT-C is a 3-item self-report scale adapted from the consumption questions in the Alcohol Use Disorders Identification Test (AUDIT) []. Compared to the AUDIT, the USAUDIT-C includes expanded response options for the first 3 AUDIT questions—from 5 to 7 categories—to allow for more precise measurements when accounting for differences in standard drink sizes and cutoff limits. The higher the total score on the USAUDIT-C, the greater the respondent’s alcohol consumption and related risk [].

Secondary Outcomes

A range of secondary outcomes, including reduced use of other substances and additional mental health constructs that the Minder app was theorized to affect, were also assessed. For the purposes of this study, we focused on examining changes from baseline to follow-up at 30 days in frequency of substance use (cannabis, alcohol, opioids, and nonmedical stimulants) and mental well-being.

Frequency of cannabis use was assessed using a single self-report question on frequency of cannabis consumption in the previous 30 days. The 3 questions on the USAUDIT-C assessed unique dimensions of alcohol consumption. Frequency of alcohol use was assessed using responses to the first question on the USAUDIT-C, which asks how often participants have a drink containing alcohol. The number of drinks consumed in a typical drinking session was assessed using the second question on the USAUDIT-C, which asks participants how many drinks containing alcohol they have on a typical day when drinking. Frequency of binge drinking was assessed using the third question in the USAUDIT-C, which asks participants how often they have ≥5 (if sex at birth was male) or ≥4 (if sex at birth was female) drinks on 1 occasion.

Frequency of any opioid use in the previous 30 days was assessed using self-reported questions that asked about any pharmaceutical opioid (eg, oxycodone; morphine; hydromorphone; meperidine; fentanyl patches; and codeine or codeine-containing products such as Tylenol 1, 2, or 3) with a physician’s prescription and taken as prescribed; any pharmaceutical opioid (eg, oxycodone; morphine; hydromorphone; meperidine; fentanyl patches; and codeine or codeine-containing products such as Tylenol 1, 2, or 3) either without a physician’s prescription or in larger doses than prescribed to get high, buzzed, or numbed out; and any street opioid (eg, heroin and fentanyl) or any other opioid obtained “on the street.” The final opioid use outcome was defined as the most frequent number among any prescribed, nonprescribed, and street opioids.

Frequency of nonmedical stimulant use in the previous 30 days was assessed using self-report questions that asked about the frequency of using any street stimulant (eg, cocaine, crack, methamphetamines, and crystal meth) or prescription stimulant (eg, amphetamine, methylphenidate, and modafinil) either without a physician’s prescription or in larger doses than prescribed to get high, buzzed, numbed out, or help them study or for any other reason. The final nonmedical stimulant use outcome was defined as the most frequent number among any prescription stimulant without a prescription or not as prescribed and any street stimulant. The exact wording and response options for the substance use secondary outcome assessments can be found in .

Mental well-being was assessed using the Short Warwick-Edinburgh Mental Wellbeing Scale. The Short Warwick-Edinburgh Mental Wellbeing Scale is a 7-item scale that has been widely validated [], with total scores ranging from 7 to 35 and higher scores indicating a better outcome (ie, higher positive mental well-being).

Sample Size Estimation

The a priori sample size calculation for a small effect assessed using the PHQ-9, GAD-7, and USAUDIT-C was performed using an effect size defined by Cohen d=Δ/σ, where Δ is the group mean difference at the completion of the study and σ is the (pooled) within-group SD []. For a small effect size (Cohen d=0.2), the sample size required to have 80% power at a P=.02 level of significance (ie, 0.05/3 primary outcomes) was 524 in each group. After incorporating a 30% attrition rate, we anticipated requiring 748 participants in each group for a total of 1496 participants for the trial to be adequately powered.

Statistical Analysis

The study used a single-blinded approach in which only the statistician, who was external to the team, was blinded to the treatment group assignment when examining the primary hypotheses. The primary analysis was intention-to-treat (ITT) including all participants who completed the baseline assessment and were randomized to either the control or intervention group following the analysis plan prespecified in our protocol paper []. The analysis considered the following 2 features of the trial design: 3 primary end points and the use of block randomization.

For the 3 primary end points (GAD-7, PHQ-9, and USAUDIT-C scores), a global test for the null hypothesis of no treatment difference in all primary end points between the control and treatment groups was conducted first using a multivariate analysis of covariance for correlated data on GAD-7, PHQ-9, and USAUDIT-C scores at 30 days after the baseline, adjusting for their values at baseline and randomization blocks. Compared with testing each outcome separately, the advantages of the global test from joint modeling included more parsimonious hypothesis tests and mitigated concerns related to multiple testing [-] as well as pooling of the information over the correlated outcomes to increase study power, especially with missing outcomes []. If the global test rejected the null hypothesis and we concluded that there was an intervention effect on at least 1 of the 3 end points, we would then analyze each end point separately to identify which of the 3 study end points were affected by the intervention. We then used the sequential Hochberg correction method [] to control the overall familywise error rate at α=.05 when testing the hypothesis for each individual primary end point. For baseline characteristics, 2-sample t tests (2-tailed) were used to compare means, chi-square tests were used to compare proportions, and the Kruskal-Wallis test was used to test for differences in medians.

The general approach for analyzing all the types of individual study outcomes separately (including the 3 primary end points and the secondary end points) was generalized linear mixed-effects models (GLMMs) for clustered measures with the randomization block as the clustering variable. These models can handle a wide range of outcome types, including continuous, binary, ordinal, and count, and can account for the correlations among observations within the same cluster. GLMMs have been widely used for conducting ITT analysis in randomized controlled trials with missing outcome data and can account for data missing at random (MAR) without the need to model why data are missing or to perform explicit imputations of the missing values []. Specifically, we analyzed each primary end point using a linear mixed-effects model, a special case of the GLMM.

For secondary outcomes, we used linear mixed-effects models to analyze mental well-being measures and a mixed-effects quasi–Poisson regression model with a log link (a special case of GLMM) for substance use frequency measures, and the zero-inflated Poisson was used for the number of drinks to deal with the excess zero counts. The treatment effect on an outcome at 30 days was assessed in these models with the treatment allocation as the main explanatory variable and with adjustment for baseline outcome assessment value as a fixed effect and the randomization block as a random effect. Robust empirical sandwich SE estimates that are robust to model misspecifications (eg, nonnormal error terms) were used for statistical inference.

The results from these GLMMs for the analysis are reported in the Results section as the model-adjusted differences in the group mean values of the continuous end points between the intervention and control groups. These intervention effect estimates, 95% CIs, and P values were obtained from the aforementioned GLMMs, clustering on randomization block effects and adjusting for the baseline outcome values. Standardized effect sizes were calculated for continuous scores by dividing the adjusted mean differences by the SDs across all participants at baseline. The incidence rate ratios (IRRs) from Poisson regression through GLMM and zero inflation for the number of drinks are also reported.

To evaluate the robustness of the results to alternative assumptions regarding missing data, sensitivity analyses were conducted on the primary outcomes via (1) using selection models to measure the potential impacts of data missing not at random [,] and (2) adjusting analysis for additional baseline covariates potentially predictive of missing data. All the primary analyses were conducted in SAS (version 9.4; SAS Institute), except for the sensitivity analysis of the selection models, which was conducted in the isni package in R (version 3.4; R Foundation for Statistical Computing) [], whereas secondary analyses were conducted in Stata (version 15.1; StataCorp) [].

Data and Privacy

Many steps were taken to ensure the privacy of participants. Each participant received an individual account that had a unique username and password. They were then able to change this password upon logging in. Only participant emails and phone numbers (if provided through access to peer coaching) were stored within the app; names were used only for consent and were never entered into the app. Instead, participants could create a unique username within the app that they were informed should not include their full name. Identifiable data (email and phone number) were stored in data files within the app back end separate from all the app use and survey data, which were recorded with only a study ID number.


ResultsRecruitment

Recruitment initially began on September 4, 2022, during an on-campus student orientation event where interested students were asked to provide contact information to receive a follow-up email; however, participants were not provided with app download information and user accounts to begin the study until September 28, 2022, due to technical delays. Recruitment concluded on June 2, 2023, with the last participant beginning the trial on June 11, 2023.

Participant Flow

In total, 2293 individuals were invited to participate in the trial following eligibility screening and provision of informed consent. Of those 2293 individuals, 1489 (64.9%) participants completed the baseline survey and were randomized into the trial, with 743 (49.9%) of the 1489 participants in the intervention group and 746 (50.1%) in the control group. A total of 79.3% (589/743) of participants in the intervention group and 83% (619/746) of participants in the control group completed the 30-day follow-up survey within the specified 44-day period to be included in the analysis (ie, 279/1489, 18.7% of participants who completed the baseline survey did not complete the 30-day follow-up survey within 44 days and were therefore considered lost to follow-up). Additional information on the participant flow is shown in .

Figure 3. CONSORT (Consolidated Standards of Reporting Trials) flow diagram completed with participant information. Participant Characteristics

The baseline characteristics of the participants in the intervention and control groups are presented in . The median age of the participants was 20 years, and 70.3% (1045/1487) self-identified as women. In terms of mental health, 33.8% (455/1347) reported a history of anxiety, and 38.9% (576/1481) reported moderate or greater levels of recent anxiety (ie, total score of ≥10) based on the GAD-7. A history of depression was reported by 28.4% (382/1347) of participants, with 43.8% (645/1474) reporting moderate or greater levels of recent depressive symptomology (ie, total score of ≥10) on the PHQ-9. The intervention group had higher baseline scores on both the GAD-7 (P=.02) and PHQ-9 (P=.02). No other statistically significant differences in baseline characteristics were found between the intervention and control groups ().

Table 1. Participant characteristics at baseline (N=1489)a.CharacteristicsOverallIntervention (n=743)Control (n=746)P valueAge (y; n=1042), median (IQR)20 (19-23)20 (19-23)20 (19-23).54Student year (n=1487), n (%).28
Undergraduate: year 1384 (25.8)195 (26.3)189 (25.3)

Undergraduate: year 2280 (18.8)126 (17)154 (20.6)

Undergraduate: year 3264 (17.8)136 (18.4)128 (17.2)

Undergraduate: year 4 or 5287 (19.3)136 (18.4)151 (20.2)

Graduate206 (13.9)114 (15.4)92 (12.3)

Other66 (4.4)34 (4.6)32 (4.3)
Gender (n=1487), n (%).71
Woman1045 (70.3)527 (71.1)518 (69.4)

Man364 (24.5)178 (24)186 (24.9)

Nonbinary78 (5.2)36 (4.9)42 (5.6)
Race or ethnicity (n=1454), n (%).66
Black18 (1.2)7 (1)11 (1.5)

East or Southeast Asian479 (32.9)240 (33.1)239 (32.8)

Hispanic or Latino46 (3.2)24 (3.3)22 (3)

Indigenous or First Nations28 (1.9)14 (1.9)14 (1.9)

South Asian198 (13.6)102 (14)96 (13.2)

West or Central Asian58 (4)30 (4.1)28 (3.8)

White436 (30)210 (28.9)226 (31)

Multiracial187 (12.9)95 (13.1)92 (12.6)

Other4 (0.3)4 (0.6)0 (0)
Mental health diagnosis or treatment history (n=1347), n (%)
History of anxiety (yes)455 (33.8)233 (34.5)222 (33).57
History of depression (yes)382 (28.4)196 (29)186 (27.7).58
History of substance (alcohol or drug) use (yes)19 (1.4)9 (1.3)10 (1.5).81
History of other mental health disorders (yes)353 (26.2)178 (26.4)175 (26).89Current mental health and substance use
GAD-7b (n=1481), mean (SD)8.7 (5)9.0 (5)8.4 (5).02

GAD-7≥10 (moderate and above), n (%)576 (38.9)297 (40.1)279 (37.7).35
PHQ-9 c n=1474), mean (SD)9.2 (5.8)9.6 (5.9)8.9 (5.6).02

PHQ-9≥10 (moderate and above), n (%)645 (43.8)326 (44.5)319 (43).55
USAUDIT-Cd (n=1483), mean (SD)3.5 (3.2)3.6 (3.2)3.5 (3.3).31
SWEMWSe (n=1474), mean (SD)21.5 (4.2)21.3 (4.3)21.6 (4.1).19

aVariations in the total n values for each characteristic were due to missing responses. We used 2-sample t tests (2-tailed) to compare means, chi-square tests to compare proportions, and the Kruskal-Wallis test to test for differences in medians.

bGAD-7: 7-item General Anxiety Disorder scale.

cPHQ-9: 9-item Patient Health Questionnaire.

dUSAUDIT-C: US Alcohol Use Disorders Identification Test–Consumption Scale.

eSWEMWS: Short Warwick-Edinburgh Mental Wellbeing Scale.

App Use

Among those in the intervention group (743/1489, 49.9%), 77.1% (573/743) accessed at least 1 app component during the 30-day study period. More specifically, 73.8% (548/743) engaged in 1 or more chatbot activities, 21.8% (162/743) accessed the community component, 27.9% (207/743) completed the services survey, and 17.2% (128/743) accessed a peer coach.

OutcomesPrimary Outcomes

For the 3 primary end points (GAD-7, PHQ-9, and USAUDIT-C scores), a global Wald test with the null hypothesis of no treatment difference in all primary end points between the control and treatment groups was conducted using a marginal multivariate analysis of covariance model for correlated data. This multivariate test of overall group differences was statistically significant (Hotelling test=0.02; P=.001; ), indicating that there were group differences in at least 1 of the primary outcomes. We then tested each of the primary outcomes using a GLMM for clustered measures, adjusting for baseline values with the randomization block as the clustering variable, and applied a sequential Hochberg correction method [] to control the overall familywise error rate at .05 when testing the hypothesis for each individual primary end point. The results of these GLMMs indicated that participants in the intervention group had significantly greater reductions in anxiety (adjusted group mean difference=−0.85, 95% CI −1.27 to −0.42; P<.001; Cohen d=−0.17, 95% CI −0.26 to −0.09) and depressive (adjusted group mean difference=−0.63, 95% CI −1.08 to −0.17; P=.007; Cohen d=−0.11, 95% CI −0.19 to −0.03) symptoms than those in the control group (). Although participants in the intervention group also demonstrated a greater reduction in alcohol risk scores on the USAUDIT-C, this difference was not statistically significant (adjusted group mean difference=−0.13, 95% CI −0.34 to 0.08; P=.23).

We conducted the following analysis to quantify the robustness of primary findings to the assumption of data MAR. First, baseline covariates potentially predictive of missing data were added to the GLMM outcome models, and the intervention effect estimates remained similar and yielded qualitatively similar P values (, last 2 columns). Second, selection models were used that permit the missingness probability to depend on the unobserved outcome values after conditioning on the observed data, after which we computed the index of local sensitivity to nonignorability []. The index of local sensitivity to nonignorability analysis results are reported in . The ISNI/SD (ISNI divided by SD) column in estimates the change in intervention effect estimates for a moderate size of nonrandom missingness, where a 1-SD (SD of the outcome) increase in the outcome is associated with an e1=2.7–fold increase in the odds of being observed conditioned on the same values of the observed predictors for missingness. For such moderately sized nonrandom missingness, the changes in the intervention effect estimates were small for both GAD-7 and PHQ-9 scores (). The MinNI column in computes the minimum magnitude of nonignorable missingness needed for substantial sensitivity such that the selection bias due to data missing not at random is of the same size as the SE. The smaller the value of the minimum nonignorable missingness, the greater the sensitivity. A minimum nonignorable missingness of 1 is suggested as the cutoff value for important sensitivity []. The minimum nonignorable missingness values for both the GAD-7 and PHQ-9 far exceeded 1, indicating that no sensitivity to potential missingness not at random was present for the primary findings. shows that both the control and intervention groups had moderate and comparable missing data percentages for both the GAD-7 and PHQ-9, which can explain why our analysis results were insensitive to the MAR assumption.

Table 2. Results for the analysis of primary outcomes at baseline and at the 30-day follow-up.Primary outcome and periodUnadjusted, mean (SD)Adjusted differencea (primary analysis; 95% CI)P valuebAdjusted differencec (robustness checking; 95% CI)P valued
Intervention (n=743)Control (n=746)



GAD-7e
Baseline9.0 (5.0)f8.4 (5.0)gN/AhN/AN/AN/A
30 days7.7 (5.1)i8.2 (5.1)j−0.85 (−1.27 to −0.42)<.001−0.82 (−1.25 to −0.39)<.001
Effect size (Cohen d)N/AN/A−0.17 (−0.26 to −0.09)<.001−0.16 (−0.25 to −0.08)<.001PHQ-9k
Baseline9.6 (5.9)l8.9 (5.6)mN/AN/AN/AN/A
30 days8.6 (6.2)n8.8 (5.8)o−0.63 (−1.08 to −0.17).007−0.58 (−1.04 to −0.11).01
Effect size (Cohen d)N/AN/A−0.11 (−0.19 to −0.03).007−0.10 (−0.18 to −0.02).01USAUDIT-Cp
Baseline3.6 (3.2)g3.4 (3.3)qN/AN/AN/AN/A
30 days3.1 (3.0)r3.1 (3.1)s−0.13 (−0.34 to 0.08).23−0.12 (−0.32 to 0.09).28

aAdjusting for block numbers and baseline outcome values.

bGlobal test: P=.001.

cAdjusting for age, gender, student status, race, substance use at baseline, block numbers, and baseline outcome values.

dGlobal test: P=.003.

eGAD-7: 7-item General Anxiety Disorder scale.

fn=741.

gn=740.

hN/A: not applicable.

in=575.

jn=589.

kPHQ-9: 9-item Patient Health Questionnaire.

ln=732.

mn=742.

nn=563.

on=587.

pUSAUDIT-C: US Alcohol Use Disorders Identification Test–Consumption Scale.

qn=743.

rn=559.

sn=590.

Table 3. Robustness checking of group comparison of primary end points to the assumption of data missing at randoma.Primary end pointsMissing data at 30 days, nm (%)ISNIb/SDMinNIc
Intervention (n=743)Control (n=746)

GAD-7d168 (22.6)157 (21.0)−0.000457.7PHQ-9e180 (24.2)159 (21.3)0.002010.3USAUDIT-Cf184 (24.8)156 (20.9)0.01251.2

aIn the missing data percentage (nm/n), nm is the number of participants not completing the outcome assessment at 30 days, and n is the number of participants in the intention-to-treat sample.

bISNI: index of sensitivity to nonignorability; SD refers to the SD of the outcomes.

cMinNI: minimum magnitude of nonignorable missingness. A MinNI of <1 indicates sensitivity to data missing not at random, and a MinNI of >1 indicates no sensitivity to data missing not at random.

dGAD-7: 7-item General Anxiety Disorder scale.

ePHQ-9: 9-item Patient Health Questionnaire.

fUSAUDIT-C: US Alcohol Use Disorders Identification Test–Consumption Scale.

Secondary Outcomes

The GLMM for clustered measures with the randomization block as the clustering variable was also used to test for differences between the intervention and control groups on secondary outcomes without any adjustment for multiple testing. The results of these GLMMs indicated that (), compared to those in the control group, participants in the intervention group had significantly greater improvements in mental well-being (adjusted mean difference=0.73, 95% CI 0.35-1.11; P<.001; Cohen d=0.17, 95% CI 0.08-0.26) and were significantly associated with a 20% reduction in their frequency of cannabis use (IRR=0.80, 95% CI 0.66-0.96; P=.02) and a 13% reduction in the typical number of drinks consumed when drinking (IRR=0.87, 95% CI 0.77-0.98; P=.03). No significant differences were found in frequency of binge drinking (IRR=0.98, 95% CI 0.86-1.13; P=.83), frequency of drinking (IRR=0.97, 95% CI 0.88-1.06; P=.48), or frequency of any opioid use (IRR=0.62, 95% CI 0.16-2.31; P=.48). The impact of the intervention on nonmedical stimulant use could not be assessed due to the small number of nonmedical stimulant users at baseline and follow-up.

Table 4. Results for the analysis of secondary outcomes at baseline and at the 30-day follow-up.Secondary outcome and periodIntervention (n=743)Control (n=746)IRRa (95% CI)P valueWarwick well-beingb, mean (SD)
Baseline21.3 (4.3)c21.6 (4.1)cN/AdN/A
30 days22.3 (4.5)e21.8 (4.3)f0.73 (0.35-1.11)<.001
Effect size (Cohen d)N/AN/A0.17 (0.08-0.26)N/ACannabis use frequency, median category (IQR)
Baseline0 (0-2g)h0 (0-2g)cN/AN/A
30 days 0 (0-2g)i0 (0-2g)j0.80 (0.66-0.96).02Binge drinking frequency, median category (IQR)
Baseline1 (0-1g)k1 (0-1g)lN/AN/A
30 days 1 (0-1g)m1 (0-1g)n0.98 (0.86-1.13).83Drinking frequency, median category (IQR)
Baseline2 (0-3g)h1 (0-2g)oN/AN/A
30 days 1 (0-3g)p1 (0-2g)q0.97 (0.88-1.06).48Number of alcoholic drinks, median category (IQR)
Baseline1 (0-2g)h1 (0-2g)rN/AN/A
30 days 0 (0-2g)s0 (0-2g)t0.87 (0.77-0.98).03Stimulant use frequency, median category (IQR)
Baseline0 (0-0g)u0 (0-0g)vNo analysesN/A
30 days 0 (0-0g)w0 (0-0g)qN/AN/AOpioid use frequency, median category (IQR)
Baseline0 (0-0g)v0 (0-0g)oN/AN/A
30 days 0 (0-0g)x0 (0-0g)y0.62 (0.16-2.31).48

aIRR: incidence rate ratio.

bAdjusted difference.

cn=737.

dN/A: not applicable.

en=567.

fn=587.

gRefer to for the category definitions for each secondary measure.

hn=742.

in=535.

jn=577.

kn=743.

ln=740.

mn=581.

nn=616.

on=746.

pn=611.

qn=645.

rn=745.

sn=594.

tn=621.

un=739.

vn=741.

wn=610.

xn=609.

留言 (0)

沒有登入
gif