Measuring Change in Coping Skills: A Pre/Post Walkthrough for MH&A Training
A simple paired analysis
program evaluation
mental health
addictions
R
pre-post
outcomes
Author
A. Srikanth
Published
November 21, 2025
Data Byte
Context
Most evaluation conversations in mental health and addictions (MH&A) start from the same place: “We ran a new group, we distributed our pre- and post-training surveys. Did it help?” I recall reading a Quick Tips guide from the University of Toronto’s Continuing Professional Development (CPD) team a short while ago, which nudged me to slow down and decide what part of the story we are actually trying to measure, who needs the answer, and how they will use it before I even think about getting started with data analysis.
An important note to be made here is that a pre/post scale is not a verdict on the entire program. It is one lens, mainly about learning and early competence, that can sit alongside participation, satisfaction, performance, and patient outcomes in a broader program evaluation plan.
Suppose we have a six-session “Coping Skills for Recovery 101” group at your trusted MH&A hospital. The group serves outpatients from two streams: one focused on mood and anxiety, and the other on concurrent substance use. At session one and session six, participants complete a brief 10-item coping confidence scale scored from 0 to 10.
Objectives
We want to know whether participants who complete the group report higher coping confidence, we want the result to inform whether to keep or adapt the group, the main audience is the clinical lead and manager, and the output will be a short summary that feeds into next year’s planning cycle.
Data Sources
Here, let’s build a small fictitious dataset in R for 270 clients from a hospital coping skills group, one row per person. Each row has a client ID (client_id), a stream label for either mood and anxiety or concurrent use (stream), and two coping knowledge scores: one before the group (pre_knowledge) and one after (post_knowledge). Pre scores are simulated from a normal distribution centred around the middle of a 0 to 10 scale, then I add an average improvement of about 1.2 points plus noise, trimming everything to stay within 0 and 10. That produces a realistic mix of partial gains, a few declines, and some ceiling effects. Once that synthetic run is working, one can swap the dataset for an export from REDCap or your clinical system with the same basic structure and reuse the rest of the analysis.
In terms of the analysis in R, that means taking post_score - pre_score within each row, summarising the mean and confidence interval of that change, and running a paired t-test with t.test(post_score, pre_score, paired = TRUE).
$summary
# A tibble: 1 × 4
n mean_pre mean_post mean_change
<int> <dbl> <dbl> <dbl>
1 270 5.55 6.77 1.23
$t_test
Paired t-test
data: dat$post_knowledge and dat$pre_knowledge
t = 16.807, df = 269, p-value < 2.2e-16
alternative hypothesis: true mean difference is not equal to 0
95 percent confidence interval:
1.082761 1.370097
sample estimates:
mean difference
1.226429
This answers a very specific question: among participants who finished the group and completed both surveys, what is the average gain in coping confidence and how precisely is it estimated? It aligns with the “learning” and early “competence” levels from above, which focus on whether people know what to do and can demonstrate that knowledge in the structured setting of the program.
The Preprocessing Step(s) Once Removed?
On a real dataset, the workflow would involve another step: reshaping the data so that each participant has one row with clearly named pre and post columns and filtering to rows where both are non-missing (super important!).
We can also check the distribution of change scores; a quick histogram will tell you whether most people are clustered around zero or shifting in a meaningful way.
As a way of ensuring our analysis steps are sound but also delivering a more rich analysis to leadership, we can repeat the summary by stream or other meaningful subgroups; by doing so, the clinical team would be able to see whether certain populations are benefiting more.
Results & Next Steps
The value of this approach is not the p-value; it is the clarity about what we are and are not claiming. A pre/post scale will not tell you whether clinicians changed their practice, whether relapse rates dropped, or whether community outcomes improved. Those sit higher in the evaluation hierarchy and rely on other methods such as chart reviews, administrative data, or follow-up surveys—the CPD calls these complementary options. What a paired analysis does give you is a compact, defensible view of learning for the subset of clients who completed the group, expressed in units that are easy to explain in supervision, at a program meeting, or in a one-page briefing to leadership.
If you are an evaluator or analyst in a similar setting, the takeaway from all of this is simple. Treat your pre/post scales as one building block in a larger evaluation design, set up a clean paired analysis with a small, well-documented R script, and connect the numerical change back to the specific decisions your team needs to make about keeping, scaling, or reshaping the program. Once that basic pattern is in place, it is easy to “rinse and repeat” the same code across groups, gradually layer in other evaluation levels, and turn scattered surveys into a reusable feedback loop for your services.