Standardized mean differences in single-group, repeated measures designs

effect size
standardized mean difference
distribution theory
Author

James E. Pustejovsky

Published

October 6, 2021

I received a question from a colleague about computing variances and covariances for standardized mean difference effect sizes from a design involving a single group, measured repeatedly over time. Deriving these quantities is a little exercise in normal distribution theory, which I find kind of relaxing sometimes (hey, we all have our coping mechanisms!).

The set-up

Consider a study in which a single group of n participants was measured at each of T+1 time-points, indexed as t=0,...,T. At the first time-point, there has not yet been any exposure to an intervention. At the second and subsequent time-points, there is some degree of exposure, and so we are interested in describing change between time point t>0 and time-point 0. For each time-point, we have a sample mean y¯t and a sample standard deviation st. For now, assume that there is complete response. Let μt denote the population mean and σt denote the population standard deviation, both at time t. Let ρst denote the correlation between outcomes measured at time s and time t, for s,t=0,..,T, where ρtt=1. We might also have sample correlations for each time point, denoted rst. We calculate a standardized mean difference for each time-point t>0 by taking dt=y¯ty¯0sP, where sp is the sample standard deviation pooled across all time-points: sP2=1T+1t=0Tst2. The question is then, what is Var(dt) and what is Cov(ds,dt), for s,t=1,...,T?

The results

Define the unstandardized mean difference between time-point t and time-point 0 as Dt=y¯ty¯0. Then, from the algebra of variances and covariances, we have Var(Dt)=1n(σ02+σt22ρt0σ0σt) and Cov(Ds,Dt)=1n[σ02+ρstσsσtσ0(ρs0σs+ρt0σt)]. From a previous post about the distribution of sample variances, we have that Cov(ss2,st2)=2(ρstσsσt)2n1. Consequently, Var(sP2)=1(T+1)2s=0Tt=0TCov(ss2,st2)=2(n1)(T+1)2s=0Tt=0T(ρstσsσt)2. Let σP2=1T+1t=0Tσt2 denote the average population variance across all T+1 time-points, and let δt denote the standardized mean difference parameter at time t. Then, following the multivariate delta method, Var(dt)Var(Dt)σP2+δt22νandCov(ds,dt)Cov(Ds,Dt)σP2+δsδt2ν, where ν=2σP4Var(sP2).

Without imposing further assumptions, and assuming that we have access to the sample correlations between time-points, a feasible estimator of the sampling variance of dt is Vt=s02+st22rt0s0stnsP2+dt22ν^, where ν^=(n1)sp41(T+1)2s=0Tt=0Trst2ss2st2. Similarly, a feasible estimator for the covariance between ds and dt is Cst=s02+rstsssts0(rs0ss+rt0st)nsP2+dsdt2ν^.

In some cases, it might be reasonable to use further assumptions about distributional structure in order to simplify these approximations. In particular, suppose we assume that the population variances are constant across time-points, σ0=σ1==σT. In this case, the variances and covariances no longer depend on the scale of the outcome, and we have ν^=(n1)(T+1)TR+1,whereR=2T(T+1)s=0T1t=s+1Trst2 (here, R is the average of the squared correlations between pairs of distinct time-points). Since R will always be less than 1, ν^ will always be larger than n1. If sample correlations aren’t reported or available, it would seem fairly reasonable to use ν^=n1, or to make a rough assumption about the average squared correlation R. With the approximate degrees of freedom ν^, the variances and covariances are then given by Vt=2(1rt0)n+dt22ν^andCst=1+rstrs0rt0n+dsdt2ν^.

Extension

In some contexts, one might encounter a design that uses over-lapping but not identical samples at each time-point. For instance, in a rotating panel survey, each participant is measured repeatedly for some small number of time-points p<T+1 (say p=2 or p=3), and new participants are added to the sample with each new time-point. The simple repeated measures set-up that I described in this post is an imperfect approximation for such designs. In dealing with such a design, suppose that one knew the total number of observations at each time-point, denoted nt for t=0,...,T, as well as the number of observations that were common across any pair of time-points, denoted as nst for s,t=0,...,T. Further suppose that the drop-outs and additions are ignorable (missing completely at random), so that any subset of participants defined by a pattern of response or non-response is still representative of the full population. I leave it as an exercise for the reader (a relaxing and fun one!) to derive Var(dt) and Cov(ds,dt) under such a model.

Back to top