A while back, I posted the outline of a problem about the number of significant effect size estimates in a study that reports multiple outcomes. This problem interests me because it connects to the issue of selective reporting of study results, which creates problems for meta-analysis. Here, I’ll re-state the problem in slightly more general terms and then make some notes about what’s going on.
Consider a study that assesses some effect size across different outcomes. (We’ll be thinking about one study at a time here, so no need to index the study as we would in a meta-analysis problem.) Let denote the effect size estimate for outcome , let denote the sampling variance of the effect size estimate for outcome , and let denote the true effect size parameter for corresponding to outcome . Assume that the study outcomes follow a correlated-and-hierarchical effects model, in which where the study-level error , the effect-specific error , and the vector of sampling errors is multivariate normal with mean , known variances , and compound symmetric correlation structure .
Define as an indicator that is equal to one if is statistically significant at level based on a one-sided test, and otherwise equal to zero. (Equivalently, let be equal to one if the effect is statistically significant at level and in the theoretically expected direction.) Formally, where is the critical value from a standard normal distribution (e.g., , ). Let denote the total number of statistically significant effect sizes in the study. The question is: what is the distribution of .
Compound symmetry to the rescue
As I noted in the previous post, this set-up means that the effect size estimates have a compound symmetric distribution. We can make this a bit more explicit by writing the sampling errors in terms of the sum of a component that’s common acrosss outcomes and a component that’s specific to each outcome. Thus, let , where and . Let me also define as the conditional mean of the effects. It then follows that the effect size estimates are conditionally independent, given the common components: Furthermore, the conditional probability of a significant effect is and are mutually independent, conditional on . Therefore, the conditional distribution of is binomial, where What about the unconditional distribution?
To get rid of the , we need to integrate over its distribution, which leads to where is a binomial density with size and probability and is a normal density with mean and variance .
This distribution is what you might call a binomial-normal convolution or a random-intercept probit model (where the random intercept is ). As far as I know, the distribution cannot be evaluated analytically but instead must be calculated using some sort of numerical integration routine.
Just the moments, please
If all we care about is the expectation of , we don’t need to bother with all the conditioning business and can just look at the marginal distribution of the effect size estimates taken individually. Marginally, is normally distributed with mean and variance , so , where By the linearity of expectations,
We can also get an approximation for the variance of by working with its conditional distribution above. By the rule of variance decomposition, where is, as defined above, a function of and thus a random variable. Now, and we can get something close to using a first-order approximation: Thus, If the amount of common variation is small, so is near zero and is near zero, then the contribution of the second term will be small, and will act more or less like a binomial random variable with size and probability . On the other hand, if the amount of independent variation in the effect sizes is small, so is near zero and is near 1, then the term on the right will approach and will approach , or the variance of times a single Bernoulli variate. So you could say that has anywhere between and variate’s worth of information in it, depending on the degree of correlation between the effect size estimates.
Interactive distribution
Here is an interactive graph of the probability mass function of , with probability points calculated using Gaussian quadrature. Below the graph, I also report , the exact mean and variance of , and the first-order approximation to the variance (denoted $V_{approx}). When and , the approximate variance is not all that accurate because the first-order approximation to isn’t that good.