Notes on regression adjustment for selectively reported results in meta-analysis

effect-size
distribution-theory
selective-reporting
Author

James E. Pustejovsky

Published

January 17, 2026

Several of the most prominent methods for investigating selective reporting bias in meta-analyis take the form of simple linear regressions, estimated by weighted least squares. Methods of this form include the Precision Effect Test (PET) and Precision Effect Estimator with Standard Error (PEESE) adjustments proposed by Stanley () and Stanley & Doucouliagos () as well as the endogenous kink (EK) meta-regression proposed by Bom & Rachinger (). These methods are widely used (PET and PEESE especially) and have been studied pretty extensively using Monte Carlo simulations (e.g., ; ; ; ; ; ). By my read, the simulation evidence is mixed: these regression adjustments seem to be less biased than regular random effects models but they are pretty crude corrections for the bias created by selective reporting, and they don’t always work very well. In this post, I want to look at how the methods work analytically. I’ll derive expressions for their bias and accuracy and consider how one might use these expressions to say something about the conditions under which regression adjustment might be expected to perform reasonably well.

Regression adjustments for publication bias

For a meta-analysis of K studies, let Ti be an effect size estimates and σi be its standard error for i=1,...,K. Many bias adjustment methods for this context involve the regression (1)Ti=α+β xi+ϵi, where xi is a predictor that is related to the precision of the effect size estimate and ϵi is a residual error. is estimated by some form of weighted least squares with weights wi, which might be the inverses of the sampling variances, wi=σi2 (i.e., fixed effects weights) or the inverses of the total residual variance wi=(τ^2+σi2)1, where τ^2 is an estimate of between-study variance in the effect size parameters.

The intuition behind this sort of estimator is that selective reporting on the basis of statistical significance tends to produce larger biases for smaller primary studies (which have larger standard errors), leading to what looks like “small-study” effects. PET, PEESE, and EK adjust for the small-study effects using some form of linear regression, treating the estimate of the intercept α in as the average effect size in a population of infinitely large studies (so that their standard errors are σi=0). The main difference between the methods is in the predictor used to extrapolate to this hypothetical population:

  • PET uses xi=σi,
  • PEESE uses xi=σi2, and
  • EK uses xi=(σic)×I(σi>c), where c is the so-called “kink” of the meta-regression.

For now, let’s work with the generic predictors x1,...,xK. Basic results from linear regression tell us that the estimator of the intercept of a linear regression fitted by weighted least squares has the form (2)α^=i=1K[wiWx¯wi(xix¯)SSx]Ti, where W=i=1Kwi, x¯=1Wi=1Kwixi, and SSx=i=1Kwi(xix¯)2. Denoting the full vector of standard errors as σ=(σ1,...,σK) and observing that the predictors are all functions of the standard errors, the conditional expectation and conditional variance of α^ can be written as (3)E(α^|σ)=i=1K[wiWx¯wi(xix¯)SSx]×E(Ti|σi) and (4)V(α^|σ)=i=1K[wiWx¯wi(xix¯)SSx]2×V(Ti|σi). If the real regression is linear in xi, then E(Ti|σi)=α+β xi. Substitute this into and you’ll find E(α^|σ)=α, so the regression estimator is indeed unbiased. But what if the real conditional expectation is not linear?

Data-generating processes

Regression adjustment methods are a bit hard to pin down because they’re not full generative models of the process by which the observed set of effect sizes arises. Instead, they’re just different forms of approximations, which don’t really commit to any particular model for how selective reporting occurs. To say anything about how well the regression adjustments work, I will need to look beyond these specific methods to other models for the selective reporting process.

Stanley & Doucouliagos () motivated the PEESE adjustment through a Taylor series approximation to the bias of effect size estimates under a simple form of a step-function selection model, in which the real form of the bias is nonlinear in σi. There are a number of other models where the strength of selection depends on the statistical significance of the effect size estimate (e.g., the beta density model), as well as models that exhibit small-study effects produced by other selection processes (e.g., the Copas selection model). Here I’ll look at a step-function model and the Copas model, mostly because these are fairly tractable to analyze. Both models posit that effect size estimates follow a conventional random effects model prior to being put through the selection process. Denoting an effect size estimate prior to selection as Ti, we assume (5)Ti=μ+vi+ei where μ is the overall average effect size, vi is a mean-zero random effect with variance τ2, which captures heterogeneity in the effect size parameters, and ei is a mean zero sampling error with known standard error σi. In the ideal, a selective reporting correction would recover an estimate of the mean of the effect size distribution μ.

Step function models

In the step function model, we assume that the generated effect size estimate Ti is observed with some probability that depends on its p-value for the one-sided hypothesis test of H0:δ0. Letting pi=1Φ(Ti/σi), the selective reporting process of the step-function model is then (6)Pr(Ti is observed|Ti,σi)={1ifpi<α1λ1ifαhpi<αh+1,h=1,...,H1λHifαHpi for some set of steps α1,...,αH and selection weights λ=(λ1,...,λH).

In my previous post on the step-function model, I showed that the conditional expectation of the observed effect size estimates is (7)E(Ti|σi)=μ+τ2+σi2×ψ(μ,τ,λ,σi), where ψ() is a certain, rather complicated non-linear function of the model parameters given by (8)ψ(μ,τ,λ,σi)=h=0Hλh(ϕ(ch+1,i)ϕ(chi))h=0Hλh(Φ(ch+1,i)Φ(chi)), where chi=σiΦ1(1αh)μτ2+σi2, with c0i= and cH+1,i=. Letting κi=κ(μ,τ,λ,σi)=h=0Hλh[chiϕ(chi)ch+1,iϕ(ch+1,i)]h=0Hλh(Φ(ch+1,i)Φ(chi)) and ψi=ψ(μ,τ,λ,σi), the conditional variance is (9)V(Ti|σi)=(τ2+σi2)(1κiψi2).

Substituting into , the bias of a regression-adjustment estimator under the step-function selection model is (10)E(α^|σ)μ=i=1K[wiWx¯wi(xix¯)SSx]×τ2+σi2×ψ(μ,τ,λ,σi). Thus, the bias of a regression adjustment depends on how well the predictor xi approximates the non-linear function τ2+σi2×ψ(μ,τ,λ,σi).

Substituting into , an expression for the sampling variance of the regression adjustment estimator is (11)V(α^|σ)=i=1K[wiWx¯wi(xix¯)SSx]2×(τ2+σi2)(1κiψi2). Combining the bias and the variance in the usual way leads to an expression for RMSE: (12)RMSE(α^|σ)=(E(α^|σ)μ)2+V(α^|σ). In general, the magnitude of the bias and RMSE will depend on:

  • the parameters of the true distribution of effects, μ and τ,
  • the strength of selective reporting, measured by λ,
  • the choice of predictors xi and weights wi, and
  • the distribution of σi values (i.e., the average and spread of the standard errors).

That’s a lot of moving pieces. However, the bias and RMSE are at least computable, and analytic expressions are much quicker to calculate than running a couple of thousand replications of a simulation.

Copas selection models

The Copas selection model posits a different selection process than the step-function model, in which the probability that a generated effect size estimate is reported is expressed in terms of an unobserved, normally distributed variable that is correlated with the sampling error of the effect size estimate (i.e., correlated with ei in ). See my previous post on the Copas model for a more detailed description, interactive illustration, and some spicy takes on why this model is pretty odd. The selection process here has three parameters, which I will denote as a, b, and ρ (my previous post used α and β, but in this post I need these to define the regression equations). The a parameter controls the overall fraction of generated effect size estimates that are observed, the b parameter controls how strongly the probability of observation depends on the precision of the effect size estimate, and the ρ parameter controls the degree to which the probability depends on the magnitude of the observed effect size estimate relative to the overall average effect. The probability that a generated effect size estimate is observed is given by (13)Pr(Oi=1|Ti,σi)=Φ(τ2+σi2τ2+(1ρ2)σi2[α+βσi+ρσi(Tiμ)τ2+σi2]). In my previous post, I gave expressions for the mean and variance of observed effect size estimate conditional on σi. The conditional expectation is (14)E(Ti|σi)=μ+ρσiξ(α+βσi) where ξ(u)=ϕ(u)/Φ(u). The conditional variance is (15)V(Ti|σi)=τ2+σi2(1ρ2ξ(α+βσi)×[α+βσi+ξ(α+βσi)]). Substituting into , the bias of a regression-adjustment estimator under the Copas selection model is (16)E(α^|σ)μ=ρi=1K[wiWx¯wi(xix¯)SSx]×σiξ(α+βσi). This expression is a bit simpler than the bias under the step function model. For one, the bias depends neither on μ nor on τ, and is multiplicative in the ρ parameter. An implication is that, if you were running a simulation where the Copas model was the true data-generating process, the bias of the regression adjustment estimator should be invariant across μ and τ and should be strictly proportional to ρ. Just as with the step-function selection model, the bias will depend on how well the predictor xi approximates the functional form of the bias curve σiξ(α+βσi), which will in turn depend on the distribution of the standard errors. In other words, the regression adjustments will perform better or worse, depending on the average σi and shape of the σi distribution across the included studies.

Why care?

I’ve just thrown down a lot of rather ugly looking formulas, and one might well ask why anyone should care about this mess of math. What are these formulas good for? To be honest, I worked through this stuff mostly for myself, but I do think there might be some useful implications to be drawn from this. For one, the expressions for bias and RMSE show how the performance of the regression adjustments depend on the model parameters, which is useful to know if you’re trying to determine how to set up a simulation study involving PET or PEESE, choose which parameters to vary, or fit a meta-model to the results.

Another, perhaps more pragmatic application would be to use the expressions to compare the bias or accuracy of different forms of regression adjustment. For a given application, should we use PET or PEESE? And should we weight by inverse sampling variances (fixed effects weights) or inverse total variances (random effects weights) or something else (like equal weighting)? We might be able to get some purchase on these questions by computing bias or RMSE for some hypothetical scenarios. Under either the step function model or the Copas model, the performance of a regression adjustment will depend on the distribution of sampling standard errors, σ1,...,σK. When conducting a meta-analysis, we will know this sample distribution but we won’t know the true values of μ, τ, or the parameters of the selection process. Still, we could evaluate the expressions above under some plausible hypothetical scenarios and see whether one regression adjustment is consistently better than the others. This will take some further fleshing out, which I will leave for another day.

Back to top

References

Bom, P. R. D., & Rachinger, H. (2019). A kinked meta-regression model for publication bias correction. Research Synthesis Methods, 10(4), 497–514. https://doi.org/10.1002/jrsm.1352
Bramley, P., López-López, J. A., & Higgins, J. P. T. (2021). Examining how meta-analytic methods perform in the presence of bias: A simulation study. Research Synthesis Methods, 12(6), 816–830. https://doi.org/10.1002/jrsm.1516
Carter, E. C., Schönbrodt, F. D., Gervais, W. M., & Hilgard, J. (2019). Correcting for bias in psychology: A comparison of meta-analytic methods. Advances in Methods and Practices in Psychological Science, 2(2). https://doi.org/10.1177/2515245919847196
Hong, S., & Reed, W. R. (2021). Using Monte Carlo experiments to select meta-analytic estimators. Research Synthesis Methods, 12(2), 192–215. https://doi.org/10.1002/jrsm.1467
McShane, B. B., Böckenholt, U., & Hansen, K. T. (2016). Adjusting for publication bias in meta-analysis an evaluation of selection methods and some cautionary notes. Perspectives on Psychological Science, 11(5), 730–749. http://pps.sagepub.com/content/11/5/730.short
Moreno, S. G., Sutton, A. J., Ades, A., Stanley, T. D., Abrams, K. R., Peters, J. L., & Cooper, N. J. (2009). Assessment of regression-based methods to adjust for publication bias through a comprehensive simulation study. BMC Medical Research Methodology, 9(1). https://doi.org/10.1186/1471-2288-9-2
Stanley, T. D. (2008). Meta-regression methods for detecting and estimating empirical effects in the presence of publication selection*. Oxford Bulletin of Economics and Statistics, 70(1), 103–127. https://doi.org/10.1111/j.1468-0084.2007.00487.x
Stanley, T. D. (2017). Limitations of PET-PEESE and other meta-analysis methods. Social Psychological and Personality Science, 8(5), 581–591. https://doi.org/10.1177/1948550617693062
Stanley, T. D., & Doucouliagos, H. (2014). Meta-regression approximations to reduce publication selection bias. Research Synthesis Methods, 5(1), 60–78.

Footnotes

  1. I will treat the standard errors as known quantities that are independent of the effect size estimates, even though the standard errors are actually stochastic and, for some effect size metrics, correlated with the effect size estimates.↩︎

  2. All of the adjustment methods will run into trouble if small study effects are the result of some mechanism other than selective reporting. For instance, if some primary study feature is systematically, positively associated with effect size and tends to be less common in larger studies, then it will act as a confounder, inducing correlation between effect size and standard error even in the absence of selective reporting.↩︎

  3. With the EK estimator, the kink is an estimated quantity, defined by a preliminary estimator of α and τ2. This makes is quite tricky to analyze, but I’m going to gloss over that detail entirely (at least for the remainder of this post).↩︎

  4. For a model with a single step at α1 and a single selection parameter λ1, ψ(μ,τ,λ1,σi)=(1λ1)ϕ(c1i)1(1λ1)Φ(c1i).↩︎

  5. Sometimes, I get a theoretical itch that I just can’t help but scratch.↩︎