Power analysis for random‐effects meta‐analysis
Abstract
One of the reasons for the popularity of meta‐analysis is the notion that these analyses will possess more power to detect effects than individual studies. This is inevitably the case under a fixed‐effect model. However, the inclusion of the between‐study variance in the random‐effects model, and the need to estimate this parameter, can have unfortunate implications for this power. We develop methods for assessing the power of random‐effects meta‐analyses, and the average power of the individual studies that contribute to meta‐analyses, so that these powers can be compared. In addition to deriving new analytical results and methods, we apply our methods to 1991 meta‐analyses taken from the Cochrane Database of Systematic Reviews to retrospectively calculate their powers. We find that, in practice, 5 or more studies are needed to reasonably consistently achieve powers from random‐effects meta‐analyses that are greater than the studies that contribute to them. Not only is statistical inference under the random‐effects model challenging when there are very few studies but also less worthwhile in such cases. The assumption that meta‐analysis will result in an increase in power is challenged by our findings.
1 INTRODUCTION
Meta‐analysis is now a very commonly used statistical tool. There are many motivations for including meta‐analyses in systematic reviews. The Cochrane Handbook1 (their section 9.1.3) gives 4 reasons for doing so: (1) to increase power, (2) to improve precision, (3) to answer questions not posed by the individual studies, and (4) to settle controversies or generate new hypotheses. The focus here is on the first of these reasons, namely, to increase power. To quote directly from the Cochrane Handbook, “Many individual studies are too small to detect small effects, but when several are combined there is a higher chance of detecting an effect.” This statement furthers our belief that there is a commonly held implicit assumption that meta‐analysis necessarily provides a way to increase statistical power and so detect effects of interest. In this paper, we will challenge this assumption. This issue should be of interest to applied analysts regardless of their preferences concerning whether to present confidence intervals or results from hypothesis tests. The former is often encouraged, and we would align ourselves with those who emphasise estimation over testing. Due to the connection between hypothesis testing and confidence intervals however, a significant hypothesis test result is equivalent to a confidence interval that excludes the null. All our results that follow are therefore immediately applicable to those who prefer to present confidence intervals because they can frame the findings of whether or not confidence intervals contain the null, and so whether they are able to detect an effect.
The power of a hypothesis test is the probability that the null hypothesis is rejected when it is false. Bayesian methods would be needed to instead calculate the probability that the null hypothesis is false, but here we focus on classical methods. Power analysis for meta‐analysis is a sufficiently important topic to warrant an entire chapter devoted to it in the introductory text by Borenstein et al2 (their chapter 29). The methods for power analysis that we develop below are of this conventional type and do not attempt to allow for multiple testing within the same review3 or sequential testing in updated reviews.4 We will compare the power of the individual studies and the power of the meta‐analysis to which they contribute, because both types of power are important considerations, which impact on each other in practice.5, 6 The planning of future studies may be based on the power or precision of meta‐analysis results.5, 7, 8
Our other main focus is on the use of random‐effects meta‐analyses. The random‐effects model9-12 relaxes the strong, and usually difficult to defend, assumption made by the fixed‐effect model that the studies estimate the same true effect. The random‐effects model introduces two reasons for doubting that the resulting meta‐analyses will possess more power than the individual studies, both of which are directly related to the between‐study variance. The first reason is that, compared to the fixed‐effect model, the need to estimate an additional parameter in random‐effects meta‐analyses could result in power loss even when there is no between‐study heterogeneity; in general, power is lost when introducing further parameters to a statistical model. However, a more obvious concern is that if the between‐study heterogeneity is very large then the variation in the study estimates could be very considerable; by (albeit, correctly) allowing for this variation, the standard error of the pooled estimate could then be very large. If this occurs, then the corresponding hypothesis test will possess very little power.
Others have previously discussed power calculations under the random‐effects model.13, 14 Hedges and Pigott (2001)15 also discuss how the unknown between‐study variance complicates power analysis in the random‐effects setting. Although these methods are directly related to our new methods, and will also be used below, a key distinction between our new methods and most previous work is that we develop methods for evaluating the power whilst taking into account the uncertainty in the estimated between‐study variance.
We should however be clear from the outset that meta‐analysis and study specific hypothesis tests involve testing different types of hypotheses: in a meta‐analysis, we test whether or not the average effect is a particular value (for example zero), and for individual studies we test whether or not the true study specific treatment effect is a particular value. The distinction between these 2 types of hypotheses is especially clear when using the random‐effects model, where we assume that the true treatment effects differ across studies. Whilst recognising that the meta‐analysis and study specific hypothesis tests differ in this way, we will still compare the power of these two types of tests. We should also recognise that the study‐specific and fixed‐effect model hypothesis tests possess, under the assumptions that these methods make, the correct significance level. However, the conventional random‐effects model hypothesis test only retains the nominal significance level approximately. To more fairly compare the power of different hypothesis tests, we should use the same actual (and not just the nominal) significance level throughout, and this is only approximately the case here. Some of the power of the conventional random‐effects model hypothesis test is therefore an artifact of the approximate nature of the methods used.
One motivation for developing methods for power analysis under the random‐effects model is so that we can retrospectively determine the powers of the random‐effects meta‐analyses from a large sample of meta‐analyses from Cochrane. We therefore calculate what is sometimes referred to as the “observed power” in this empirical investigation. Then, by comparing the power of these meta‐analyses to the average power of the studies that contribute to them, we can empirically investigate the validity of the notion that random‐effects meta‐analyses result in an increase in power. This empirical investigation of a large number of random‐effects meta‐analyses is one important contribution of this paper. The new Monte Carlo method that we develop for this purpose is another important contribution. Retrospective power calculations have some value in practice because they have the potential to explain why effects were not detected. For example, a systematic reviewer might be disappointed or surprised not to detect a particular effect, but this is likely to be mitigated or explained by a calculation that reveals that the power was in any case very low. Although low power also manifests itself as wide confidence intervals, a power calculation provides a much more direct statement about the difficulty in detecting effects than a confidence interval. However this should not be taken to suggest that the usual statistical inferences, made for example by using confidence intervals, are in any sense deficient because they do not involve power calculations. Readers may also be able to suggest other reasons why retrospective power calculations could be of interest.
Despite this, we would not advocate the routine use of retrospective power calculations in meta‐analysis, rather they are likely to be useful in some instances to clearly convey the difficulty in detecting particular effects. In our empirical investigation, we retrospectively investigate the powers of random‐effects meta‐analyses so that these powers can be compared to study specific powers. Our retrospective power calculations are performed to answer the more general question of whether or not random‐effects meta‐analyses usually provide an increase in power, rather than to retrospectively investigate the power for any specific meta‐analysis. Those who might advocate the routine use of retrospective power calculations should examine the arguments made by Hoenig and Heisey16 and, in our opinion, consider amending their position.
Another motivation is to develop a method for power analysis that is most suitable for those who wish to perform such an analysis at the planning stage. As Borenstein et al2 point out, ideally “a power analysis should be performed when the review is being planned, and not after the review has been completed.” However, determining suitable information for a power calculation at the planning stage of a meta‐analysis or systematic review is challenging, because the power depends on quantities like the number and size of studies, which the systematic reviewer cannot directly control (although as Borenstein et al point out, we could modify the inclusion criterion to include more or fewer studies). We therefore develop a method that facilitates a meaningful power calculation for those planning reviews and that also allows for the uncertainty in the estimation of the between‐study variance. This is another important contribution of this paper. Power calculations are especially useful at this early stage, because they can be used to inform systematic reviewers about important practical issues such as whether they should modify the inclusion criterion to accommodate more studies, or wait until more studies are available, and so on. We strongly encourage power calculations at the planning stage for this reason, and suggest that these could be included in protocols.
The rest of the paper is set out as follows. In section 2, we explain how to derive the power of the studies that contribute to a meta‐analysis, so that the study specific powers can subsequently be computed and compared to those from a meta‐analysis that combines these studies. In section 3, we examine the implications for the fixed‐effect model, where we quickly find that a fixed‐effect meta‐analysis necessarily results in an increase in power. In section 4, we examine the random‐effects model and propose a new way to derive the average study specific power and 3 ways to derive the power of random‐effects meta‐analyses. In section 5, we assess the performance of the methods proposed in section 4. In section 6, we explore the study specific and meta‐analysis powers empirically in a large database from Cochrane. Together the findings in sections 5 and 6 enable us to reach some important conclusions that we discuss in section 7.
2 THE POWER OF THE INDIVIDUAL STUDIES THAT CONTRIBUTES TO THE META‐ANALYSIS
In this section, we derive the power of the individual studies that contribute to the meta‐analysis. At this stage, we make no use of meta‐analysis methodology, because we make no assumptions about how the true treatment effects for each study relate to each other.
We let μi denote the true effect in study i and let k denotes the number of studies. We let Yi denotes this study's estimate of μi and let σi denotes the corresponding standard error. These standard errors are usually estimated in practice but treated as if fixed and known in analysis. We suppress the fact that the within‐study standard errors are estimated prior to performing the meta‐analysis, and so write σi instead of
. We also use normal within‐study approximations
, as is conventional in meta‐analysis and common when analysing data from individual trials. We assume that 2‐tailed hypothesis tests are used throughout. We make no attempt to distinguish “accepting the null hypothesis” and “not rejecting the null hypothesis” and other more subtle issues related to the interpretation of hypothesis and significance testing.
, and this hypothesis is accepted if |Zi|<Za, where Za is a suitable critical value from a standard normal distribution; Za=1.96 gives the conventional 5% significance level that we will assume is used in our investigations below. The probability of accepting the null hypothesis is therefore equal to Φ(Za−δi/σi)−Φ(−Za−δi/σi), where Φ(·) is the standard normal cumulative distribution function. Hence, the power is given by the probability of correctly rejecting the null hypothesis when it is false, which is
(1)2.1 The probability of rejecting the null hypothesis and inferring the correct directional effect

3 THE FIXED‐EFFECT MODEL
The power for each individual study is easily calculated as explained in section 2. We now examine the power of meta‐analyses that combine such studies, so that the study specific and the meta‐analysis powers can be compared. The fixed‐effect (or common‐effect) model assumes that there is no between‐study variation, so that μi=μ, and δi=δ=μ−μ0 for all i. This means that we assume that
, and calculating powers of hypothesis tests are straightforward.
3.1 The power of the individual studies
(2)
(3)3.2 The power of fixed‐effect meta‐analyses
, where
. Hence, the power of the 2‐sided hypothesis test H0:μ=μ0 is given by expression 2 with σi replaced by
, which is
(4)Assuming that there is more than a single study in the meta‐analysis, it is straightforward to show that
for all i. Using this fact, it is then straightforward to show that the power of the fixed effect meta‐analysis is greater than all the study specific powers in Equation 2. A similar analysis shows that the fixed‐effect meta‐analysis also possesses more power than the individual studies in the context of 1‐sided hypothesis tests, provided that μ lies in the direction of the alternative hypothesis. Of course if there is a single study then the study specific and meta‐analysis powers are the same.
This analysis shows that in the case of a fixed‐effects model, the claim that “meta‐analyses increase power” is completely justifiable. This conclusion was also reached by Cohn and Becker,13but we also give details here to motivate our analysis of the random‐effects model, which provides our main interest.
4 THE RANDOM‐EFFECTS MODEL
The analyses of individual studies (section 2) and the fixed‐effect model (section 3) are straightforward. However, matters are more complicated under the random‐effects model. The random‐effects model relaxes the assumption that μi=μ for all i and instead assumes μi∼N(μ,τ2), so that δi∼N(δ,τ2). If τ2=0, then we have δi=δ, and we recover the fixed‐effect model as a special case. The random‐effects model is often presented as a slight modification of the fixed‐effect model. This is because τ2 is typically estimated and then treated as fixed and known in analysis, so that
is replaced by
in analysis. However, this does not take into account the fact that τ2 is estimated, and the uncertainty in τ2 is considerable in typical meta‐analyses with few studies.18-20 Another complication when comparing powers is that the powers of the individual studies that contribute to a random‐effects meta‐analysis now depend on random δi.
4.1 The average power of the individual studies
Now that δi is a random variable, we can obtain the average study specific power
by taking the expectation of β(δi,σi) (from Equation 2) over the joint distribution of (δi,σi). For the fixed‐effect model, only the σi differed across studies, and we took this expectation over their empirical distribution. Under the random‐effects model, we have δi∼N(δ,τ2), and we continue to approximate the distribution of σi with their empirical distribution, where we further assume that δi and σi are independent. In situations where an association between study specific estimates and their precision is observed then this is generally attributed to small study effects or publication bias; we assume that no such phenomena are present.
4.2 Fitting the random‐effects model

,
. Under the assumptions of the random‐effects model we have

, which provides the DerSimonian and Laird estimate

, where
, and the distribution of
is approximately
. The resulting test statistic for testing H0:μ=μ0 is given by T, the ratio of
and its approximate standard error, which can be written as
(6)4.3 The power of random‐effects meta‐analyses
The distribution of T in Equation 6 is, at best, very difficult to obtain analytically so that suitable power formulae are harder to derive than for the fixed‐effect model. We therefore suggest 3 approaches for evaluating the power of the test based on Equation 6, that have different advantages and disadvantages. The first 2 methods require values of δ and τ2, and the within‐study variances, and are very closely related. The first method is a well established approximate analytical approach, and the second is a more computationally expensive numerical analogue of the first method that allows for the uncertainty in estimates of τ2. The third method assumes that all studies are the same size and requires just the number of studies, the proportion of variation that is due to between‐study heterogeneity (which we will denote as I2, where this is the quantity that the I2 statistic proposed by Higgins and Thompson21 estimates) and a noncentrality parameter that depends on δ, the number of studies and the study size.
4.3.1 An analytical approach
is replaced by
, where τ2 is usually taken to be its estimated value. Hence, the power can be taken to be as described in Equation 4 for the fixed‐effect model but where VF is replaced with
, which gives
(7)4.3.2 A Monte Carlo approach
, so that Equation 6 becomes
(8)
in Equation 8. Hence, we can obtain the power of the random‐effects meta‐analysis, for true values of δ and τ2 and a set of within‐study variances, by simulating many meta‐analyses as
, i=1,⋯k, and then using standard meta‐analysis software to perform random‐effects meta‐analyses using these outcome data. The metafor package and the command rma.uni will be used with the “DL”10 option for this purpose throughout. Then the proportion of simulated random‐effects meta‐analyses that are statistically significant at the appropriate level gives the power denoted as βR(δ,τ2,σ), but obtained differently, in Equation 7.
The difference between this approach and the previous one is that it clearly distinguishes between the true and estimated τ2. Although this Monte Carlo method requires the analyst to determine the true value of τ2 to use in the power calculation, the estimated between‐study variance is used when computing the simulated test statistics that are used to determine the power of the test. In meta‐analyses with large numbers of studies, the approximation
is appropriate when making inferences about the average effect, and we will see below that allowing for the uncertainty in the estimated between‐study variance becomes unimportant in meta‐analyses with very large numbers of studies. The advantage of this approach is that it allows for the uncertainty in the estimated between‐study variance when making inferences about the average effect, and so can be expected to provide more accurate powers. The disadvantage of this method is that it requires simulation and so is computationally more expensive and subject to Monte Carlo error.
4.3.3 An analytical approach assuming that all studies are the same “size”
for all i; this means that all studies provide the same amount of information). We show in the web supplementary materials that the cumulative distribution function of T is given by

(9)
is a noncentrality parameter,
is the probability density function of the χ2 distribution with (k−1) degrees of freedom, and we define the incomplete Gamma function:

, where
is the typical within‐study variance used by Higgins and Thompson,21
(10)
, and the power of the random‐effects meta‐analysis is then obtained by subtracting this probability from 1. This method gives the same power as the second method when all studies are the same size (although the second method is subject to Monte Carlo error), and so also computes the power denoted as βR(δ,τ2,σ), but obtained differently, in Equation 7.
The main advantage of this method is that it takes into account the uncertainty in τ2 without resorting to simulation. As we will explain below, the quantities required for this method are more amenable to researchers performing power calculations at the planning stage. The main disadvantage of this approach is that it requires considering an artificial special case, but this type of special case was also used by Hedges and Pigott (2001)15 to make the first method more accessible to applied analysts.
5 ASSESSING THE PERFORMANCE OF THE 3 METHODS FOR CALCULATING META‐ANALYSIS POWER UNDER THE RANDOM‐EFFECTS MODEL
We now have 3 main proposals for computing the power of random‐effects meta‐analyses. The main differences between the methods are whether or not uncertainty in τ2 is taken into account, and whether or not all studies are assumed to be the same size. Hence, the relative importance of these factors are crucial in determining which method should be recommended in practice.
5.1 The implications of allowing for the uncertainty in the estimated between‐study variance
The more conventional method for power analysis in random‐effects meta‐analysis, described in section 4.3.1, does not allow for the uncertainty in the estimation of τ2 when making inferences about the average effect. To investigate how important this consideration is for meta‐analysis power calculations, we will initially consider the case where all studies are the same size so that we can perform the investigation analytically.
5.2 An important comparison
is equal to VF. However, under the random‐effects model, writing
, the “average” study‐specific power (Equation 5) can be written as
(12)
(13)
, whereas the meta‐analysis hypothesis test uses the more usual standard normal critical values ±Za. We have
so that
, which means that the power of the study‐specific hypothesis test is equivalent to the meta‐analysis hypothesis test but where the study‐specific hypothesis test has used smaller critical values. This means that the study‐specific power in Equation 12 is greater than the random‐effects meta‐analysis power in Equation 13 if between‐study heterogeneity is present. This analysis clarifies the different natures of the 2 types of hypothesis tests and also indicates that sufficient numbers of studies will be needed for the random‐effects meta‐analysis to achieve greater power than the individual studies that contribute to it.
5.3 Comparing the powers from calculations that do, and do not, take into account the uncertainty in τ2
The powers calculated from Equation 11 depend only on the noncentrality parameter
and I2=τ2/(σ2+τ2). However the powers calculated from Equation 9, and as explained in section 4.3.3., further depend on k, where this dependence can also be conveniently expressed in the degrees of freedom (k−1). It is obvious that Equation 11 does not depend on the sign of Δ, and only on its magnitude |Δ|, because Φ(z)=1−Φ(−z). It is perhaps less obvious that this is also the case in powers calculated in the way explained in section 4.3.3, but the same identity also establishes this. Hence, we only need consider, for example, positive Δ, because the results for −Δ are the same. This is in any case intuitively obvious because we perform conventional 2‐tailed tests, so that the power depends only on the magnitude of δ and so on the magnitude of Δ.
In Figure 1, we show the contour plots of the resulting powers in Δ and I2, obtained using Equation 9. In each plot, we show power contours at 0.1, 0.2, ⋯, 0.9, and we also show the corresponding contours using Equation 11 as dotted lines (without labelling them, to avoid cluttered figures). As the sample size k increases, the results using Equation 9 to calculate the power become more similar to those from Equation 11, and for inordinately large k(1000 say, results not shown), the powers obtained using Equation 9 and from Equation 11 become indistinguishable. This is because as the sample size increases, the uncertainty in τ2 becomes negligible. For small k and large I2, much smaller Δ are apparently required to achieve low power from Equation 9 than Equation 11, but this is an artifact of the standard methods for random‐effects meta‐analysis being anticonservative in small samples where the heterogeneity is severe; the random‐effects model's hypothesis test possesses considerably less than its nominal significance level in such situations, which results in artificially small Δ to achieve powers that are only slightly greater than the nominal significance level. Note also that Δ is an increasing function in k. Hence, for a fixed effect δ and within‐study standard error σ2, as k increases so does Δ. Hence, Δ corresponds to a smaller effect δ as k increases, and from Figure 1, we can see that the power to detect any fixed value of δ increases as the sample size k becomes larger.

when performing power calculations. This figure explores the special case where all studies are the same size. The 4 plots show the power of the standard random‐effects model's hypothesis test for k=3, 5, 10, and 50 studies, as a function of Δ and I2. These plots allow for the fact that the between‐study variance is estimated in practice. The dotted lines on each plot show the power of this test when ignoring the uncertainty in the estimated between‐study variance, or equivalently as the sample size tends towards infinity. Note that Δ is an increasing function in k, so that as the sample size increases Δ corresponds to a decreasing effect δFigure 1 shows that conventional methods for power analysis under the random‐effects model, that do not allow for the fact that τ2 must be estimated, serve as a reasonable guide to the actual power, especially for large k. Hence, ignoring the uncertainty in τ2 in the power calculation is not a very serious source of concern. However, small k is extremely common place in practice,22 and we can also see that the conventional method performs least accurately in such instances.
We also investigated this issue empirically using a database of 1991 meta‐analyses from the Cochrane Database of Systematic Reviews (Issue 1, 2008). We used the log risk ratio as the measure of treatment effect; we had access to the raw count data so that any differences in the method of analysis used in the individual reviews presented no difficulties. Most Cochrane reviews contain multiple meta‐analyses, corresponding to different pairwise comparisons of interventions and different outcomes examined. Davey et al22 classified each meta‐analysis by outcome type, the type of interventions compared and the medical specialty. Here, we use data on the first reported binary outcome meta‐analysis within each of the 1991 Cochrane reviews reporting at least 1 binary outcome meta‐analysis in the full database extracted by Davey et al.22 We performed retrospective power calculations for all 1991 random‐effects meta‐analyses using the analytical (section 4.3.1) and Monte Carlo (section 4.3.2) approaches, that do not, and do, allow for the uncertainty in τ2, respectively. Now that we apply the random‐effects model to the Cochrane data, it is pertinent to recall that this model can be a quite crude approximation in practice. For example, see Hoaglin23 and Shuster and Walker24 for good discussions of this issue.
For each meta‐analysis, we took δ to be the absolute value of the random‐effects pooled estimate of the average relative risk (though its sign is irrelevant, as explained above) and τ2 to be the DerSimonian and Laird estimate of the between‐study variance. By taking δ to be the (absolute) pooled estimate in this way, we perform retrospective power calculations for the null hypothesis that there is no treatment effect. We used 10 000 iterations when using the Monte Carlo approach, and the resulting 1991 pairs of powers are shown in Figure 2, where the powers obtained using the Monte Carlo method (section 4.3.2.) are shown on the vertical axis, and the powers obtained using the analytical approach (section 4.3.1.) are shown on the horizontal axis.

when performing power calculations. This figure shows the results of the empirical investigation of power in 1991 meta‐analyses. A line of equality is also shown
Figure 2 shows that the powers are in reasonable agreement. Hence ignoring the uncertainty in τ2 in the analytical approach continues to appear not to be a very serious concern. Despite this, it is also evident that the power calculated from the Monte Carlo method can differ substantially from the analytical power. This is because the actual powers of meta‐analysis hypothesis tests differ from the powers obtained using methods that ignore the uncertainty in estimates of τ2. The analytical powers in Figure 2 are bounded below by 0.05, but this is not the case for the Monte Carlo method because of the approximate nature of the standard random‐effects methods for meta‐analysis; when the between‐study variance is zero or very small, the standard methods are conservative so that actual powers of less than 0.05 are possible. More commonly, in other instances where the analytical power is very low, much higher powers using the Monte Carlo method are possible. This too is an artifact of the approximate nature of standard methods for random‐effects meta‐analysis, where if there are very small numbers of studies and very considerable between‐study variation, standard methods for random‐effects meta‐analysis are highly anticonservative and artificially high powers can be obtained; this is also evident in the top left hand plot in Figure 1 for k=3.
We conclude that it is desirable, but not essential, to use methods for power analysis under the random‐effects model that take into account the uncertainty in the estimation of τ2 when making inferences about the average effect.
5.4 Using the method assuming all studies are the same “size” to serve as a guide for retrospective power calculations
The suggestion in section 4.3.3 was to use an analytical approach where all studies are the same size, where the within‐study standard error was taken to be the square root of corresponding typical within‐study variance suggested by Higgins and Thompson.21 This type of approach has been suggested previously (eg, Jackson and Bowden,25) for obtaining a guide to how methods for meta‐analysis perform. It is tempting to consider this type of approach in large scale empirical investigations, like the one below, because the analytical results are obtained almost instantly. But this computational ease comes at the price of using an analogous meta‐analysis to the one that has been observed to obtain an indication of the power in the real meta‐analysis. We therefore also investigated this particular issue carefully using the database of meta‐analyses from the Cochrane Database of Systematic Reviews.
Specifically, we compared the 1991 powers obtained using the Monte Carlo method (as explained in the previous section) to those from the analytical approach in section 4.3.3 (top plot in Figure 3). We can see that, in general, the analytical powers are in good agreement with those obtained using simulation and the observed distribution of the within‐study variances. Given the relative computational and conceptual simplicity of the analytical powers, Figure 3 provides strong evidence that it is more than adequate for giving a good indication of the power. However, the agreement between the 2 powers for some meta‐analyses is not so strong, and a good indicator of whether this is the case or not is of course the amount of variation in the within‐study variances; the analytical powers generally agree less well for meta‐analyses with the greatest variation in study sizes, as expected.

. Lines of equality are also shown but are barely visible
We also considered using alternatives to the typical within‐study variance (Equation 10) as a measure of typical study size, such as the mean and median of the within‐study variances or the corresponding standard deviations. An alternative that also resulted in good agreement was taking the typical within‐study standard deviation to be the reciprocal of the mean of the within‐study “precisions,” ie, instead taking
, and the corresponding results are shown in the bottom plot in Figure 3. However, no obvious alternative resulted in any visible improvement to the agreement level between the powers shown in Figure 3, which indicates that performing power calculations that take all studies to be the same size can give a reasonable indication of the power but also that powers of this type are not very accurate in every case.
In conclusion, it would appear to be desirable, but not essential, to take the distribution of the within‐study variances and the uncertainty in the estimation of τ2 into account. The Monte Carlo method described in section 4.3.2 is computationally feasible and achieves both aims so it would seem reasonable to make the general recommendation that this method should be used for retrospective power calculations.
5.5 Using the methodology assuming all studies are the same “size” to perform power analyses at the planning stage
The method in Section 4.3.3. is, despite its limitations, useful for power calculations at the planning stage, where detailed knowledge of the distribution of the within‐study variances is unlikely to be available. We therefore suggest that this particular method should be considered when performing this type of power analysis. Here, the analyst specifies the anticipated number of studies, a typical within‐study variance and an I2 statistic. The typical within‐study variance can be obtained from the relevant formulae for their calculation them and using information such as a representative study size. This is quite a lot of information to posit, and a variety of possibilities could be explored. Then figures like those in Figure 1 may be used to give an indication of the value of Δ, and hence δ, that is needed to obtain powers of interest. In the supplementary materials that accompany this paper, we provide R code to produce plots like those in Figure 1 with an arbitrary k, to facilitate power calculations at the planning stage in this way.
6 AN EMPIRICAL INVESTIGATION COMPARING META‐ANALYSIS AND STUDY SPECIFIC POWERS UNDER THE RANDOM‐EFFECTS MODEL
Now that we have determined the most suitable way to perform retrospective power calculations under the random‐effects model, we are able to compare the meta‐analysis and average study specific powers. Specifically, we will compare the powers obtained in Section 5 using the Monte Carlo method to the average study specific power from Equation 5 with the same values of δ and τ2. By taking δ to be the pooled estimates, we therefore continue to assume that interest lies in testing the null hypothesis of no treatment effect.
The meta‐analysis power was found to be greater than the average study specific power in 303/609 (49.8%) random‐effects meta‐analyses where k=2. For k=3, this was 211/322 (65.5%); for k=4, this was 170/236 (72.0%); for k=5, this was 134/169 (79.3%); for k=6,7,8,9, this was 290/355 (81.7%); and for
, this was 263/300 (87.7%). We used these groups to provide reasonably large proportions of meta‐analyses in each group. A trend where the power increases with the number of studies is clear, as we should expect, although the observational nature of this conclusion should be emphasised because it does not control for other important factors such as the assumed average effect or the variance structures in the meta‐analyses that contribute to each of the 6 groupings. Under the strong assumption that our data are representative of meta‐analysis datasets, we estimate that random‐effects meta‐analysis results in an increase in power in 1371/1991 of meta‐analyses, which is just under 70%. However, it is clear that our sample of meta‐analyses contains many meta‐analyses where k is small; an increase in power when using random‐effects meta‐analysis is much more likely in subject areas where k is typically much larger.
Using the median, rather than the mean, study specific power in this comparison resulted in rather similar proportions of 303/609 (49.8%), 215/322 (66.8%), 173/236 (73.3%), 135/169 (79.9%), 302/355 (85.1%), and 269/300 (89.7%); using the median makes the power of the meta‐analysis look slightly better in this comparison. Instead using the maximum study specific power in this comparison resulted in proportions of 135/609 (22.2%), 107/322 (33.2%), 105/236 (44.5%), 88/169 (52.1%), 194/355 (54.6%), and 222/300 (74.0%). This suggests that in many meta‐analyses, the largest study will possess more power than the random‐effects meta‐analysis that it contributes to.
The overall picture from this empirical investigation is that if there are less than 5 studies then obtaining less power from the random‐effects meta‐analysis than from the individual studies that contribute to this meta‐analysis is quite likely in practice. Not only is statistical inference most difficult in this type of situation,24 but also it is less worthwhile in such cases. Most meta‐analysis powers are greater than the average study specific power however, so this investigation does not entirely discourage the use of meta‐analysis to obtain greater power. Despite this, our investigation certainly challenges the notion that meta‐analyses necessarily provide greater power.
7 DISCUSSION
We have investigated three different methods for performing power calculations for random‐effects meta‐analyses. We suggest that the Monte Carlo method should be used for retrospective power calculations because it is computationally feasible and allows for the uncertainty in the estimate of τ2 and the distribution of the within‐study variances. We also suggest that our new approximate analytical method is very suitable power calculations at the planning stage. We suspect that advocates of random‐effects meta‐analysis will find the comparison of these two powers disappointing. Researchers working in very different application areas to those represented in the Cochrane database might argue that key parameters, such as the number of studies and effect sizes, differ considerably in their subject area, so that our empirical conclusion that meta‐analysis powers are disappointing does not necessarily apply in their work. Our analysis in section 3 demonstrates that meta‐analysis necessarily results in an increase in power when the data are correctly assumed to be homogenous. Combining this observation with our findings under the random‐effects model clarifies that between‐study heterogeneity has serious consequences for the power of meta‐analyses. The main difficulty for obtaining high power in random‐effects meta‐analysis would seem to be due to the presence of considerable between‐study heterogeneity. However, in small samples, there are further difficulties associated with estimating this parameter.
In meta‐analyses which lack power, precision in estimation is also lacking, and therefore the summary intervention effect is imprecisely estimated. Thorlund et al26 have warned that if meta‐analyses are performed too early, before enough studies are available, there is a danger that incorrect conclusions may be drawn. They recommend therefore that results from underpowered meta‐analyses are interpreted with caution. Trial sequential analysis methods proposed by Wetterslev et al27 can be used to evaluate whether or not a particular meta‐analysis contains enough information to be regarded as providing conclusive evidence. Ideally, extremely underpowered meta‐analyses should not be performed. However, we recognise that it is often difficult for researchers to know how many eligible studies provide data for any given outcome until most of the reviewing work has been performed, and at that stage withholding the results might result in reporting bias. In some settings, a lack of power is caused by high observed heterogeneity rather than by the number of studies, and this is even more difficult to predict in advance.
We have assumed that the conventional DerSimonian and Laird method for random‐effects meta‐analysis is used throughout. Calculating power using further alternative methods for random‐effects meta‐analysis is also possible, for example, using an altenative estimator of the between‐study variance.20 The Monte Carlo method that we advocate for retrospective power calculations can easily be adapted for other τ2 estimators, and our approximate analytical method described in section 4.3.3 applies to a variety of these estimators, as we have explained. Other modifications and small sample corrections have been suggested (eg)28, 29 which inevitably have some implications for the power. For example, the artificially small values of Δ that achieve low powers, as seen in Figure 1, will not result in this difficulty when using the Hartung and Knapp modification in conjunction with common estimators of τ2 when all studies are the same size. This is because the modification results in exact inference for this special case under the random‐effects model.30 Investigating the power when using more sophisticated methods is an important avenue for further research. In particular, for binary outcome data such that from the Cochrane database, methods that use binomial within‐study distributions31, 32 are in principle preferable to those that we have used here. However, the power formulae that we have provided have the merit of being simple and transparent and, provided that they can be shown to represent the amount of information and so the power when using more sophisticated methods, they may prove valuable long after current methodologies for meta‐analysis, such as the DerSimonian and Laird method, might be consigned to history.
Our methodology allows for the testing that the true effect is any value μ0, but in our empirical investigations, we have only explored the null value μ0=0. There is also interest in testing for clinically significant effects, and indeed we suggest that this should be considered more often in application. However, the testing of the null hypothesis that there is no treatment effect is so ubiquitous in application that we have restricted our investigation to this special case. Investigating meta‐analysis and study specific powers to detect other effects of interest could form the subject of future work.
There are many powerful and persuasive reasons for performing meta‐analyses; the desire to increase the power to detect an effect is just one such reason. Our investigations show that random‐effects meta‐analyses can achieve this aim and that they generally do. Provided that sufficient numbers of studies can be found then a gain in power is of course assured. However, we have found that the powers of real random‐effects meta‐analyses compare less favourably to the powers of the studies that contribute to them than many might suppose. As we also pointed out in the introduction, the standard meta‐analysis methods fail to provide the nominal significance level. We have not taken this into account, which means that some of the power that we have attributed to random‐effects meta‐analyses is likely to be artificial and due to using statistical methods that are often anticonservative. Hence, the observation that standard methods for meta‐analysis are merely approximate strengthens our case that random‐effects meta‐analyses possess less power, relative to the studies that contribute to them, than many might otherwise suppose. Our findings in no way diminish the valuable impact that systematic reviews and meta‐analyses have had, but do lead us to conclude that the notion that meta‐analyses necessarily increase power is one that we should be much more critical of.
In summary, we have provided new methods for power analysis for random‐effects meta‐analysis and have investigated how the power of this type of meta‐analysis compares to that of the studies that contribute outcome data. We hope that our new methods are a useful addition to the literature and that this article will serve to emphasise the importance of considerations of power in meta‐analysis. Finally, we hope that our empirical investigations will make applied analysts think more critically about whether random‐effects meta‐analyses, when applied to highly heterogeneous datasets with very few studies, are likely to provide more power than individual studies.
REFERENCES
Citing Literature
Number of times cited according to CrossRef: 88
- Terri D. Pigott, Power of Statistical Tests for Subgroup Analysis in Meta-Analysis, Design and Analysis of Subgroups with Biopharmaceutical Applications, 10.1007/978-3-030-40105-4_17, (347-368), (2020).
- Kelly A. Romano, Kristin E. Heron, Rachel Amerson, Lindsay M. Howard, Rachel I. MacIntyre, Tyler B. Mason, Changes in disordered eating behaviors over 10 or more years: A meta‐analysis, International Journal of Eating Disorders, 10.1002/eat.23288, 53, 7, (1034-1055), (2020).
- James A Watson, Cormac G Ryan, Greg Atkinson, Philip Williamson, Dominic Ellington, Robbie Whittle, John Dixon, Denis J Martin, Inter-individual differences in the responses to pain neuroscience education in adults with chronic musculoskeletal pain: A systematic review and meta-analysis of randomised controlled trials, The Journal of Pain, 10.1016/j.jpain.2020.03.006, (2020).
- Lianne Wood, Nadine E Foster, Martyn Lewis, Annette Bishop, Exercise interventions for persistent non-specific low back pain – does matching outcomes to treatment targets make a difference? A systematic review and meta-analysis, The Journal of Pain, 10.1016/j.jpain.2020.04.002, (2020).
- A L Barker, Sze-Ee Soh, Kerrie M Sanders, Julie Pasco, Sundeep Khosla, Peter R Ebeling, Stephanie A Ward, Geeske Peeters, Jason Talevski, Robert G Cumming, Ego Seeman, John J McNeil, Aspirin and fracture risk: a systematic review and exploratory meta-analysis of observational studies, BMJ Open, 10.1136/bmjopen-2018-026876, 10, 2, (e026876), (2020).
- David T Turner, Elena Riedel, Loulou Hassan Kobeissi, Eirini Karyotaki, Claudia Garcia-Moreno, Lale Say, Pim Cuijpers, Psychosocial interventions for intimate partner violence in low and middle income countries: A meta-analysis of randomised controlled trials, Journal of Global Health, 10.7189/jogh.10.010409, 10, 1, (2020).
- Isabella Ingram, Peter J. Kelly, Frank P. Deane, Amanda L. Baker, Melvin C. W. Goh, Dayle K. Raftery, Genevieve A. Dingle, Loneliness among people with substance use problems: A narrative systematic review, Drug and Alcohol Review, 10.1111/dar.13064, 39, 5, (447-483), (2020).
- Delia Hillmayr, Lisa Ziernwald, Frank Reinhold, Sarah I. Hofer, Kristina M. Reiss, The potential of digital tools to enhance mathematics and science learning in secondary schools: A context-specific meta-analysis, Computers & Education, 10.1016/j.compedu.2020.103897, (103897), (2020).
- Olga Sánchez de Ribera, Nicolás Trajtenberg, Larissa S. Christensen, Evaluating the quality of meta-analytical reviews using the AMSTAR-2: A systematic review of meta-analytical reviews regarding child sexual abuse interventions, Child Abuse & Neglect, 10.1016/j.chiabu.2020.104463, 104, (104463), (2020).
- Ronny Scherer, Fazilat Siddiq, Bárbara Sánchez Viveros, A meta-analysis of teaching and learning computer programming: Effective instructional approaches and conditions, Computers in Human Behavior, 10.1016/j.chb.2020.106349, (106349), (2020).
- Getúlio Rodrigues de Oliveira Filho, Raquel Spilere Kammer, Heloísa de Cássia dos Santos, Duloxetine for the treatment acute postoperative pain in adult patients: A systematic review with meta-analysis, Journal of Clinical Anesthesia, 10.1016/j.jclinane.2020.109785, 63, (109785), (2020).
- Elisa S. Danthinne, Francesca E. Giorgianni, Rachel F. Rodgers, Labels to prevent the detrimental effects of media on body image: A systematic review and meta‐analysis, International Journal of Eating Disorders, 10.1002/eat.23242, 53, 5, (647-661), (2020).
- Annika Theodoulou, Jeganath Krishnan, Edoardo Aromataris, Risk of complications in patients who are obese following upper limb arthroplasty: A systematic review and meta-analysis, Obesity Research & Clinical Practice, 10.1016/j.orcp.2020.01.001, (2020).
- Dawei Yang, Dan Li, Jinshui Li, Zhenghan Yang, Zhenchang Wang, Systematic review: The diagnostic efficacy of gadoxetic acid-enhanced MRI for liver fibrosis staging, European Journal of Radiology, 10.1016/j.ejrad.2020.108857, (108857), (2020).
- Victoria Leclercq, Charlotte Beaudart, Sara Ajamieh, Ezio Tirelli, Olivier Bruyère, Methodological quality of meta-analyses indexed in PsycINFO: leads for enhancements: a meta-epidemiological study, BMJ Open, 10.1136/bmjopen-2019-036349, 10, 8, (e036349), (2020).
- Lauren Orr, Keith Reisinger-Kindle, Amrita Roy, Lisa Levine, Katherine Connolly, Paul Visintainer, Corina N. Schoen, Combination of Foley and prostaglandins versus Foley and oxytocin for cervical ripening: a network meta-analysis, American Journal of Obstetrics and Gynecology, 10.1016/j.ajog.2020.05.007, (2020).
- Kai Wei Lee, Hong Chuan Loh, Seng Choi Chong, Siew Mooi Ching, Navin Kumar Devaraj, Maiza Tusimin, Habibah Abdul Hamid, Fan Kee Hoo, Prevalence of anxiety among gestational diabetes mellitus patients: A systematic review and meta-analysis, World Journal of Meta-Analysis, 10.13105/wjma.v8.i3.275, 8, 3, (275-284), (2020).
- Dayle Raftery, Peter J. Kelly, Frank P. Deane, Amanda L. Baker, Isabella Ingram, Melvin C.W. Goh, Dan I. Lubman, Greg Carter, Alyna Turner, Olivia M. Dean, Barbara Sinclair, Rebecca McKetin, Insight in substance use disorder: A systematic review of the literature, Addictive Behaviors, 10.1016/j.addbeh.2020.106549, (106549), (2020).
- Vinod K. Gupta, Minsuk Kim, Utpal Bakshi, Kevin Y. Cunningham, John M. Davis, Konstantinos N. Lazaridis, Heidi Nelson, Nicholas Chia, Jaeyun Sung, A predictive index for health status using species-level gut microbiome profiling, Nature Communications, 10.1038/s41467-020-18476-8, 11, 1, (2020).
- Ivana Hanzlíková, Josie Athens, Kim Hébert-Losier, Factors influencing the Landing Error Scoring System: Systematic review with meta-analysis, Journal of Science and Medicine in Sport, 10.1016/j.jsams.2020.08.013, (2020).
- Taymour Mostafa, Ibrahim A. Abdel-Hamid, Mai Taymour, Omar I. Ali, Gene Variants in Premature Ejaculation: Systematic Review and Future Directions, Sexual Medicine Reviews, 10.1016/j.sxmr.2020.07.002, (2020).
- Practice Guidelines for Central Venous Access 2020, Anesthesiology, 10.1097/ALN.0000000000002864, 132, 1, (8-43), (2020).
- Andrew R. Stephens, Angela P. Presson, Mary M. McFarland, Chong Zhang, Kai Sirniö, Marjolein A.M. Mulders, Niels W.L. Schep, Andrew R. Tyser, Nikolas H. Kazmers, Volar Locked Plating Versus Closed Reduction and Casting for Acute, Displaced Distal Radial Fractures in the Elderly, Journal of Bone and Joint Surgery, 10.2106/JBJS.19.01442, 102, 14, (1280-1288), (2020).
- Neil Dagnall, Kenneth G. Drinkwater, Ciarán O’Keeffe, Annalisa Ventola, Brian Laythe, Michael A. Jawer, Brandon Massullo, Giovanni B. Caputo, James Houran, Things That Go Bump in the Literature: An Environmental Appraisal of “Haunted Houses”, Frontiers in Psychology, 10.3389/fpsyg.2020.01328, 11, (2020).
- Elena M. C. Riedel, David T. Turner, Loulou Hassan Kobeissi, Eirini Karyotaki, Lale Say, Pim Cuijpers, The impact of psychosocial interventions on condom and contraceptive use in LMICs: Meta-analysis of randomised controlled trials, Global Public Health, 10.1080/17441692.2020.1744679, (1-18), (2020).
- David T. Turner, Mandy X. Hu, Ellen Generaal, Daniel Bos, M. Kamran Ikram, Alis Heshmatollah, Lana Fani, M. Arfan Ikram, Brenda W. J. H. Penninx, Pim Cuijpers, Physical Exercise Interventions Targeting Cognitive Functioning and the Cognitive Domains in Nondementia Samples: A Systematic Review of Meta-Analyses, Journal of Geriatric Psychiatry and Neurology, 10.1177/0891988720915523, (089198872091552), (2020).
- Chelsea L Ratcliff, Ye Sun, Overcoming Resistance Through Narratives: Findings from a Meta-Analytic Review, Human Communication Research, 10.1093/hcr/hqz017, (2020).
- Damoon Ashtary-Larky, Reza Bagheri, Amir Abbasnezhad, Grant M. Tinsley, Meysam Alipour, Alexei Wong, Effects of gradual weight loss v . rapid weight loss on body composition and RMR: a systematic review and meta-analysis , British Journal of Nutrition, 10.1017/S000711452000224X, (1-12), (2020).
- Wenge Xu, Hai-Ning Liang, Nilufar Baghaei, Bing Wu Berberich, Yong Yue, Health Benefits of Digital Videogames for the Aging Population: A Systematic Review, Games for Health Journal, 10.1089/g4h.2019.0130, (2020).
- Nathan Walter, John J. Brooks, Camille J. Saucier, Sapna Suresh, Evaluating the Impact of Attempts to Correct Health Misinformation on Social Media: A Meta-Analysis, Health Communication, 10.1080/10410236.2020.1794553, (1-9), (2020).
- Kirsti Jane Loughran, Greg Atkinson, Marla K Beauchamp, John Dixon, Denis Martin, Shaera Rahim, Samantha Louise Harrison, Balance impairment in individuals with COPD: a systematic review with meta-analysis, Thorax, 10.1136/thoraxjnl-2019-213608, (thoraxjnl-2019-213608), (2020).
- David T Turner, Simone Burger, Filip Smit, Lucia R Valmaggia, Mark van der Gaag, What Constitutes Sufficient Evidence for Case Formulation–Driven CBT for Psychosis? Cumulative Meta-analysis of the Effect on Hallucinations and Delusions, Schizophrenia Bulletin, 10.1093/schbul/sbaa045, (2020).
- Katie Eklund, Matthew K. Burns, Kari Oyen, Sarah DeMarchena, Elizabeth M. McCollom, Addressing Chronic Absenteeism in Schools: A Meta-Analysis of Evidence-Based Interventions, School Psychology Review, 10.1080/2372966X.2020.1789436, (1-17), (2020).
- Hannah Dickson, Emily P. Hedges, Shin Y. Ma, Alexis E. Cullen, James H. MacCabe, Matthew J. Kempton, Johnny Downs, Kristin R. Laurens, Academic achievement and schizophrenia: a systematic meta-analysis, Psychological Medicine, 10.1017/S0033291720002354, (1-17), (2020).
- C. Cotrena, L. Damiani Branco, A. Ponsoni, C. Samamé, F. Milman Shansis, R. Paz Fonseca, Executive functions and memory in bipolar disorders I and II: new insights from meta‐analytic results, Acta Psychiatrica Scandinavica, 10.1111/acps.13121, 141, 2, (110-130), (2019).
- Raquel Ruiz‐Íñiguez, Miguel Á. Santed Germán, Francisco A. Burgos‐Julián, Cintia Díaz‐Silveira, Ana Carralero Montero, Effectiveness of mindfulness‐based interventions on anxiety for children and adolescents: A systematic review and meta‐analysis, Early Intervention in Psychiatry, 10.1111/eip.12849, 14, 3, (263-274), (2019).
- M. Bekheit, R. Smith, G. Ramsay, F. Soggiu, M. Ghazanfar, I. Ahmed, Meta‐analysis of laparoscopic transcystic versus transcholedochal common bile duct exploration for choledocholithiasis, BJS Open, 10.1002/bjs5.50132, 3, 3, (242-251), (2019).
- Janus Christian Jakobsen, Christian Ovesen, Per Winkel, Jørgen Hilden, Christian Gluud, Jørn Wetterslev, Power estimations for non-primary outcomes in randomised clinical trials, BMJ Open, 10.1136/bmjopen-2018-027092, 9, 6, (e027092), (2019).
- Donald J Robinaugh, Meredith J Ward, Emma R Toner, Mackenzie L Brown, Olivia M Losiewicz, Eric Bui, Scott P Orr, Assessing vulnerability to panic: a systematic review of psychological and physiological responses to biological challenges as prospective predictors of panic attacks and panic disorder, General Psychiatry, 10.1136/gpsych-2019-100140, 32, 6, (e100140), (2019).
- Philipp von Trott, Shiao Li Oei, Christina Ramsenthaler, Re: Acupuncture for breathlessness in advanced diseases: methodological issues. – Reply, Journal of Pain and Symptom Management, 10.1016/j.jpainsymman.2019.11.019, (2019).
- Ioana R. Podina, Andreea Vîslă, Liviu A. Fodor, Christoph Flückiger, Is there a sleeper effect of exposure-based vs. cognitive-only intervention for anxiety disorders? A longitudinal multilevel meta-analysis, Clinical Psychology Review, 10.1016/j.cpr.2019.101774, (101774), (2019).
- Ronny Scherer, Fazilat Siddiq, The relation between students’ socioeconomic status and ICT literacy: Findings from a meta-analysis, Computers & Education, 10.1016/j.compedu.2019.04.011, 138, (13-32), (2019).
- Claire Hawcroft, Rachael Hughes, Amira Shaheen, Jinan Usta, Hannah Elkadi, Tom Dalton, Khadijah Ginwalla, Gene Feder, Prevalence and health outcomes of domestic violence amongst clinical populations in Arab countries: a systematic review and meta-analysis, BMC Public Health, 10.1186/s12889-019-6619-2, 19, 1, (2019).
- Bernd Weiß, Michael Wagner, Meta-Analyse, Handbuch Methoden der empirischen Sozialforschung, 10.1007/978-3-658-21308-4, (1511-1522), (2019).
- Cosette Fakih El Khoury, Mirey Karavetian, Ruud J.G. Halfens, Rik Crutzen, Lama Khoja, Jos M.G.A. Schols, The Effects of Dietary Mobile Apps on Nutritional Outcomes in Adults with Chronic Diseases: A Systematic Review, Journal of the Academy of Nutrition and Dietetics, 10.1016/j.jand.2018.11.010, (2019).
- James A Watson, Cormac G Ryan, Lesley Cooper, Dominic Ellington, Robbie Whittle, Michael Lavender, John Dixon, Greg Atkinson, Kay Cooper, Denis J Martin, Pain neuroscience education for adults with chronic musculoskeletal pain: a mixed-methods systematic review and meta-analysis, The Journal of Pain, 10.1016/j.jpain.2019.02.011, (2019).
- Michael K. Musyl, Eric L. Gilman, Meta‐analysis of post‐release fishing mortality in apex predatory pelagic sharks and white marlin, Fish and Fisheries, 10.1111/faf.12358, 20, 3, (466-500), (2019).
- U. Matterne, S.E. Baumeister, C.J. Apfelbacher, Suicidality and risk of suicidality in psoriasis: a critical appraisal of two systematic reviews and meta‐analyses, British Journal of Dermatology, 10.1111/bjd.18108, 181, 4, (717-721), (2019).
- Lorenzo Zanetto, Liviana Da Dalt, Marco Daverio, Joel Dunning, Anna Chiara Frigo, Lise E. Nigrovic, Silvia Bressan, Systematic review and meta‐analysis found significant risk of brain injury and neurosurgery in alert children after a post‐traumatic seizure, Acta Paediatrica, 10.1111/apa.14810, 108, 10, (1841-1849), (2019).
- Xuangao Wu, Hye Jeong Yang, Min Jung Kim, Ting Zhang, Jing Yi Qiu, Sunmin Park, Association between PTPN22‐1123G/C and susceptibility to rheumatoid arthritis: A systematic review and meta‐analysis, International Journal of Rheumatic Diseases, 10.1111/1756-185X.13535, 22, 5, (769-780), (2019).
- Sarah Barakat, Sarah Maguire, Kathryn E. Smith, Tyler B. Mason, Ross D. Crosby, Stephen Touyz, Evaluating the role of digital intervention design in treatment outcomes and adherence to eTherapy programs for eating disorders: A systematic review and meta‐analysis, International Journal of Eating Disorders, 10.1002/eat.23131, 52, 10, (1077-1094), (2019).
- Annika Theodoulou, Jeganath Krishnan, Edoardo Aromataris, Risk of poor outcomes in patients who are obese following total shoulder arthroplasty and reverse total shoulder arthroplasty: a systematic review and meta-analysis, Journal of Shoulder and Elbow Surgery, 10.1016/j.jse.2019.06.017, 28, 11, (e359-e376), (2019).
- Jagjit Mankelow, Cormac Ryan, Paul Taylor, Dominic Simpson, Effectiveness of pain education to improve pain related knowledge, attitudes and behaviours in health care students and professionals: a systematic review protocol, International Journal of Therapy and Rehabilitation, 10.12968/ijtr.2018.0083, 26, 8, (1-8), (2019).
- Caroline Marker, Timo Gnambs, Markus Appel, Exploring the myth of the chubby gamer: A meta-analysis of studies on sedentary video gaming and body mass, Social Science & Medicine, 10.1016/j.socscimed.2019.05.030, (2019).
- Michiel A.I. Jsseldijk, Melina Shoni, Charles Siegert, Jan Seegers, Anton K.C. van Engelenburg, Thomas C. Tsai, Abraham Lebenthal, Richard P.G. ten Broek, Oncological outcomes of lobar resection, segmentectomy and wedge resection for T1a non-small cell lung carcinoma: a systematic review and meta-analysis, Seminars in Thoracic and Cardiovascular Surgery, 10.1053/j.semtcvs.2019.08.004, (2019).
- Golareh Agha, Michael M. Mendelson, Cavin K. Ward-Caviness, Roby Joehanes, TianXiao Huan, Rahul Gondalia, Elias Salfati, Jennifer A. Brody, Giovanni Fiorito, Jan Bressler, Brian H. Chen, Symen Ligthart, Simonetta Guarrera, Elena Colicino, Allan C. Just, Simone Wahl, Christian Gieger, Amy R. Vandiver, Toshiko Tanaka, Dena G. Hernandez, Luke C. Pilling, Andrew B. Singleton, Carlotta Sacerdote, Vittorio Krogh, Salvatore Panico, Rosario Tumino, Yun Li, Guosheng Zhang, James D. Stewart, James S. Floyd, Kerri L. Wiggins, Jerome I. Rotter, Michael Multhaup, Kelly Bakulski, Steven Horvath, Philip S. Tsao, Devin M. Absher, Pantel Vokonas, Joel Hirschhorn, M. Daniele Fallin, Chunyu Liu, Stefania Bandinelli, Eric Boerwinkle, Abbas Dehghan, Joel D. Schwartz, Bruce M. Psaty, Andrew P. Feinberg, Lifang Hou, Luigi Ferrucci, Nona Sotoodehnia, Giuseppe Matullo, Annette Peters, Myriam Fornage, Themistocles L. Assimes, Eric A. Whitsel, Daniel Levy, Andrea A. Baccarelli, Blood Leukocyte DNA Methylation Predicts Risk of Future Myocardial Infarction and Coronary Heart Disease, Circulation, 10.1161/CIRCULATIONAHA.118.039357, 140, 8, (645-657), (2019).
- Nathan Walter, Jonathan Cohen, R. Lance Holbert, Yasmin Morag, Fact-Checking: A Meta-Analysis of What Works and for Whom, Political Communication, 10.1080/10584609.2019.1668894, (1-26), (2019).
- Andrea Sandstrom, Rudolf Uher, Barbara Pavlova, Prospective Association between Childhood Behavioral Inhibition and Anxiety: a Meta-Analysis, Journal of Abnormal Child Psychology, 10.1007/s10802-019-00588-5, (2019).
- Amber Berdenis van Berlekom, Cita H Muflihah, Gijsje J L J Snijders, Harold D MacGillavry, Jinte Middeldorp, Elly M Hol, René S Kahn, Lot D de Witte, Synapse Pathology in Schizophrenia: A Meta-analysis of Postsynaptic Elements in Postmortem Brain Studies, Schizophrenia Bulletin, 10.1093/schbul/sbz060, (2019).
- Ilan Cerna-Turoff, Hanna-Tina Fischer, Susannah Mayhew, Karen Devries, Violence against children and natural disasters: A systematic review and meta-analysis of quantitative evidence, PLOS ONE, 10.1371/journal.pone.0217719, 14, 5, (e0217719), (2019).
- Ari Alex Ramos, Amer Cavalheiro Hamdan, Liana Machado, A meta-analysis on verbal working memory in children and adolescents with ADHD, The Clinical Neuropsychologist, 10.1080/13854046.2019.1604998, (1-26), (2019).
- Nicholas Mikolajewicz, Svetlana V. Komarova, Meta-Analytic Methodology for Basic Research: A Practical Guide, Frontiers in Physiology, 10.3389/fphys.2019.00203, 10, (2019).
- Nada Salhab, Mirey Karavetian, Jeroen Kooman, Enrico Fiaccadori, Cosette F. El Khoury, Effects of intradialytic aerobic exercise on hemodialysis patients: a systematic review and meta-analysis, Journal of Nephrology, 10.1007/s40620-018-00565-z, (2019).
- Felix J Hüttner, Pascal Probst, Eva Kalkum, Matthes Hackbusch, Katrin Jensen, Alexis Ulrich, Jürgen Debus, Dirk Jäger, Markus K Diener, Addition of Platinum Derivatives to Fluoropyrimidine-Based Neoadjuvant Chemoradiotherapy for Stage II/III Rectal Cancer: Systematic Review and Meta-Analysis, JNCI: Journal of the National Cancer Institute, 10.1093/jnci/djz081, (2019).
- Arnaud J. P. E. Vincent, Willem S. van Hoogstraten, Antoinette Maassen Van Den Brink, Joost van Rosmalen, Bibi L. J. Bouwen, Extracranial Trigger Site Surgery for Migraine: A Systematic Review With Meta-Analysis on Elimination of Headache Symptoms, Frontiers in Neurology, 10.3389/fneur.2019.00089, 10, (2019).
- Anastasia Bougea, Leonidas Stefanis, George P Paraskevas, Evangelia Emmanouilidou, Kostas Vekrelis, Elisabeth Kapaki, Plasma alpha-synuclein levels in patients with Parkinson’s disease: a systematic review and meta-analysis, Neurological Sciences, 10.1007/s10072-019-03738-1, (2019).
- Naomi L. Burn, Matthew Weston, Neil Maguire, Greg Atkinson, Kathryn L. Weston, Effects of Workplace-Based Physical Activity Interventions on Cardiorespiratory Fitness: A Systematic Review and Meta-Analysis of Controlled Trials, Sports Medicine, 10.1007/s40279-019-01125-6, (2019).
- Robin L. Nabi, Nathan Walter, Neekaan Oshidary, Camille G. Endacott, Jessica Love-Nichols, Z. J. Lew, Alex Aune, Can Emotions Capture the Elusive Gain-Loss Framing Effect? A Meta-Analysis, Communication Research, 10.1177/0093650219861256, (009365021986125), (2019).
- Phillip K. Martin, Ryan W. Schroeder, Daniel H. Olsen, Halley Maloy, Anneliese Boettcher, Nathan Ernst, Hayrettin Okut, A systematic review and meta-analysis of the Test of Memory Malingering in adults: Two decades of deception detection, The Clinical Neuropsychologist, 10.1080/13854046.2019.1637027, (1-32), (2019).
- Jack Hollingdale, Emma Woodhouse, Susan Young, Aide Fridman, Will Mandy, Autistic spectrum disorder symptoms in children and adolescents with attention-deficit/hyperactivity disorder: a meta-analytical review, Psychological Medicine, 10.1017/S0033291719002368, (1-14), (2019).
- Haylee K. DeLuca, Shannon E. Claxton, Manfred H. M. Dulmen, The Peer Relationships of Those Who Have Experienced Adoption or Foster Care: A Meta‐Analysis, Journal of Research on Adolescence, 10.1111/jora.12421, 29, 4, (796-813), (2018).
- Leo E. Akioyamen, Jacques Genest, Shubham D. Shan, Happy Inibhunu, Anna Chu, Jack V. Tu, Anxiety, depression, and health-related quality of life in heterozygous familial hypercholesterolemia: A systematic review and meta-analysis, Journal of Psychosomatic Research, 10.1016/j.jpsychores.2018.03.170, 109, (32-43), (2018).
- Erlend G Faltinsen, Ole Jakob Storebø, Janus C Jakobsen, Kim Boesen, Theis Lange, Christian Gluud, Network meta-analysis: the highest level of medical evidence?, BMJ Evidence-Based Medicine, 10.1136/bmjebm-2017-110887, 23, 2, (56-59), (2018).
- Kathryn E. Smith, Tyler B. Mason, Jason M. Lavender, Rumination and eating disorder psychopathology: A meta-analysis, Clinical Psychology Review, 10.1016/j.cpr.2018.03.004, 61, (9-23), (2018).
- Annemarie Uhlig, Ali Seif Amir Hosseini, Jörg Simon, Joachim Lotz, Lutz Trojan, Marianne Schmid, Johannes Uhlig, Gender Specific Differences in Disease-Free, Cancer Specific and Overall Survival after Radical Cystectomy for Bladder Cancer: A Systematic Review and Meta-Analysis, Journal of Urology, 10.1016/j.juro.2017.11.150, 200, 1, (48-60), (2018).
- Nicholas Mikolajewicz, Ali Mohammed, Martin Morris, Svetlana V. Komarova, Mechanically stimulated ATP release from mammalian cells: systematic review and meta-analysis, Journal of Cell Science, 10.1242/jcs.223354, 131, 22, (jcs223354), (2018).
- Bethan Dalton, Savani Bartholdy, Lauren Robinson, Marco Solmi, Mohammad A.A. Ibrahim, Gerome Breen, Ulrike Schmidt, Hubertus Himmerich, A meta-analysis of cytokine concentrations in eating disorders, Journal of Psychiatric Research, 10.1016/j.jpsychires.2018.06.002, 103, (252-264), (2018).
- Taylor P. Kohn, Jaden R. Kohn, Ranjith Ramasamy, Effect of Sperm Morphology on Pregnancy Success via Intrauterine Insemination: A Systematic Review and Meta-Analysis, The Journal of Urology, 10.1016/j.juro.2017.11.045, 199, 3, (812-822), (2018).
- Felix J. Hüttner, Pascal Probst, Eva Kalkum, Matthes Hackbusch, Katrin Jensen, Alexis Ulrich, Markus W. Büchler, Markus K. Diener, Addition of platinum derivatives to neoadjuvant single-agent fluoropyrimidine chemoradiotherapy in patients with stage II/III rectal cancer: protocol for a systematic review and meta-analysis (PROSPERO CRD42017073064), Systematic Reviews, 10.1186/s13643-018-0678-9, 7, 1, (2018).
- Matthew Joe Grima, Mourad Boufi, Martin Law, Dan Jackson, Kate Stenson, Benjamin Patterson, Ian Loftus, Matt Thompson, Alan Karthikesalingam, Peter Holt, Editor's Choice – The Implications of Non-compliance to Endovascular Aneurysm Repair Surveillance: A Systematic Review and Meta-analysis, European Journal of Vascular and Endovascular Surgery, 10.1016/j.ejvs.2017.11.030, 55, 4, (492-502), (2018).
- Hung-Bin Sheu, Robert W. Lent, Matthew J. Miller, Lee T. Penn, Megan E. Cusick, Nancy N. Truong, Sources of self-efficacy and outcome expectations in science, technology, engineering, and mathematics domains: A meta-analysis, Journal of Vocational Behavior, 10.1016/j.jvb.2018.10.003, (2018).
- Nathan Walter, Sheila T. Murphy, How to unring the bell: A meta-analytic approach to correction of misinformation, Communication Monographs, 10.1080/03637751.2018.1467564, 85, 3, (423-441), (2018).
- Emanuele Felice Osimo, Katherine Beck, Tiago Reis Marques, Oliver D Howes, Synaptic loss in schizophrenia: a meta-analysis and systematic review of synaptic protein and mRNA measures, Molecular Psychiatry, 10.1038/s41380-018-0041-5, (2018).
- Joshua K. Swift, Bruce E. Wampold, Inclusion and exclusion strategies for conducting meta-analyses, Psychotherapy Research, 10.1080/10503307.2017.1405169, 28, 3, (356-366), (2017).
- Thelma B. González-Castro, Yazmín Hernandez-Diaz, Isela E. Juárez-Rojop, Lilia López-Narváez, Carlos A. Tovilla-Zárate, José M. Rodriguez-Perez, Juan P. Sánchez-de la Cruz, The role of the Cys23Ser (rs6318) polymorphism of the HTR2C gene in suicidal behavior, Psychiatric Genetics, 10.1097/YPG.0000000000000184, 27, 6, (199-209), (2017).
- Joshua Feinberg, Emil Eik Nielsen, Janette Greenhalgh, Juliet Hounsome, Naqash J Sethi, Sanam Safi, Christian Gluud, Janus C Jakobsen, Drug-eluting stents versus bare-metal stents for acute coronary syndrome, Cochrane Database of Systematic Reviews, 10.1002/14651858.CD012481.pub2, (2017).
- Hilary Norman, Andrea Oskis, Lisa Marzano, Mark Coulson, The relationship between self‐harm and alexithymia: A systematic review and meta‐analysis, Scandinavian Journal of Psychology, 10.1111/sjop.12668, 0, 0, (undefined).
- Tina Thomas, Yusra Ahmed, Mei Tan, Elena L. Grigorenko, Cognitive and Educational Interventions for Orphans and Vulnerable Children Affected by HIV/AIDS: A Meta‐Analysis, Child Development, 10.1111/cdev.13438, 0, 0, (undefined).




, in Equation
, so that Equation 

