Volume 7, Issue 1
Commentary
Open Access

Absolute or relative effects? Arm‐based synthesis of trial data

S. Dias

Corresponding Author

School of Social and Community Medicine, University of Bristol, USA

Correspondence to: S Dias, School of Social and Community Medicine, University of Bristol, Canynge Hall, 39 Whatley Road, Bristol, UK.

E‐mail: s.dias@bristol.ac.uk

Search for more papers by this author
A. E. Ades

School of Social and Community Medicine, University of Bristol, USA

Search for more papers by this author
First published: 13 October 2015
Citations: 41

We congratulate Hwanhee Hong and colleagues on another fascinating paper (Hong et al., 2015a) arguing the case for arm‐based models for meta‐analysis.

The standard approach to meta‐analysis is the contrast‐based model where the information that is pooled over trials is the information of the trial‐specific relative treatment effect, expressed for example as a log relative risk, log odds ratio, or as a mean treatment difference. In an arm‐based model, it is the absolute log risk, log odds, or mean outcome on each arm that are pooled.

There is no doubt that arm‐based models are an intriguing alternative to the accepted understanding of meta‐analysis, and that they provide a very elegant alternative approach to network meta‐analysis (NMA). However, readers of Research Synthesis Methods (RSM) will have no difficulty in recognising that arm‐based models represent a radical – even revolutionary – departure from current meta‐analytic practice.

In this commentary, we begin by outlining the key differences between arm‐ and contrast‐based meta‐analysis to make clear the true extent of the implications of the claims being made. We then argue that contrast based models are to be preferred on both theoretical and practical grounds.

Hong and colleagues present a number of arguments about the core assumptions of contrast‐based models, both in their RSM paper (Hong et al., 2015a) and previously (Zhang et al., 2014, Ohlssen et al., 2014, Zhang et al., 2015, Hong et al., 2015b), which we believe are mistaken. We offer some counter‐arguments and comment on the simulation study presented by Hong et al. (2015a).

1. Classic contrast‐based, contrast‐based plus baseline and arm‐based models

Hong et al. (2015a) contrast three kinds of models. We will call these (a) the “Classic Contrast‐Based” (Classic CB) model, which corresponds to the standard model for pair‐wise meta‐analysis (Higgins and Green, 2008), extended to NMA (Higgins and Green, 2008, Dias et al., 2013a, Lu and Ades, 2004, Cooper et al., 2009, Lu and Ades, 2006, Caldwell et al., 2005), (b) “Contrast‐Based plus Baseline” (CB + Baseline), and (c) “Arm Based” (AB). It is important to be clear about notation and terminology, so we start with pair‐wise meta‐analysis and then turn to NMA.

With the contrast‐based models, particularly those used in a frequentist framework, the data to be pooled are the relative effect measures, and in its most common implementation the absolute effects are not even available to estimate. Indexing trials by i, the data Di,XY are the observed relative effect measures comparing treatment Y to X, δi,XY, which in a Bayesian setting are the “shrunken” trial‐specific estimates, and dXY are the mean relative effects. Thus, the “Classic CB” model for pair‐wise (two‐treatment) data is stated as follows:
urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0001(1)
Arm‐based likelihoods (not to be confused with arm‐based models) have also been proposed (Prevost et al., 2000, Smith et al., 1995) particularly in a Bayesian context, partly because they avoid normal approximations for count data. This obliges us to introduce the parameter μi,X representing the absolute effect of the control arm X, on the appropriate scale (log risk, log odds, etc). The scale is determined by the link function G which transforms the relative effect parameter of interest, Θi,XY (for example a probability or rate), onto a scale where its effects can be assumed to be additive:
urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0002(2)

This form of contrast based model has always come in two flavours (Prevost et al., 2000). The one that corresponds exactly to the “Classic CB” model 1 treats the μi,X as unrelated nuisance parameters. This is the approach adopted in the NMA models developed in Bristol (Dias et al., 2013a, Lu and Ades, 2004, Prevost et al., 2000).

The alternative is to put another hierarchical model on the trial‐specific control‐arms: urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0003 (Achana et al., 2013, Dias et al., 2013c). This is the “CB + baseline model”, or alternatively a bivariate normal model on the μi,X and the δi,XY could be adopted (van Houwelingen et al., 1993).

To see how this idea works in NMA, we need to complicate the notation a little. First, we will consider trial‐specific treatment indices k = 1, 2, 3 …, and we distinguish between the relative effect of the k‐th treatment relative to the control treatment on that trial urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0004, and its effect relative to the reference treatment 1, urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0005, which may or may not be present in trial i. With this extended notation, the “Classic CB” model makes the trial‐specific control arms arbitrary and unrelated.
urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0006(3)

Note that often, subscript t1 is dropped from urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0007, as it is redundant. Clearly, there is no way of estimating an absolute effect for any treatment from the “Classic CB” model.

We note in passing that in another paper Hong et al. (2015b) claim that the assumption of exchangeability of “effects relative to an arbitrary and nonconstant baseline remains much to assume”. This misses the point that exchangeability of the effects δi,1k relative to treatment 1 implies exchangeability of all the δi,jk (Lu and Ades, 2009).

In the “CB + baseline” model, a model is placed on the absolute effect of reference treatment 1 in trial i. Note that it is not put on the trial control arms, as these are not necessarily the same treatment—care should be taken that only the effect of treatment 1 is modelled. This allows absolute effects on the chosen scale (for example the log‐odds or log), ak, to be calculated for any treatment based on the included trial data. The model can be written as
urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0008(4)

Note that in the “CB + baseline” model, information on the baseline will back‐propagate onto the relative effect parameters. It is this property of “CB + Baseline” models that makes the relative effect estimates vulnerable to misspecification of the absolute effects.

We can now set out the arm‐based models proposed by Hong et al. (2015a) as
urn:x-wiley:17592879:media:jrsm1184:jrsm1184-math-0009(5)

The first feature we notice is that there are no relative treatment effects in equation 5. Mean “treatment effects” on the linear predictor scale are manufactured from the treatment‐specific means as dXY = aY − aX. Note that while the CB modeller working on binary data is committed to specify whether effects are additive on a continuous, log, or log‐odds scale, the AB modeller is free to report risk differences, G− 1(aY) − G− 1(aX), relative risks G− 1(aY)/G− 1(aX), or odds ratios (G− 1(aY)G− 1(1 − aX))/(G− 1(aX)G− 1(1 − aY)). For binary outcomes, both probit and logit link functions have been proposed in arm‐based models, presumably giving virtually identical results. This “model‐free” approach to relative effect estimation has the advantage that it makes no assumptions, but, as we see below, it comes with a heavy price in posterior variance of the relative effects.

In the latest paper in RSM Hong et al. (2015a) extend AB modelling to multiple outcomes, and run simulations that compare “Classic CB”, “CB + baseline”, and “AB” models in various styles. In previous work developing AB models Zhang et al. (2014) have criticised the “Classic CB” model on the grounds that, because it contains no model for absolute effects, it cannot report overall event rates without making further assumptions (although some would see this as an advantage of “Classic CB”). Other papers have examined the concept of inconsistency in NMA (Zhang et al., 2015) and looked at the role of AB models in drug development (Ohlssen et al., 2014).

1.1. Comparing arm‐based and contrast based models

  1. Fundamentals of meta‐analysis
Although much of the arm‐based literature casts the debate as specifically relevant to NMA, it should be clear that the central claim being made is equally relevant to pair‐wise (i.e. two‐treatment) meta‐analysis (MA). The proposal that one should pool absolute arm effects rather than relative treatment effects goes against the entire tradition of MA, in which the absolute event rates have been viewed as nuisance parameters, and as highly variable, while the relative effects, expressed on a suitable scale, have been seen as relatively stable. One might go further, and say that arm‐based models run against the entire tradition of epidemiological statistics in which relative effects – for example of risk factors – are pooled in multiple 2‐by‐2 tables (Mantel and Haenszel, 1959, Zelen, 1971).

We also believe that arm‐based pooling effectively breaks randomisation, and in fact runs against the entire way in which randomised controlled trials are designed, analysed, and used.

While “AB” models take an extreme position, many of us have also argued against “CB + baseline” models, because, unless the baseline model is correctly specified, the relative effect estimates will be biased. For example, one often finds that outcomes on the control arm have improved over time, because less severe populations are entered into more recent trials while the relative effect has stayed constant. If one random effects model is put on the baselines and another on the relative effects, and both are estimated at the same time, we expect the baseline effects in earlier studies to be shrunk upwards, and relative effects under‐estimated, while the relative effects in more recent studies will be over‐estimated. Senn (2010) has argued forcefully against “CB + baseline” models for these reasons. “AB” models will be even more likely to produce biased treatment effects, if one accepts the commonly made assumption that it is the relative effects which are exchangeable across trials.

  1. Evidence synthesis for decision‐making
Working as we do on evidence synthesis in the decision making context, we are well aware the absolute event rates often play a more important role than the relative treatment effects when treatment recommendations are made. Thus, we fully accept that there needs to be a model for the absolute effect that one would see on at least one of the treatments, from which absolute effects can be recovered for all treatments.

Where we differ from proponents of AB models is that we believe that, while randomised controlled trials are unquestionably the best data sources to inform relative effects, the data sources that best inform the absolute effects might be cohort studies, a carefully selected subset of the trials included in the meta‐analysis, or expert opinion. Manufacturers making submissions on new products to the National Institute for Health and Care Excellence (NICE) in the UK are asked to explain the basis on which they have chosen the evidence for absolute effects, and this is quite distinct from trial evidence from which relative effects are sourced (National Institute for Health and Clinical Excellence, 2012).

The disadvantage of an arm based model is that it obliges the modeller to use exactly the same data sources for both absolute and relative effects, whereas the natural way to carry out health technology assessments of alternative treatments is to apply a “transportable” relative effect estimate to an independently estimated baseline model, which can be carefully tailored to represent the relevant target population. This is what is recommended in standard texts (Hunink et al., 2001, Drummond et al., 1997, National Institute for Health and Clinical Excellence, 2012).

One might add that in many cases the entire debate about whether a treatment should be adopted hinges on the “threshold” baseline risk at which the treatment – whose relative effect is assumed to remain constant in ratio terms – becomes cost effective, or at which the clinical benefits outweigh the side‐effects. Statin treatment is perhaps the prime example (Stone et al., 2013, National Clinical Guideline Centre, 2014). It is not clear how an arm‐based approach could be modified or extended to produce the kind of analysis that is needed to support decision makers in this context, or indeed in any context.

  1. The empirical question
We are not sure whether proponents of “AB” models would agree, but to some extent the choice between arm‐ and contrast based models could be regarded as an empirical question: do we find in practice that relative effects, given that an appropriate scale has been chosen, are more stable than absolute effects? This is the kind of question that has been studied previously using L'Abbé plots (L'Abbe et al., 1987, Song, 1999).

In the example presented by Hong et al. (2015a), the between study variances in “AB” models were considerably greater than in the “CB” models. It is important to note that the posterior precision of the mean relative effects d1k was quite severely degraded as a result, as can be seen from the wide credible intervals of AB models in Fig. 6. Of course, poor posterior precision is not necessarily inappropriate posterior precision, but the effect of the arm‐based model is to allow the huge variation in absolute effects to affect the estimates of relative effects, while contrast‐based pooling is precisely designed to insulate them from this. Global Goodness of fit and Deviance Information Criteria do not provide a good basis for model choice as different patterns of shrinkage can be expected. The choice should be made on principled and empirical grounds, with a view to minimising bias, not maximising goodness of fit or parsimony.

1.2. Missing‐ness of treatment arms in NMA

The key assumption in all the CB models is the exchangeability of the trial‐specific treatment effects δi,XY across the entire ensemble of trials. Specifically this means that the δi,XY in XY trials are exchangeable with δi,XY in XZ and YZ trials and indeed in all the trials, including those that have neither X nor Y arms. In previous papers (Caldwell et al., 2005, Dias et al., 2011) we have said that this was equivalent to the arms being “missing at random” (MAR). By this we meant that missing‐ness was unrelated to any factors that influenced the relative treatment effects δi,XY. Hong et al. (2015a) claim that CB models require “missing completely at random” (MCAR) and will not work with MAR, because MAR would rule out missing‐ness being related to absolute efficacy. Of course, if “Classic CB” models for NMA were vulnerable to missing‐ness that was related to absolute effects, this would be a most serious criticism because this type of missing‐ness is surely a very common occurrence. But “Classic CB” models are quite obviously not vulnerable to this kind of missing‐ness, because it does not imply that missing‐ness depends on relative effects. The only requirement for CB models is that missing‐ness is unrelated to the relative effects, which is related to the requirement that the trial‐specific relative effects are exchangeable (Lu and Ades, 2009).

Hong et al. (2015a) correctly note that CB models are vulnerable to relationships between relative treatment effects and baseline severity. In a CB framework this – as with any relative effect modification – is typically handled by a meta‐regression of the trial‐specific relative effects against absolute response to the reference treatment, but without putting a model on the reference treatment (Dias et al., 2013b). What is less clear is how AB modellers would cope with this pattern in the data. Covariates can readily be added to AB models, but there appears to be no way of adding a covariate to a relative effect, or putting different covariates on relative and absolute effects.

2. Simulation studies

The simulation studies reported in Hong et al. (2015a) relate to an extension of AB models to multiple outcomes, and they compare certain variants of AB, and “CB + baseline” models, with “Classic CB” (labelled “LAREhom” in Hong et al. (2015a), for Lu & Ades Random Effects homogeneous variance). The purpose is to compare the three models in three‐treatment networks under different forms of missing‐ness: MCAR (completely at random), MAR, and MNAR (not at random), and with different degrees of missing‐ness, approximately 10%, 20%, and 40%. The target outputs of the simulations are (a) the arm means a1, a2, a3 and (b) the mean relative effects d12,d13 (an identity link is used so we ignore the back transformation G1).

Readers should immediately be asking themselves, how can the “Classic CB” model, which does not concern itself at all with absolute effects be made to generate estimates of the absolute effects a1, a2, a3?! It is at this point that we must confess that in a number of publications we have, to our eternal regret, created an estimate of a1 by averaging the absolute event rates on the reference treatment 1. We have done this simply to illustrate the process by which absolute effects ak can be composed from the relative effects d1k and an estimate of the absolute effect on the reference treatment that does not originate from the NMA itself. In spite of this tactical blunder, it should be clear, and we believe it is clear from our specification of the “Classic CB” models in previous publications (Dias et al., 2013c, Lu and Ades, 2004), that these estimates of a1 have nothing to do with the “Classic CB” relative effects model and they have the effect of allowing us to estimate absolute effects without taking the risk of introducing bias into the relative effects. We are therefore quite surprised that Hong et al. (2015a) claim that their simulations show that “Classic CB” “fails to estimate the ak correctly”. The truth is that “Classic CB” has nothing whatsoever to say about absolute effects at all, and this is perfectly obvious from our publications.

Needless to say, “AB” and “CB + baseline” models both perform well in terms of bias, MSE, and coverage, on ak in simulations, but, as we explain above, the estimates of ak obtained by these models are unlikely to be of practical value, unless they happen to best represent the target population. Moreover, it is highly likely that they will be accompanied by biased estimates of the relative effects d1k. In the case of “AB” models the estimates of ak are, in our view, completely arbitrary and unusable. They have an interpretation that makes about as much sense as the average distance between carrots and Cadillacs.

When it comes to the d1k, the simulation studies show that “Classic CB” performs as well as “CB + baseline” model in MCAR scenarios, and better than “CB + baseline” model for “MAR” and MNAR scenarios. (We put the MAR in quotes as they are generated from AB models, and we are uncertain as to whether they would be MAR with respect to relative effects). The main problems for “Classic CB” are the “MAR” scenarios. It is important to appreciate that all the missing‐ness scenarios have the missing‐ness of treatments depending on a function of the simulated absolute effects on one or both of the outcomes. For example in “MAR” the presence of treatment 1 depends on the sum over the two outcomes of the absolute effects on treatment 2, while the presence of treatment 3 depends on the difference between outcomes on treatment 2. It is not immediately clear what this implies about missing‐ness with respect to relative effects. It is therefore difficult to interpret the results, in which “Classic CB” performs worse as the degree of missing‐ness increases, because the data are generated with correlations while the “Classic CB” models used to analyse each simulated dataset assume zero correlations. We note that “Classic CB” models for multiple outcomes which incorporate correlations have been proposed (Efthimiou et al., 2014, Bujkiewicz et al., 2014, Bujkiewicz et al., 2013, Schmid et al., 2014) and they would perhaps provide a fairer comparison.

Regarding the MNAR scenarios, Hong et al. (2015a) note that “Classic CB” produces “oddly unbiased” estimates of the d1k; “oddly” because both a1 and ak are biased, but “to the same degree”. This is, in fact, exactly what would be expected. A great many MNAR scenarios were tested, in which missing‐ness was an extremely complex function of simulated absolute effects on different treatments and outcomes, but none clearly implying a failure of MAR on relative effects. We take these results as a vindication of the arguments set out above to the effect that placing a model on baseline effects makes relative effects vulnerable to misspecification of the baseline model.

3. Conclusions

We believe that AB models are thoroughly misguided, and a huge step back from the separation of absolute and relative effects that has been at the core of modern epidemiology and biostatistics for so long. The use of AB models would risk seriously biased estimates of relative treatment effects, with over‐inflated posterior variance, and block the ability to “transport” relative effects into new absolute effect scenarios, a basic principle in health technology assessment.

Absolute effects should be estimated independently, and the appropriate evidence sources are most likely not to be trials, let alone the same trials which inform the relative effects. The exchangeability assumptions on which “Classic CB” models are based imply MAR with respect to relative effects, and we remain unconvinced by the simulation studies so far presented.

The presentations of AB models in RSM and other journals are superb examples of Bayesian data analysis, and while we disagree profoundly with their assumptions, the authors have succeeded in making us reflect more precisely on the assumptions we are making in evidence synthesis models, and on how well these assumptions are supported by theory or empirical evidence. One particular issue that their papers raise is scale. Some of the important benefits enjoyed by CB modellers depend on an appropriate choice of link function, and it is perhaps not always recognised that choice of which treatment is best can be sensitive to this (Caldwell et al., 2012, van Valkenhoef and Ades, 2013, Norton et al., 2012). Hong et al.'s presentation of “scale‐free” AB models surely challenges our usually blind adoption of log‐odds, log, or even identity links, and it encourages us to pay much more careful attention to the fundamental properties of both frequency counts in different situations, and the scales of measurement used for continuous outcomes.

Acknowledgments

Sofia Dias is partly funded by a Medical Research Council grant (MR/M005232/1).

      Number of times cited according to CrossRef: 41

      • Bayesian network meta-analysis of multiple outcomes in dental research, Journal of Evidence Based Dental Practice, 10.1016/j.jebdp.2020.101403, (101403), (2020).
      • Revisiting and expanding the meta‐analysis of variation: The log coefficient of variation ratio, Research Synthesis Methods, 10.1002/jrsm.1423, 11, 4, (553-567), (2020).
      • Multivariate network meta‐analysis of survival function parameters, Research Synthesis Methods, 10.1002/jrsm.1405, 11, 3, (443-456), (2020).
      • Application of network meta-analysis in the field of physical activity and health promotion, Journal of Sport and Health Science, 10.1016/j.jshs.2020.07.011, (2020).
      • Illustrating the importance of meta‐analysing variances alongside means in ecology and evolution, Journal of Evolutionary Biology, 10.1111/jeb.13661, 33, 9, (1216-1223), (2020).
      • The impact of covariance priors on arm‐based Bayesian network meta‐analyses with binary outcomes, Statistics in Medicine, 10.1002/sim.8580, 39, 22, (2883-2900), (2020).
      • Nonproportional Hazards in Network Meta-Analysis: Efficient Strategies for Model Building and Analysis, Value in Health, 10.1016/j.jval.2020.03.010, (2020).
      • Exploiting relationships between outcomes in Bayesian multivariate network meta‐analysis with an application to relapsing‐remitting multiple sclerosis, Statistics in Medicine, 10.1002/sim.8668, 39, 24, (3329-3346), (2020).
      • Comparing direct acting antivirals for hepatitis C using observational data – Why and how?, Pharmacology Research & Perspectives, 10.1002/prp2.650, 8, 5, (2020).
      • A Bayesian approach to discrete multiple outcome network meta-analysis, PLOS ONE, 10.1371/journal.pone.0231876, 15, 4, (e0231876), (2020).
      • A variance shrinkage method improves arm-based Bayesian network meta-analysis, Statistical Methods in Medical Research, 10.1177/0962280220945731, (096228022094573), (2020).
      • Component network meta‐analysis compared to a matching method in a disconnected network: A case study, Biometrical Journal, 10.1002/bimj.201900339, 0, 0, (2020).
      • Degree irregularity and rank probability bias in network meta‐analysis, Research Synthesis Methods, 10.1002/jrsm.1454, 0, 0, (2020).
      • Random‐effects meta‐analysis of few studies involving rare events, Research Synthesis Methods, 10.1002/jrsm.1370, 11, 1, (74-90), (2019).
      • Integrated evaluation of targeted and non‐targeted therapies in a network meta‐analysis, Biometrical Journal, 10.1002/bimj.201800322, 62, 3, (777-789), (2019).
      • A comparison of arm‐based and contrast‐based models for network meta‐analysis, Statistics in Medicine, 10.1002/sim.8360, 38, 27, (5197-5213), (2019).
      • Incorporating single‐arm evidence into a network meta‐analysis using aggregate level matching: Assessing the impact, Statistics in Medicine, 10.1002/sim.8139, 38, 14, (2505-2523), (2019).
      • Modelling time‐course relationships with multiple treatments: Model‐based network meta‐analysis for continuous summary outcomes, Research Synthesis Methods, 10.1002/jrsm.1351, 10, 2, (267-286), (2019).
      • Evaluating network meta‐analysis and inconsistency using arm‐parameterized model in structural equation modeling, Research Synthesis Methods, 10.1002/jrsm.1344, 10, 2, (240-254), (2019).
      • Meta‐analysis of rare events under the assumption of a homogeneous treatment effect, Biometrical Journal, 10.1002/bimj.201800381, 61, 6, (1557-1574), (2019).
      • A unified method for improved inference in random effects meta-analysis, Biostatistics, 10.1093/biostatistics/kxz020, (2019).
      • Bayesian inference for network meta-regression using multivariate random effects with applications to cholesterol lowering drugs, Biostatistics, 10.1093/biostatistics/kxy014, 20, 3, (499-516), (2018).
      • Bayesian hierarchical methods for meta-analysis combining randomized-controlled and single-arm studies, Statistical Methods in Medical Research, 10.1177/0962280218754928, 28, 5, (1293-1310), (2018).
      • Complications Related to the Initial Trocar Insertion of Three Different Techniques: A Systematic Review and Meta-Analysis, Journal of Minimally Invasive Gynecology, 10.1016/j.jmig.2018.06.023, (2018).
      • Differences in the placebo response between trials do not necessarily preclude network meta‐analysis, Acta Psychiatrica Scandinavica, 10.1111/acps.12974, 138, 6, (615-615), (2018).
      • Multivariate network meta‐analysis to mitigate the effects of outcome reporting bias, Statistics in Medicine, 10.1002/sim.7815, 37, 22, (3254-3266), (2018).
      • A comparison of seven random‐effects models for meta‐analyses that estimate the summary odds ratio, Statistics in Medicine, 10.1002/sim.7588, 37, 7, (1059-1085), (2018).
      • Estimating the variance for heterogeneity in arm‐based network meta‐analysis, Pharmaceutical Statistics, 10.1002/pst.1857, 17, 3, (264-277), (2018).
      • A design‐by‐treatment interaction model for network meta‐analysis and meta‐regression with integrated nested Laplace approximations, Research Synthesis Methods, 10.1002/jrsm.1285, 9, 2, (179-194), (2018).
      • References, Network Meta‐Analysis for Decision Making, 10.1002/9781118951651, (409-445), (2018).
      • Network Meta-Analysis of Diagnostic Test Accuracy Studies, Diagnostic Meta-Analysis, 10.1007/978-3-319-78966-8, (183-197), (2018).
      • Mapping the characteristics of network meta-analyses on drug therapy: A systematic review, PLOS ONE, 10.1371/journal.pone.0196644, 13, 4, (e0196644), (2018).
      • Bartlett‐type corrections and bootstrap adjustments of likelihood‐based inference methods for network meta‐analysis, Statistics in Medicine, 10.1002/sim.7578, 37, 7, (1178-1190), (2017).
      • A review of methods for comparing treatments evaluated in studies that form disconnected networks of evidence, Research Synthesis Methods, 10.1002/jrsm.1278, 9, 2, (148-162), (2017).
      • A Bayesian hierarchical model for network meta-analysis of multiple diagnostic tests, Biostatistics, 10.1093/biostatistics/kxx025, 19, 1, (87-102), (2017).
      • Network meta‐analysis of disconnected networks: How dangerous are random baseline treatment effects?, Research Synthesis Methods, 10.1002/jrsm.1256, 8, 4, (465-474), (2017).
      • Paule‐Mandel estimators for network meta‐analysis with random inconsistency effects, Research Synthesis Methods, 10.1002/jrsm.1244, 8, 4, (416-434), (2017).
      • An investigation of the impact of using different methods for network meta-analysis: a protocol for an empirical evaluation, Systematic Reviews, 10.1186/s13643-017-0511-x, 6, 1, (2017).
      • Disconnected by design: analytic approach in treatment networks having no common comparator, Research Synthesis Methods, 10.1002/jrsm.1204, 7, 4, (420-432), (2016).
      • Meta-analysis of variance: an illustration comparing the effects of two dietary interventions on variability in weight, Evolution, Medicine, and Public Health, 10.1093/emph/eow020, 2016, 1, (244-255), (2016).
      • Two new methods to fit models for network meta-analysis with random inconsistency effects, BMC Medical Research Methodology, 10.1186/s12874-016-0184-5, 16, 1, (2016).

      The full text of this article hosted at iucr.org is unavailable due to technical difficulties.