Absolute or relative effects? Arm‐based synthesis of trial data
We congratulate Hwanhee Hong and colleagues on another fascinating paper (Hong et al., 2015a) arguing the case for arm‐based models for meta‐analysis.
The standard approach to meta‐analysis is the contrast‐based model where the information that is pooled over trials is the information of the trial‐specific relative treatment effect, expressed for example as a log relative risk, log odds ratio, or as a mean treatment difference. In an arm‐based model, it is the absolute log risk, log odds, or mean outcome on each arm that are pooled.
There is no doubt that arm‐based models are an intriguing alternative to the accepted understanding of meta‐analysis, and that they provide a very elegant alternative approach to network meta‐analysis (NMA). However, readers of Research Synthesis Methods (RSM) will have no difficulty in recognising that arm‐based models represent a radical – even revolutionary – departure from current meta‐analytic practice.
In this commentary, we begin by outlining the key differences between arm‐ and contrast‐based meta‐analysis to make clear the true extent of the implications of the claims being made. We then argue that contrast based models are to be preferred on both theoretical and practical grounds.
Hong and colleagues present a number of arguments about the core assumptions of contrast‐based models, both in their RSM paper (Hong et al., 2015a) and previously (Zhang et al., 2014, Ohlssen et al., 2014, Zhang et al., 2015, Hong et al., 2015b), which we believe are mistaken. We offer some counter‐arguments and comment on the simulation study presented by Hong et al. (2015a).
1. Classic contrast‐based, contrast‐based plus baseline and arm‐based models
Hong et al. (2015a) contrast three kinds of models. We will call these (a) the “Classic Contrast‐Based” (Classic CB) model, which corresponds to the standard model for pair‐wise meta‐analysis (Higgins and Green, 2008), extended to NMA (Higgins and Green, 2008, Dias et al., 2013a, Lu and Ades, 2004, Cooper et al., 2009, Lu and Ades, 2006, Caldwell et al., 2005), (b) “Contrast‐Based plus Baseline” (CB + Baseline), and (c) “Arm Based” (AB). It is important to be clear about notation and terminology, so we start with pair‐wise meta‐analysis and then turn to NMA.
(1)
(2)This form of contrast based model has always come in two flavours (Prevost et al., 2000). The one that corresponds exactly to the “Classic CB” model 1 treats the μi,X as unrelated nuisance parameters. This is the approach adopted in the NMA models developed in Bristol (Dias et al., 2013a, Lu and Ades, 2004, Prevost et al., 2000).
The alternative is to put another hierarchical model on the trial‐specific control‐arms:
(Achana et al., 2013, Dias et al., 2013c). This is the “CB + baseline model”, or alternatively a bivariate normal model on the μi,X and the δi,XY could be adopted (van Houwelingen et al., 1993).
, and its effect relative to the reference treatment 1,
, which may or may not be present in trial i. With this extended notation, the “Classic CB” model makes the trial‐specific control arms arbitrary and unrelated.
(3)Note that often, subscript t1 is dropped from
, as it is redundant. Clearly, there is no way of estimating an absolute effect for any treatment from the “Classic CB” model.
We note in passing that in another paper Hong et al. (2015b) claim that the assumption of exchangeability of “effects relative to an arbitrary and nonconstant baseline remains much to assume”. This misses the point that exchangeability of the effects δi,1k relative to treatment 1 implies exchangeability of all the δi,jk (Lu and Ades, 2009).
(4)Note that in the “CB + baseline” model, information on the baseline will back‐propagate onto the relative effect parameters. It is this property of “CB + Baseline” models that makes the relative effect estimates vulnerable to misspecification of the absolute effects.
(5)The first feature we notice is that there are no relative treatment effects in equation 5. Mean “treatment effects” on the linear predictor scale are manufactured from the treatment‐specific means as dXY = aY − aX. Note that while the CB modeller working on binary data is committed to specify whether effects are additive on a continuous, log, or log‐odds scale, the AB modeller is free to report risk differences, G− 1(aY) − G− 1(aX), relative risks G− 1(aY)/G− 1(aX), or odds ratios (G− 1(aY)G− 1(1 − aX))/(G− 1(aX)G− 1(1 − aY)). For binary outcomes, both probit and logit link functions have been proposed in arm‐based models, presumably giving virtually identical results. This “model‐free” approach to relative effect estimation has the advantage that it makes no assumptions, but, as we see below, it comes with a heavy price in posterior variance of the relative effects.
In the latest paper in RSM Hong et al. (2015a) extend AB modelling to multiple outcomes, and run simulations that compare “Classic CB”, “CB + baseline”, and “AB” models in various styles. In previous work developing AB models Zhang et al. (2014) have criticised the “Classic CB” model on the grounds that, because it contains no model for absolute effects, it cannot report overall event rates without making further assumptions (although some would see this as an advantage of “Classic CB”). Other papers have examined the concept of inconsistency in NMA (Zhang et al., 2015) and looked at the role of AB models in drug development (Ohlssen et al., 2014).
1.1. Comparing arm‐based and contrast based models
- Fundamentals of meta‐analysis
We also believe that arm‐based pooling effectively breaks randomisation, and in fact runs against the entire way in which randomised controlled trials are designed, analysed, and used.
While “AB” models take an extreme position, many of us have also argued against “CB + baseline” models, because, unless the baseline model is correctly specified, the relative effect estimates will be biased. For example, one often finds that outcomes on the control arm have improved over time, because less severe populations are entered into more recent trials while the relative effect has stayed constant. If one random effects model is put on the baselines and another on the relative effects, and both are estimated at the same time, we expect the baseline effects in earlier studies to be shrunk upwards, and relative effects under‐estimated, while the relative effects in more recent studies will be over‐estimated. Senn (2010) has argued forcefully against “CB + baseline” models for these reasons. “AB” models will be even more likely to produce biased treatment effects, if one accepts the commonly made assumption that it is the relative effects which are exchangeable across trials.
- Evidence synthesis for decision‐making
Where we differ from proponents of AB models is that we believe that, while randomised controlled trials are unquestionably the best data sources to inform relative effects, the data sources that best inform the absolute effects might be cohort studies, a carefully selected subset of the trials included in the meta‐analysis, or expert opinion. Manufacturers making submissions on new products to the National Institute for Health and Care Excellence (NICE) in the UK are asked to explain the basis on which they have chosen the evidence for absolute effects, and this is quite distinct from trial evidence from which relative effects are sourced (National Institute for Health and Clinical Excellence, 2012).
The disadvantage of an arm based model is that it obliges the modeller to use exactly the same data sources for both absolute and relative effects, whereas the natural way to carry out health technology assessments of alternative treatments is to apply a “transportable” relative effect estimate to an independently estimated baseline model, which can be carefully tailored to represent the relevant target population. This is what is recommended in standard texts (Hunink et al., 2001, Drummond et al., 1997, National Institute for Health and Clinical Excellence, 2012).
One might add that in many cases the entire debate about whether a treatment should be adopted hinges on the “threshold” baseline risk at which the treatment – whose relative effect is assumed to remain constant in ratio terms – becomes cost effective, or at which the clinical benefits outweigh the side‐effects. Statin treatment is perhaps the prime example (Stone et al., 2013, National Clinical Guideline Centre, 2014). It is not clear how an arm‐based approach could be modified or extended to produce the kind of analysis that is needed to support decision makers in this context, or indeed in any context.
- The empirical question
In the example presented by Hong et al. (2015a), the between study variances in “AB” models were considerably greater than in the “CB” models. It is important to note that the posterior precision of the mean relative effects d1k was quite severely degraded as a result, as can be seen from the wide credible intervals of AB models in Fig. 6. Of course, poor posterior precision is not necessarily inappropriate posterior precision, but the effect of the arm‐based model is to allow the huge variation in absolute effects to affect the estimates of relative effects, while contrast‐based pooling is precisely designed to insulate them from this. Global Goodness of fit and Deviance Information Criteria do not provide a good basis for model choice as different patterns of shrinkage can be expected. The choice should be made on principled and empirical grounds, with a view to minimising bias, not maximising goodness of fit or parsimony.
1.2. Missing‐ness of treatment arms in NMA
The key assumption in all the CB models is the exchangeability of the trial‐specific treatment effects δi,XY across the entire ensemble of trials. Specifically this means that the δi,XY in XY trials are exchangeable with δi,XY in XZ and YZ trials and indeed in all the trials, including those that have neither X nor Y arms. In previous papers (Caldwell et al., 2005, Dias et al., 2011) we have said that this was equivalent to the arms being “missing at random” (MAR). By this we meant that missing‐ness was unrelated to any factors that influenced the relative treatment effects δi,XY. Hong et al. (2015a) claim that CB models require “missing completely at random” (MCAR) and will not work with MAR, because MAR would rule out missing‐ness being related to absolute efficacy. Of course, if “Classic CB” models for NMA were vulnerable to missing‐ness that was related to absolute effects, this would be a most serious criticism because this type of missing‐ness is surely a very common occurrence. But “Classic CB” models are quite obviously not vulnerable to this kind of missing‐ness, because it does not imply that missing‐ness depends on relative effects. The only requirement for CB models is that missing‐ness is unrelated to the relative effects, which is related to the requirement that the trial‐specific relative effects are exchangeable (Lu and Ades, 2009).
Hong et al. (2015a) correctly note that CB models are vulnerable to relationships between relative treatment effects and baseline severity. In a CB framework this – as with any relative effect modification – is typically handled by a meta‐regression of the trial‐specific relative effects against absolute response to the reference treatment, but without putting a model on the reference treatment (Dias et al., 2013b). What is less clear is how AB modellers would cope with this pattern in the data. Covariates can readily be added to AB models, but there appears to be no way of adding a covariate to a relative effect, or putting different covariates on relative and absolute effects.
2. Simulation studies
The simulation studies reported in Hong et al. (2015a) relate to an extension of AB models to multiple outcomes, and they compare certain variants of AB, and “CB + baseline” models, with “Classic CB” (labelled “LAREhom” in Hong et al. (2015a), for Lu & Ades Random Effects homogeneous variance). The purpose is to compare the three models in three‐treatment networks under different forms of missing‐ness: MCAR (completely at random), MAR, and MNAR (not at random), and with different degrees of missing‐ness, approximately 10%, 20%, and 40%. The target outputs of the simulations are (a) the arm means a1, a2, a3 and (b) the mean relative effects d12,d13 (an identity link is used so we ignore the back transformation G−1).
Readers should immediately be asking themselves, how can the “Classic CB” model, which does not concern itself at all with absolute effects be made to generate estimates of the absolute effects a1, a2, a3?! It is at this point that we must confess that in a number of publications we have, to our eternal regret, created an estimate of a1 by averaging the absolute event rates on the reference treatment 1. We have done this simply to illustrate the process by which absolute effects ak can be composed from the relative effects d1k and an estimate of the absolute effect on the reference treatment that does not originate from the NMA itself. In spite of this tactical blunder, it should be clear, and we believe it is clear from our specification of the “Classic CB” models in previous publications (Dias et al., 2013c, Lu and Ades, 2004), that these estimates of a1 have nothing to do with the “Classic CB” relative effects model and they have the effect of allowing us to estimate absolute effects without taking the risk of introducing bias into the relative effects. We are therefore quite surprised that Hong et al. (2015a) claim that their simulations show that “Classic CB” “fails to estimate the ak correctly”. The truth is that “Classic CB” has nothing whatsoever to say about absolute effects at all, and this is perfectly obvious from our publications.
Needless to say, “AB” and “CB + baseline” models both perform well in terms of bias, MSE, and coverage, on ak in simulations, but, as we explain above, the estimates of ak obtained by these models are unlikely to be of practical value, unless they happen to best represent the target population. Moreover, it is highly likely that they will be accompanied by biased estimates of the relative effects d1k. In the case of “AB” models the estimates of ak are, in our view, completely arbitrary and unusable. They have an interpretation that makes about as much sense as the average distance between carrots and Cadillacs.
When it comes to the d1k, the simulation studies show that “Classic CB” performs as well as “CB + baseline” model in MCAR scenarios, and better than “CB + baseline” model for “MAR” and MNAR scenarios. (We put the MAR in quotes as they are generated from AB models, and we are uncertain as to whether they would be MAR with respect to relative effects). The main problems for “Classic CB” are the “MAR” scenarios. It is important to appreciate that all the missing‐ness scenarios have the missing‐ness of treatments depending on a function of the simulated absolute effects on one or both of the outcomes. For example in “MAR” the presence of treatment 1 depends on the sum over the two outcomes of the absolute effects on treatment 2, while the presence of treatment 3 depends on the difference between outcomes on treatment 2. It is not immediately clear what this implies about missing‐ness with respect to relative effects. It is therefore difficult to interpret the results, in which “Classic CB” performs worse as the degree of missing‐ness increases, because the data are generated with correlations while the “Classic CB” models used to analyse each simulated dataset assume zero correlations. We note that “Classic CB” models for multiple outcomes which incorporate correlations have been proposed (Efthimiou et al., 2014, Bujkiewicz et al., 2014, Bujkiewicz et al., 2013, Schmid et al., 2014) and they would perhaps provide a fairer comparison.
Regarding the MNAR scenarios, Hong et al. (2015a) note that “Classic CB” produces “oddly unbiased” estimates of the d1k; “oddly” because both a1 and ak are biased, but “to the same degree”. This is, in fact, exactly what would be expected. A great many MNAR scenarios were tested, in which missing‐ness was an extremely complex function of simulated absolute effects on different treatments and outcomes, but none clearly implying a failure of MAR on relative effects. We take these results as a vindication of the arguments set out above to the effect that placing a model on baseline effects makes relative effects vulnerable to misspecification of the baseline model.
3. Conclusions
We believe that AB models are thoroughly misguided, and a huge step back from the separation of absolute and relative effects that has been at the core of modern epidemiology and biostatistics for so long. The use of AB models would risk seriously biased estimates of relative treatment effects, with over‐inflated posterior variance, and block the ability to “transport” relative effects into new absolute effect scenarios, a basic principle in health technology assessment.
Absolute effects should be estimated independently, and the appropriate evidence sources are most likely not to be trials, let alone the same trials which inform the relative effects. The exchangeability assumptions on which “Classic CB” models are based imply MAR with respect to relative effects, and we remain unconvinced by the simulation studies so far presented.
The presentations of AB models in RSM and other journals are superb examples of Bayesian data analysis, and while we disagree profoundly with their assumptions, the authors have succeeded in making us reflect more precisely on the assumptions we are making in evidence synthesis models, and on how well these assumptions are supported by theory or empirical evidence. One particular issue that their papers raise is scale. Some of the important benefits enjoyed by CB modellers depend on an appropriate choice of link function, and it is perhaps not always recognised that choice of which treatment is best can be sensitive to this (Caldwell et al., 2012, van Valkenhoef and Ades, 2013, Norton et al., 2012). Hong et al.'s presentation of “scale‐free” AB models surely challenges our usually blind adoption of log‐odds, log, or even identity links, and it encourages us to pay much more careful attention to the fundamental properties of both frequency counts in different situations, and the scales of measurement used for continuous outcomes.
Acknowledgments
Sofia Dias is partly funded by a Medical Research Council grant (MR/M005232/1).
References
Citing Literature
Number of times cited according to CrossRef: 41
- Menglu Liang, Qinshu Lian, Georgios A. Kotsakis, Bryan S. Michalowicz, Mike T. John, Haitao Chu, Bayesian network meta-analysis of multiple outcomes in dental research, Journal of Evidence Based Dental Practice, 10.1016/j.jebdp.2020.101403, (101403), (2020).
- Alistair M. Senior, Wolfgang Viechtbauer, Shinichi Nakagawa, Revisiting and expanding the meta‐analysis of variation: The log coefficient of variation ratio, Research Synthesis Methods, 10.1002/jrsm.1423, 11, 4, (553-567), (2020).
- Shannon Cope, Keith Chan, Jeroen P. Jansen, Multivariate network meta‐analysis of survival function parameters, Research Synthesis Methods, 10.1002/jrsm.1405, 11, 3, (443-456), (2020).
- Xiwen Su, Daniel J. McDonough, Haitao Chu, Minghui Quan, Zan Gao, Application of network meta-analysis in the field of physical activity and health promotion, Journal of Sport and Health Science, 10.1016/j.jshs.2020.07.011, (2020).
- Alfredo Sánchez‐Tójar, Nicholas P. Moran, Rose E. O’Dea, Klaus Reinhold, Shinichi Nakagawa, Illustrating the importance of meta‐analysing variances alongside means in ecology and evolution, Journal of Evolutionary Biology, 10.1111/jeb.13661, 33, 9, (1216-1223), (2020).
- Zhenxun Wang, Lifeng Lin, James S. Hodges, Haitao Chu, The impact of covariance priors on arm‐based Bayesian network meta‐analyses with binary outcomes, Statistics in Medicine, 10.1002/sim.8580, 39, 22, (2883-2900), (2020).
- Anna Wiksten, Neil Hawkins, Hans-Peter Piepho, Sandro Gsteiger, Nonproportional Hazards in Network Meta-Analysis: Efficient Strategies for Model Building and Analysis, Value in Health, 10.1016/j.jval.2020.03.010, (2020).
- Ed Waddingham, Paul M. Matthews, Deborah Ashby, Exploiting relationships between outcomes in Bayesian multivariate network meta‐analysis with an application to relapsing‐remitting multiple sclerosis, Statistics in Medicine, 10.1002/sim.8668, 39, 24, (3329-3346), (2020).
- Jim Young, Stanley Wong, Naveed Z. Janjua, Marina B. Klein, Comparing direct acting antivirals for hepatitis C using observational data – Why and how?, Pharmacology Research & Perspectives, 10.1002/prp2.650, 8, 5, (2020).
- Rebecca Graziani, Sergio Venturini, A Bayesian approach to discrete multiple outcome network meta-analysis, PLOS ONE, 10.1371/journal.pone.0231876, 15, 4, (e0231876), (2020).
- Zhenxun Wang, Lifeng Lin, James S Hodges, Richard MacLehose, Haitao Chu, A variance shrinkage method improves arm-based Bayesian network meta-analysis, Statistical Methods in Medical Research, 10.1177/0962280220945731, (096228022094573), (2020).
- Gerta Rücker, Susanne Schmitz, Guido Schwarzer, Component network meta‐analysis compared to a matching method in a disconnected network: A case study, Biometrical Journal, 10.1002/bimj.201900339, 0, 0, (2020).
- Annabel L. Davies, Tobias Galla, Degree irregularity and rank probability bias in network meta‐analysis, Research Synthesis Methods, 10.1002/jrsm.1454, 0, 0, (2020).
- Burak Kürsad Günhan, Christian Röver, Tim Friede, Random‐effects meta‐analysis of few studies involving rare events, Research Synthesis Methods, 10.1002/jrsm.1370, 11, 1, (74-90), (2019).
- Tanja Proctor, Katrin Jensen, Meinhard Kieser, Integrated evaluation of targeted and non‐targeted therapies in a network meta‐analysis, Biometrical Journal, 10.1002/bimj.201800322, 62, 3, (777-789), (2019).
- Ian R. White, Rebecca M. Turner, Amalia Karahalios, Georgia Salanti, A comparison of arm‐based and contrast‐based models for network meta‐analysis, Statistics in Medicine, 10.1002/sim.8360, 38, 27, (5197-5213), (2019).
- Joy Leahy, Howard Thom, Jeroen P. Jansen, Emma Gray, Aisling O'Leary, Arthur White, Cathal Walsh, Incorporating single‐arm evidence into a network meta‐analysis using aggregate level matching: Assessing the impact, Statistics in Medicine, 10.1002/sim.8139, 38, 14, (2505-2523), (2019).
- Hugo Pedder, Sofia Dias, Margherita Bennetts, Martin Boucher, Nicky J. Welton, Modelling time‐course relationships with multiple treatments: Model‐based network meta‐analysis for continuous summary outcomes, Research Synthesis Methods, 10.1002/jrsm.1351, 10, 2, (267-286), (2019).
- Ming‐Chieh Shih, Yu‐Kang Tu, Evaluating network meta‐analysis and inconsistency using arm‐parameterized model in structural equation modeling, Research Synthesis Methods, 10.1002/jrsm.1344, 10, 2, (240-254), (2019).
- Romain Piaget‐Rossel, Patrick Taffé, Meta‐analysis of rare events under the assumption of a homogeneous treatment effect, Biometrical Journal, 10.1002/bimj.201800381, 61, 6, (1557-1574), (2019).
- Shonosuke Sugasawa, Hisashi Noma, A unified method for improved inference in random effects meta-analysis, Biostatistics, 10.1093/biostatistics/kxz020, (2019).
- Hao Li, Ming-Hui Chen, Joseph G Ibrahim, Sungduk Kim, Arvind K Shah, Jianxin Lin, Andrew M Tershakovec, Bayesian inference for network meta-regression using multivariate random effects with applications to cholesterol lowering drugs, Biostatistics, 10.1093/biostatistics/kxy014, 20, 3, (499-516), (2018).
- Jing Zhang, Chia-Wen Ko, Lei Nie, Yong Chen, Ram Tiwari, Bayesian hierarchical methods for meta-analysis combining randomized-controlled and single-arm studies, Statistical Methods in Medical Research, 10.1177/0962280218754928, 28, 5, (1293-1310), (2018).
- Mai Nishimura, Sachiko Matsumoto, Yasuhiro Ohara, Kaoru Minowa, Risa Tsunematsu, Kanako Takimoto, Kazuaki Imai, Yoko Tsuzuki, Hajime Ota, Ayako Nakajima, Yoshiyuki Fukushi, Shinichiro Wada, Takafumi Fujino, Yoichi M. Ito, Complications Related to the Initial Trocar Insertion of Three Different Techniques: A Systematic Review and Meta-Analysis, Journal of Minimally Invasive Gynecology, 10.1016/j.jmig.2018.06.023, (2018).
- G. Rücker, G. Schwarzer, Differences in the placebo response between trials do not necessarily preclude network meta‐analysis, Acta Psychiatrica Scandinavica, 10.1111/acps.12974, 138, 6, (615-615), (2018).
- Hyunsoo Hwang, Stacia M. DeSantis, Multivariate network meta‐analysis to mitigate the effects of outcome reporting bias, Statistics in Medicine, 10.1002/sim.7815, 37, 22, (3254-3266), (2018).
- Dan Jackson, Martin Law, Theo Stijnen, Wolfgang Viechtbauer, Ian R. White, A comparison of seven random‐effects models for meta‐analyses that estimate the summary odds ratio, Statistics in Medicine, 10.1002/sim.7588, 37, 7, (1059-1085), (2018).
- Hans‐Peter Piepho, Laurence V. Madden, James Roger, Roger Payne, Emlyn R. Williams, Estimating the variance for heterogeneity in arm‐based network meta‐analysis, Pharmaceutical Statistics, 10.1002/pst.1857, 17, 3, (264-277), (2018).
- Burak Kürsad Günhan, Tim Friede, Leonhard Held, A design‐by‐treatment interaction model for network meta‐analysis and meta‐regression with integrated nested Laplace approximations, Research Synthesis Methods, 10.1002/jrsm.1285, 9, 2, (179-194), (2018).
- Sofia Dias, A. E. Ades, Nicky J. Welton, Jeroen P. Jansen, Alexander J. Sutton, References, Network Meta‐Analysis for Decision Making, 10.1002/9781118951651, (409-445), (2018).
- Gerta Rücker, Network Meta-Analysis of Diagnostic Test Accuracy Studies, Diagnostic Meta-Analysis, 10.1007/978-3-319-78966-8, (183-197), (2018).
- Fernanda S. Tonin, Laiza M. Steimbach, Antonio M. Mendes, Helena H. Borba, Roberto Pontarolo, Fernando Fernandez-Llimos, Mapping the characteristics of network meta-analyses on drug therapy: A systematic review, PLOS ONE, 10.1371/journal.pone.0196644, 13, 4, (e0196644), (2018).
- Hisashi Noma, Kengo Nagashima, Kazushi Maruo, Masahiko Gosho, Toshi A. Furukawa, Bartlett‐type corrections and bootstrap adjustments of likelihood‐based inference methods for network meta‐analysis, Statistics in Medicine, 10.1002/sim.7578, 37, 7, (1178-1190), (2017).
- John W. Stevens, Christine Fletcher, Gerald Downey, Anthea Sutton, A review of methods for comparing treatments evaluated in studies that form disconnected networks of evidence, Research Synthesis Methods, 10.1002/jrsm.1278, 9, 2, (148-162), (2017).
- Xiaoye Ma, Qinshu Lian, Haitao Chu, Joseph G Ibrahim, Yong Chen, A Bayesian hierarchical model for network meta-analysis of multiple diagnostic tests, Biostatistics, 10.1093/biostatistics/kxx025, 19, 1, (87-102), (2017).
- Audrey Béliveau, Sarah Goring, Robert W. Platt, Paul Gustafson, Network meta‐analysis of disconnected networks: How dangerous are random baseline treatment effects?, Research Synthesis Methods, 10.1002/jrsm.1256, 8, 4, (465-474), (2017).
- Dan Jackson, Areti Angeliki Veroniki, Martin Law, Andrea C. Tricco, Rose Baker, Paule‐Mandel estimators for network meta‐analysis with random inconsistency effects, Research Synthesis Methods, 10.1002/jrsm.1244, 8, 4, (416-434), (2017).
- Amalia Karahalios, Georgia Salanti, Simon L. Turner, G. Peter Herbison, Ian R. White, Areti Angeliki Veroniki, Adriani Nikolakopoulou, Joanne E. Mckenzie, An investigation of the impact of using different methods for network meta-analysis: a protocol for an empirical evaluation, Systematic Reviews, 10.1186/s13643-017-0511-x, 6, 1, (2017).
- S. M. Goring, P. Gustafson, Y. Liu, S. Saab, S. K. Cline, R. W. Platt, Disconnected by design: analytic approach in treatment networks having no common comparator, Research Synthesis Methods, 10.1002/jrsm.1204, 7, 4, (420-432), (2016).
- Alistair M. Senior, Alison K. Gosby, Jing Lu, Stephen J. Simpson, David Raubenheimer, Meta-analysis of variance: an illustration comparing the effects of two dietary interventions on variability in weight, Evolution, Medicine, and Public Health, 10.1093/emph/eow020, 2016, 1, (244-255), (2016).
- Martin Law, Dan Jackson, Rebecca Turner, Kirsty Rhodes, Wolfgang Viechtbauer, Two new methods to fit models for network meta-analysis with random inconsistency effects, BMC Medical Research Methodology, 10.1186/s12874-016-0184-5, 16, 1, (2016).




