When the efficacy of a new medical drug is compared against that of an established competitor in a randomized controlled trial, the difference in patient-relevant outcomes, such as mortality, is usually measured directly. In diagnostic research, however, the impact of diagnostic procedures is of an indirect nature as test results do influence downstream clinical decisions, but test performance (as characterized by sensitivity, specificity, and the predictive values of a procedure) is, at best, only a surrogate endpoint for patient outcome and does not necessarily translate into it. Not many randomized controlled trials have been conducted so far in diagnostic research, and, hence, we need alternative approaches to close the gap between test characteristics and patient outcomes. Several informal approaches have been suggested in order to close this gap, and decision modeling has been advocated as a means of obtaining formal approaches. Recently, the expected benefit has been proposed as a quantity that allows a simple formal approach, and we take up this suggestion in this paper. We regard the expected benefit as an estimation problem and consider two approaches to statistical inference. Moreover, using data from a previously published study, we illustrate the possible insights to be gained from the application of formal inference techniques to determine the expected benefit.

]]>In observational studies with dichotomous outcome of a population, researchers usually report treatment effect alone, although both baseline risk and treatment effect are needed to evaluate the significance of the treatment effect to the population. In this article, we study point and interval estimates including confidence region of baseline risk and treatment effect based on logistic model, where baseline risk is the risk of outcome of the population under control treatment while treatment effect is measured by the risk difference between outcomes of the population under active versus control treatments. Using approximate normal distribution of the maximum-likelihood (ML) estimate of the model parameters, we obtain an approximate joint distribution of the ML estimate of the baseline risk and the treatment effect. Using the approximate joint distribution, we obtain point estimate and confidence region of the baseline risk and the treatment effect as well as point estimate and confidence interval of the treatment effect when the ML estimate of the baseline risk falls into specified range. These interval estimates reflect nonnormality of the joint distribution of the ML estimate of the baseline risk and the treatment effect. The method can be easily implemented by using any software that generates normal distribution. The method can also be used to obtain point and interval estimates of baseline risk and any other measure of treatment effect such as risk ratio and the number needed to treat. The method can also be extended from logistic model to other models such as log-linear model.

]]>Covariate measurement error may cause biases in parameters of regression coefficients in generalized linear models. The influence of measurement error on interaction parameters has, however, only rarely been investigated in depth, and if so, attenuation effects were reported. In this paper, we show that also reverse attenuation of interaction effects may emerge, namely when heteroscedastic measurement error or sampling variances of a mismeasured covariate are present, which are not unrealistic scenarios in practice. Theoretical findings are illustrated with simulations. A Bayesian approach employing integrated nested Laplace approximations is suggested to model the heteroscedastic measurement error and covariate variances, and an application shows that the method is able to reveal approximately correct parameter estimates.

]]>We consider modelling the movements of larvae using individual bioassays in which data are collected at a high-frequency rate of five observations per second. The aim is to characterize the behaviour of the larvae when exposed to attractant and repellent compounds. Mixtures of diffusion processes, as well as Hidden Markov models, are proposed as models of larval movement. These models account for directed and localized movements, and successfully distinguish between the behaviour of larvae exposed to attractant and repellent compounds. A simulation study illustrates the advantage of using a Hidden Markov model rather than a simpler mixture model. Practical aspects of model estimation and inference are considered on extensive data collected in a study of novel approaches for the management of cabbage root fly.

]]>Using the conditional likelihood, a model-specific score test can be derived under a given genetic model to test genetic association for case-parents triad family data. When the underlying genetic model is correctly specified, the score test is most powerful. However, it can lose substantial power when the model is misspecified. Several robust tests have been proposed to deal with the problem, such as the maximum test statistic, the maximin efficiency robust test, and the constrained likelihood ratio test. These tests have been shown to be robust against model misspecification compared with those model-specific score tests, but they are either time-consuming in computation or not sufficiently high in power robustness under some situations. In this study, a data-driven procedure is proposed to construct two adaptive robust genetic association tests *W*_{MERT} and *W*_{MAX}. The *W*_{MERT} is simple in calculation and has fairly high power robustness. The empirical power of *W*_{MAX} is quite stable and close to those of the model-specific score tests. The two proposed tests should be beneficial to practical genetic association studies. A real dataset consisting of neural tube defect triad families is used for illustration of the methods. R-scripts are also provided for numerical calculation of the proposed methods in practical studies.

To reduce the lengthy duration of a crossover trial for comparing three treatments, the incomplete block design has been often considered. A sample size calculation procedure for testing nonequality between either of the two experimental treatments and a placebo under such a design is developed. To evaluate the performance of the proposed sample size calculation procedure, Monte Carlo simulation is employed. The accuracy of the sample size calculation procedure developed here is demonstrated in a variety of situations. As compared with the parallel groups design, a substantial proportional reduction in the total minimum required sample size in use of the incomplete block crossover design is found. A crossover trial comparing two different doses of formoterol with a placebo on the forced expiratory volume is applied to illustrate the use of the sample size calculation procedure.

]]>The development of methods for dealing with continuous data with a spike at zero has lagged behind those for overdispersed or zero-inflated count data. We consider longitudinal ecological data corresponding to an annual average of 26 weekly maximum counts of birds, and are hence effectively continuous, bounded below by zero but also with a discrete mass at zero. We develop a Bayesian hierarchical Tweedie regression model that can directly accommodate the excess number of zeros common to this type of data, whilst accounting for both spatial and temporal correlation. Implementation of the model is conducted in a Markov chain Monte Carlo (MCMC) framework, using reversible jump MCMC to explore uncertainty across both parameter and model spaces. This regression modelling framework is very flexible and removes the need to make strong assumptions about mean-variance relationships *a priori*. It can also directly account for the spike at zero, whilst being easily applicable to other types of data and other model formulations. Whilst a correlative study such as this cannot prove causation, our results suggest that an increase in an avian predator may have led to an overall decrease in the number of one of its prey species visiting garden feeding stations in the United Kingdom. This may reflect a change in behaviour of house sparrows to avoid feeding stations frequented by sparrowhawks, or a reduction in house sparrow population size as a result of sparrowhawk increase.

Shape analysis is of great importance in many fields of medical imaging and computational biology. In this paper, we consider the shape space as the set of smooth planar immersed curves in (parameterized curves) and, using the property of being isometric to a classical manifold immersed in a Euclidean space, we introduce a new extrinsic sample mean and a new extrinsic variance for a finite set of shapes, which are not necessarily star shaped. This is a fundamental tool in medical image analysis, for instance, to assess uncertainties that arise in locating anatomical structures such as the prostate and the bladder. We apply it to a dataset consisting of parallel planar axial CT sections of human prostate, in order to study the variability between boundaries that have been manually delineated by several observers.

]]>A surrogate endpoint is intended to replace a clinical endpoint for the evaluation of new treatments when it can be measured more cheaply, more conveniently, more frequently, or earlier than that clinical endpoint. A surrogate endpoint is expected to predict clinical benefit, harm, or lack of these. Besides the biological plausibility of a surrogate, a quantitative assessment of the strength of evidence for surrogacy requires the demonstration of the prognostic value of the surrogate for the clinical outcome, and evidence that treatment effects on the surrogate reliably predict treatment effects on the clinical outcome. We focus on these two conditions, and outline the statistical approaches that have been proposed to assess the extent to which these conditions are fulfilled. When data are available from a single trial, one can assess the “individual level association” between the surrogate and the true endpoint. When data are available from several trials, one can additionally assess the “trial level association” between the treatment effect on the surrogate and the treatment effect on the true endpoint. In the latter case, the “surrogate threshold effect” can be estimated as the minimum effect on the surrogate endpoint that predicts a statistically significant effect on the clinical endpoint. All these concepts are discussed in the context of randomized clinical trials in oncology, and illustrated with two meta-analyses in gastric cancer.

]]>This paper proposes a semiparametric methodology for modeling multivariate and conditional distributions. We first build a multivariate distribution whose dependence structure is induced by a Gaussian copula and whose marginal distributions are estimated nonparametrically via mixtures of B-spline densities. The conditional distribution of a given variable is obtained in closed form from this multivariate distribution. We take a Bayesian approach, using Markov chain Monte Carlo methods for inference. We study the frequentist properties of the proposed methodology via simulation and apply the method to estimation of conditional densities of summary statistics, used for computing conditional local false discovery rates, from genetic association studies of schizophrenia and cardiovascular disease risk factors.

]]>When performing single arm meta-analyses of rare events in small populations, if the outcome of interest is incidence, it is not uncommon to have at least one study with zero events, especially in the presence of competing risks. In this paper, we address the problem of how to include studies with zero events in inverse variance meta-analyses when individual patient data are not available, going beyond the naïve approach of not including the study or the use of a continuity correction. The proposed solution is the arcsine transformation of the crude cumulative incidence as its approximate variance, which is inversely proportional to the sample size, can be calculated also for studies with a zero estimate. As an alternative, generalized linear mixed models (GLMM) can be used. Simulations were performed to compare the results from inverse variance method meta-analyses of the arcsine transformed cumulative incidence to those obtained from meta-analyses of the cumulative incidence itself and of the logit transformation of the cumulative incidence. The comparisons have been carried out for different scenarios of heterogeneity, incidence, and censoring and for competing and not competing risks. The arcsine transformation showed the smallest bias and the highest coverage among models assuming within study normality. At the same time, the GLMM model had the best performance at very low incidences. The proposed method was applied to the clinical context that motivated this work, i.e. a meta-analysis of 5-year crude cumulative incidence of central nervous system recurrences in children treated for acute lymphoblastic leukemia.

]]>Instead of assessing the overall fit of candidate models like the traditional model selection criteria, the focused information criterion focuses attention directly on the parameter of the primary interest and aims to select the model with the minimum estimated mean squared error for the estimate of the focused parameter. In this article we apply the focused information criterion for personalized medicine. By using individual-level information from clinical observations, demographics, and genetics, we obtain the personalized predictive models to make the prognosis and diagnosis individually. The consideration of the heterogeneity among the individuals helps reduce the prediction uncertainty and improve the prediction accuracy. Two real data examples from biomedical research are studied as illustrations.

]]>Multiple imputation can be used as a tool in the process of constructing prediction models in medical and epidemiological studies with missing covariate values. Such models can be used to make predictions for model performance assessment, but the task is made more complicated by the multiple imputation structure. We summarize various predictions constructed from covariates, including multiply imputed covariates, and either the set of imputation-specific prediction model coefficients or the pooled prediction model coefficients. We further describe approaches for using the predictions to assess model performance. We distinguish between *ideal model performance* and *pragmatic model performance*, where the former refers to the model's performance in an ideal clinical setting where all individuals have fully observed predictors and the latter refers to the model's performance in a real-world clinical setting where some individuals have missing predictors. The approaches are compared through an extensive simulation study based on the UK700 trial. We determine that measures of ideal model performance can be estimated within imputed datasets and subsequently pooled to give an overall measure of model performance. Alternative methods to evaluate pragmatic model performance are required and we propose constructing predictions either from a second set of covariate imputations which make no use of observed outcomes, or from a set of partial prediction models constructed for each potential observed pattern of covariate. Pragmatic model performance is generally lower than ideal model performance. We focus on model performance within the derivation data, but describe how to extend all the methods to a validation dataset.

Quantitative decision models such as multiple criteria decision analysis (MCDA) can be used in benefit-risk assessment to formalize trade-offs between benefits and risks, providing transparency to the assessment process. There is however no well-established method for propagating uncertainty of treatment effects data through such models to provide a sense of the variability of the benefit-risk balance. Here, we present a Bayesian statistical method that directly models the outcomes observed in randomized placebo-controlled trials and uses this to infer indirect comparisons between competing active treatments. The resulting treatment effects estimates are suitable for use within the MCDA setting, and it is possible to derive the distribution of the overall benefit-risk balance through Markov Chain Monte Carlo simulation. The method is illustrated using a case study of natalizumab for relapsing-remitting multiple sclerosis.

]]>While benefit-risk assessment is a key component of the drug development and maintenance process, it is often described in a narrative. In contrast, structured benefit-risk assessment builds on established ideas from decision analysis and comprises a qualitative framework and quantitative methodology. We compare two such frameworks, applying multi-criteria decision-analysis (MCDA) within the PrOACT-URL framework and weighted net clinical benefit (wNCB), within the BRAT framework. These are applied to a case study of natalizumab for the treatment of relapsing remitting multiple sclerosis. We focus on the practical considerations of applying these methods and give recommendations for visual presentation of results. In the case study, we found structured benefit-risk analysis to be a useful tool for structuring, quantifying, and communicating the relative benefit and safety profiles of drugs in a transparent, rational and consistent way. The two frameworks were similar. MCDA is a generic and flexible methodology that can be used to perform a structured benefit-risk in any common context. wNCB is a special case of MCDA and is shown to be equivalent to an extension of the number needed to treat (NNT) principle. It is simpler to apply and understand than MCDA and can be applied when all outcomes are measured on a binary scale.

]]>In randomized clinical trials where the times to event of two treatment groups are compared under a proportional hazards assumption, it has been established that omitting prognostic factors from the model entails an underestimation of the hazards ratio. Heterogeneity due to unobserved covariates in cancer patient populations is a concern since genomic investigations have revealed molecular and clinical heterogeneity in these populations. In HIV prevention trials, heterogeneity is unavoidable and has been shown to decrease the treatment effect over time. This article assesses the influence of trial duration on the bias of the estimated hazards ratio resulting from omitting covariates from the Cox analysis. The true model is defined by including an unobserved random frailty term in the individual hazard that reflects the omitted covariate. Three frailty distributions are investigated: gamma, log-normal, and binary, and the asymptotic bias of the hazards ratio estimator is calculated. We show that the attenuation of the treatment effect resulting from unobserved heterogeneity strongly increases with trial duration, especially for continuous frailties that are likely to reflect omitted covariates, as they are often encountered in practice. The possibility of interpreting the long-term decrease in treatment effects as a bias induced by heterogeneity and trial duration is illustrated by a trial in oncology where adjuvant chemotherapy in stage 1B NSCLC was investigated.

]]>Discrimination statistics describe the ability of a survival model to assign higher risks to individuals who experience earlier events: examples are Harrell's C-index and Royston and Sauerbrei's D, which we call the D-index. Prognostic covariates whose distributions are controlled by the study design (e.g. age and sex) influence discrimination and can make it difficult to compare model discrimination between studies. Although covariate adjustment is a standard procedure for quantifying disease-risk factor associations, there are no covariate adjustment methods for discrimination statistics in censored survival data.

To develop extensions of the C-index and D-index that describe the prognostic ability of a model adjusted for one or more covariate(s).

We define a covariate-adjusted C-index and D-index for censored survival data, propose several estimators, and investigate their performance in simulation studies and in data from a large individual participant data meta-analysis, the Emerging Risk Factors Collaboration.

The proposed methods perform well in simulations. In the Emerging Risk Factors Collaboration data, the age-adjusted C-index and D-index were substantially smaller than unadjusted values. The study-specific standard deviation of baseline age was strongly associated with the unadjusted C-index and D-index but not significantly associated with the age-adjusted indices.

The proposed estimators improve meta-analysis comparisons, are easy to implement and give a more meaningful clinical interpretation.

In many areas of science where empirical data are analyzed, a task is often to identify important variables with influence on an outcome. Most often this is done by using a variable selection strategy in the context of a multivariable regression model. Using a study on ozone effects in children (*n* = 496, 24 covariates), we will discuss aspects relevant for deriving a suitable model. With an emphasis on model stability, we will explore and illustrate differences between predictive models and explanatory models, the key role of stopping criteria, and the value of bootstrap resampling (with and without replacement). Bootstrap resampling will be used to assess variable selection stability, to derive a predictor that incorporates model uncertainty, check for influential points, and visualize the variable selection process. For the latter two tasks we adapt and extend recent approaches, such as stability paths, to serve our purposes. Based on earlier experiences and on results from the example, we will argue for simpler models and that predictions are usually very similar, irrespective of the selection method used. Important differences exist for the corresponding variances, and the model uncertainty concept helps to protect against serious underestimation of the variance of a predictor-derived data dependently. Results of stability investigations illustrate severe difficulties in the task of deriving a suitable explanatory model. It seems possible to identify a small number of variables with an important and probably true influence on the outcome, but too often several variables are included whose selection may be a result of chance or may depend on a small number of observations.

We develop an asymptotic likelihood ratio test for multivariate lognormal data with a point mass at zero in each dimension. The test generalizes Wilks' lambda and Hotelling *T*-test to the case of semicontinuous data. Simulations show that the resulting test statistic attains the nominal Type I error rate and has good power for reasonable alternatives. We conclude with an application to exploration of ecological niches of trees of South Africa.

New markers may improve prediction of diagnostic and prognostic outcomes. We aimed to review options for graphical display and summary measures to assess the predictive value of markers over standard, readily available predictors. We illustrated various approaches using previously published data on 3264 participants from the Framingham Heart Study, where 183 developed coronary heart disease (10-year risk 5.6%). We considered performance measures for the incremental value of adding HDL cholesterol to a prediction model. An initial assessment may consider statistical significance (HR = 0.65, 95% confidence interval 0.53 to 0.80; likelihood ratio *p* < 0.001), and distributions of predicted risks (densities or box plots) with various summary measures. A range of decision thresholds is considered in predictiveness and receiver operating characteristic curves, where the area under the curve (AUC) increased from 0.762 to 0.774 by adding HDL. We can furthermore focus on reclassification of participants with and without an event in a reclassification graph, with the continuous net reclassification improvement (NRI) as a summary measure. When we focus on one particular decision threshold, the changes in sensitivity and specificity are central. We propose a net reclassification risk graph, which allows us to focus on the number of reclassified persons and their event rates. Summary measures include the binary AUC, the two-category NRI, and decision analytic variants such as the net benefit (NB). Various graphs and summary measures can be used to assess the incremental predictive value of a marker. Important insights for impact on decision making are provided by a simple graph for the net reclassification risk.

A method for simultaneously assessing noninferiority with respect to efficacy and superiority with respect to another endpoint in two-arm noninferiority trials is presented. The procedure controls both the average type I error rate for the intersection-union test problem and the frequentist type I error rate for the noninferiority test by α while allowing an increased level for the superiority test. For normally distributed outcomes, two methods are presented to deal with the uncertainty about the correlation between the endpoints which defines the adjusted levels. The operating characteristics of these procedures are investigated. Furthermore, the sample size required when applying the proposed method is compared with that of alternative procedures. Application of the method in the situation of binary endpoints and mixed normal and binary endpoints, respectively, is sketched. An illustrative example is provided demonstrating implementation of the proposed approach in a clinical trial.

]]>Survey data often contain measurements for variables that are semicontinuous in nature, i.e. they either take a single fixed value (we assume this is zero) or they have a continuous, often skewed, distribution on the positive real line. Standard methods for small area estimation (SAE) based on the use of linear mixed models can be inefficient for such variables. We discuss SAE techniques for semicontinuous variables under a two part random effects model that allows for the presence of excess zeros as well as the skewed nature of the nonzero values of the response variable. In particular, we first model the excess zeros via a generalized linear mixed model fitted to the probability of a nonzero, i.e. strictly positive, value being observed, and then model the response, given that it is strictly positive, using a linear mixed model fitted on the logarithmic scale. Empirical results suggest that the proposed method leads to efficient small area estimates for semicontinuous data of this type. We also propose a parametric bootstrap method to estimate the MSE of the proposed small area estimator. These bootstrap estimates of the MSE are compared to the true MSE in a simulation study.

]]>This paper presents an extension of the joint modeling strategy for the case of multiple longitudinal outcomes and repeated infections of different types over time, motivated by postkidney transplantation data. Our model comprises two parts linked by shared latent terms. On the one hand is a multivariate mixed linear model with random effects, where a low-rank thin-plate spline function is incorporated to collect the nonlinear behavior of the different profiles over time. On the other hand is an infection-specific Cox model, where the dependence between different types of infections and the related times of infection is through a random effect associated with each infection type to catch the *within* dependence and a shared frailty parameter to capture the dependence *between* infection types. We implemented the parameterization used in joint models which uses the fitted longitudinal measurements as time-dependent covariates in a relative risk model. Our proposed model was implemented in OpenBUGS using the MCMC approach.

In this paper, we introduce a new model for recurrent event data characterized by a baseline rate function fully parametric, which is based on the exponential-Poisson distribution. The model arises from a latent competing risk scenario, in the sense that there is no information about which cause was responsible for the event occurrence. Then, the time of each recurrence is given by the minimum lifetime value among all latent causes. The new model has a particular case, which is the classical homogeneous Poisson process. The properties of the proposed model are discussed, including its hazard rate function, survival function, and ordinary moments. The inferential procedure is based on the maximum likelihood approach. We consider an important issue of model selection between the proposed model and its particular case by the likelihood ratio test and score test. Goodness of fit of the recurrent event models is assessed using Cox-Snell residuals. A simulation study evaluates the performance of the estimation procedure in the presence of a small and moderate sample sizes. Applications on two real data sets are provided to illustrate the proposed methodology. One of them, first analyzed by our team of researchers, considers the data concerning the recurrence of malaria, which is an infectious disease caused by a protozoan parasite that infects red blood cells.

]]>Recurrent event data arise in longitudinal follow-up studies, where each subject may experience the same type of events repeatedly. The work in this article is motivated by the data from a study of repeated peritonitis for patients on peritoneal dialysis. Due to the aspects of medicine and cost, the peritonitis cases were classified into two types: Gram-positive and non-Gram-positive peritonitis. Further, since the death and hemodialysis therapy preclude the occurrence of recurrent events, we face multivariate recurrent event data with a dependent terminal event. We propose a flexible marginal model, which has three characteristics: first, we assume marginal proportional hazard and proportional rates models for terminal event time and recurrent event processes, respectively; second, the inter-recurrences dependence and the correlation between the multivariate recurrent event processes and terminal event time are modeled through three multiplicative frailties corresponding to the specified marginal models; third, the rate model with frailties for recurrent events is specified only on the time before the terminal event. We propose a two-stage estimation procedure for estimating unknown parameters. We also establish the consistency of the two-stage estimator. Simulation studies show that the proposed approach is appropriate for practical use. The methodology is applied to the peritonitis cohort data that motivated this study.

]]>The problem of variable selection in the generalized linear-mixed models (GLMMs) is pervasive in statistical practice. For the purpose of variable selection, many methodologies for determining the best subset of explanatory variables currently exist according to the model complexity and differences between applications. In this paper, we develop a “higher posterior probability model with bootstrap” (HPMB) approach to select explanatory variables without fitting all possible GLMMs involving a small or moderate number of explanatory variables. Furthermore, to save computational load, we propose an efficient approximation approach with Laplace's method and Taylor's expansion to approximate intractable integrals in GLMMs. Simulation studies and an application of HapMap data provide evidence that this selection approach is computationally feasible and reliable for exploring true candidate genes and gene–gene associations, after adjusting for complex structures among clusters.

]]>Marginal structural models (MSMs) have been proposed for estimating a treatment's effect, in the presence of time-dependent confounding. We aimed to evaluate the performance of the Cox MSM in the presence of missing data and to explore methods to adjust for missingness. We simulated data with a continuous time-dependent confounder and a binary treatment. We explored two classes of missing data: (i) missed visits, which resemble clinical cohort studies; (ii) missing confounder's values, which correspond to interval cohort studies. Missing data were generated under various mechanisms. In the first class, the source of the bias was the extreme treatment weights. Truncation or normalization improved estimation. Therefore, particular attention must be paid to the distribution of weights, and truncation or normalization should be applied if extreme weights are noticed. In the second case, bias was due to the misspecification of the treatment model. Last observation carried forward (LOCF), multiple imputation (MI), and inverse probability of missingness weighting (IPMW) were used to correct for the missingness. We found that alternatives, especially the IPMW method, perform better than the classic LOCF method. Nevertheless, in situations with high marker's variance and rarely recorded measurements none of the examined method adequately corrected the bias.

]]>In recent years, the evaluation of healthcare provider performance has become standard for governments, insurance companies, and other stakeholders. Often, performance is compared across providers using indicators in one time period, for example a year. However it is often important to assess changes in the performance of individual providers over time. Such analyses can be used to determine if any providers have significant improvements, deteriorations, unusual patterns or systematic changes in performance. Studies which monitor healthcare provider performance in this way have to date typically been limited to comparing performance in the most recent period with performance in a previous period. It is also important to consider a longer-term view of performance and assess changes over more than two periods. In this paper, we develop test statistics that account for variable numbers of prior performance indicators, and show that these are particularly useful for assessing consecutive improvements or deteriorations in performance. We apply the tests to coronary artery bypass graft mortality rates in New York State hospitals, and mortality data from Australian and New Zealand intensive care units. Although our applications are to medical data, the new tests have broad application in other areas.

]]>Methods to examine whether genetic and/or environmental sources can account for the residual variation in ordinal family data usually assume proportional odds. However, standard software to fit the non-proportional odds model to ordinal family data is limited because the correlation structure of family data is more complex than for other types of clustered data. To perform these analyses we propose the non-proportional odds multivariate logistic regression model and take a simulation-based approach to model fitting using Markov chain Monte Carlo methods, such as partially collapsed Gibbs sampling and the Metropolis algorithm. We applied the proposed methodology to male pattern baldness data from the Victorian Family Heart Study.

]]>Evaluation of diagnostic performance is typically based on the receiver operating characteristic (ROC) curve and the area under the curve (AUC) as its summary index. The partial area under the curve (pAUC) is an alternative index focusing on the range of practical/clinical relevance. One of the problems preventing more frequent use of the pAUC is the perceived loss of efficiency in cases of noncrossing ROC curves. In this paper, we investigated statistical properties of comparisons of two correlated pAUCs. We demonstrated that outside of the classic model there are practically reasonable ROC types for which comparisons of noncrossing concave curves would be more powerful when based on a part of the curve rather than the entire curve. We argue that this phenomenon stems in part from the exclusion of noninformative parts of the ROC curves that resemble straight-lines. We conducted extensive simulation studies in families of binormal, straight-line, and bigamma ROC curves. We demonstrated that comparison of pAUCs is statistically more powerful than comparison of full AUCs when ROC curves are close to a “straight line”. For less flat binormal ROC curves an increase in the integration range often leads to a disproportional increase in pAUCs’ difference, thereby contributing to an increase in statistical power. Thus, efficiency of differences in pAUCs of noncrossing ROC curves depends on the shape of the curves, and for families of ROC curves that are nearly straight-line shaped, such as bigamma ROC curves, there are multiple practical scenarios in which comparisons of pAUCs are preferable.

]]>This paper presents a collection of dissimilarity measures to describe and then classify spatial point patterns when multiple replicates of different types are available for analysis. In particular, we consider a range of distances including the spike-time distance and its variants, as well as cluster-based distances and dissimilarity measures based on classical statistical summaries of point patterns. We review and explore, in the form of a tutorial, their uses, and their pros and cons. These distances are then used to summarize and describe collections of repeated realizations of point patterns via prototypes and multidimensional scaling. We also show a simulation study to evaluate the performance of multidimensional scaling with two types of selected distances. Finally, a multivariate spatial point pattern of a natural plant community is analyzed through various of these measures of dissimilarity.

]]>