Volume 35, Issue 5
Research Article

A new test and graphical tool to assess the goodness of fit of logistic regression models

Giovanni Nattino

Corresponding Author

GiViTI Coordinating Center, Laboratory of Clinical Epidemiology, IRCCS – Istituto di Ricerche Farmacologiche 'Mario Negri', Villa Camozzi, Ranica (BG), Italy

Correspondence to: Giovanni Nattino, Laboratory of Clinical Epidemiology, IRCCS – Istituto di Ricerche Farmacologiche Mario Negri, Villa Camozzi, Ranica (BG), Italy.

E‐mail: giovanni.nattino@marionegri.it

Search for more papers by this author
Stefano Finazzi

Laboratoire Matériaux et Phénomènes Quantiques, Université Paris Diderot‐Paris 7, Paris, France

Search for more papers by this author
Guido Bertolini

GiViTI Coordinating Center, Laboratory of Clinical Epidemiology, IRCCS – Istituto di Ricerche Farmacologiche 'Mario Negri', Villa Camozzi, Ranica (BG), Italy

Search for more papers by this author
First published: 05 October 2015
Citations: 20

Abstract

A prognostic model is well calibrated when it accurately predicts event rates. This is first determined by testing for goodness of fit with the development dataset. All existing tests and graphic tools designed for the purpose suffer several drawbacks, related mainly to the subgrouping of observations or to heavy dependence on arbitrary parameters. We propose a statistical test and a graphical method to assess the goodness of fit of logistic regression models, obtained through an extension of similar techniques developed for external validation. We analytically computed and numerically verified the distribution of the underlying statistic. Simulations on a set of realistic scenarios show that this test and the well‐known Hosmer–Lemeshow approach have similar type I error rates. The main advantage of this new approach is that the relationship between model predictions and outcome rates across the range of probabilities can be represented in the calibration belt plot, together with its statistical confidence. By readily spotting any deviations from the perfect fit, this new graphical tool is designed to identify, during the process of model development, poorly modeled variables that call for further investigation. This is illustrated through an example based on real data. Copyright © 2015 John Wiley & Sons, Ltd.

Number of times cited according to CrossRef: 20

  • Predictive scores for major bleeding after coronary artery bypass surgery in low operative risk patients, The Journal of Cardiovascular Surgery, 10.23736/S0021-9509.20.11048-6, 61, 2, (2020).
  • Assessing the goodness of fit of logistic regression models in large samples: A modification of the Hosmer‐Lemeshow test, Biometrics, 10.1111/biom.13249, 76, 2, (549-560), (2020).
  • Case-mix affects calibration of cardiosurgical severity scores, Minerva Anestesiologica, 10.23736/S0375-9393.20.14280-9, 86, 7, (2020).
  • Development of Climate Impact Response Functions for highly regulated water resource systems, Journal of Hydrology, 10.1016/j.jhydrol.2020.125251, 590, (125251), (2020).
  • AutoScore: A Machine Learning-Based Automatic Clinical Score Generator and Its Application to Mortality Prediction Using Electronic Health Records (Preprint), JMIR Medical Informatics, 10.2196/21798, (2020).
  • Machine learning-based dynamic mortality prediction after traumatic brain injury, Scientific Reports, 10.1038/s41598-019-53889-6, 9, 1, (2019).
  • Assessing the Calibration of Dichotomous Outcome Models with the Calibration Belt, The Stata Journal: Promoting communications on statistics and Stata, 10.1177/1536867X1801700414, 17, 4, (1003-1014), (2019).
  • Trends and prediction of antimicrobial susceptibility in urinary bacteria isolated in European emergency departments: the EuroUTI 2010-2016 Study, Journal of Antimicrobial Chemotherapy, 10.1093/jac/dkz274, (2019).
  • The predictive performance of SAPS 2 and SAPS 3 in an intermediate care unit for internal medicine at a German university transplant center; A retrospective analysis, PLOS ONE, 10.1371/journal.pone.0222164, 14, 9, (e0222164), (2019).
  • Mortality attributable to different Klebsiella susceptibility patterns and to the coverage of empirical antibiotic therapy: a cohort study on patients admitted to the ICU with infection, Intensive Care Medicine, 10.1007/s00134-018-5360-0, 44, 10, (1709-1719), (2018).
  • Validation of the ICH score in patients with spontaneous intracerebral haemorrhage admitted to the intensive care unit in Southern Spain, BMJ Open, 10.1136/bmjopen-2018-021719, 8, 8, (e021719), (2018).
  • Validation of the return of spontaneous circulation after cardiac arrest (RACA) score in two different national territories, Resuscitation, 10.1016/j.resuscitation.2018.11.012, (2018).
  • Assessing the Calibration of Dichotomous Outcome Models with the Calibration Belt, The Stata Journal: Promoting communications on statistics and Stata, 10.1177/1536867X1701700414, 17, 4, (1003-1014), (2018).
  • Development and validation of a risk-prediction nomogram for patients with ureteral calculi associated with urosepsis: A retrospective analysis, PLOS ONE, 10.1371/journal.pone.0201515, 13, 8, (e0201515), (2018).
  • A prediction model to identify people with spinal cord injury who are at high risk of dying within 5 years of discharge from hospital in Bangladesh, Spinal Cord, 10.1038/s41393-018-0211-y, (2018).
  • Common intensive care scoring systems do not outperform age and glasgow coma scale score in predicting mid-term mortality in patients with spontaneous intracerebral hemorrhage treated in the intensive care unit, Scandinavian Journal of Trauma, Resuscitation and Emergency Medicine, 10.1186/s13049-017-0448-z, 25, 1, (2017).
  • Are high nurse workload/staffing ratios associated with decreased survival in critically ill patients? A cohort study, Annals of Intensive Care, 10.1186/s13613-017-0269-2, 7, 1, (2017).
  • Risk prediction models for delirium in the intensive care unit after cardiac surgery: a systematic review and independent external validation, BJA: British Journal of Anaesthesia, 10.1093/bja/aew476, 118, 3, (391-399), (2017).
  • Intensive care medicine in 2050: statistical tools for development of prognostic models (why clinicians should not be ignored), Intensive Care Medicine, 10.1007/s00134-017-4825-x, 43, 9, (1403-1406), (2017).
  • Reconsidering lactate as a sepsis risk biomarker, PLOS ONE, 10.1371/journal.pone.0185320, 12, 10, (e0185320), (2017).

The full text of this article hosted at iucr.org is unavailable due to technical difficulties.