Skip to main content

Table 1 Summary points on calibration

From: Calibration: the Achilles heel of predictive analytics

Why calibration matters- Decisions are often based on risk, so predicted risks should be reliable
- Poor calibration may make a prediction model clinically useless or even harmful
Causes of poor calibration- Statistical overfitting and measurement error
- Heterogeneity in populations in terms of patient characteristics, disease incidence or prevalence, patient management, and treatment policies
Assessment of calibration in practice- Perfect calibration, where predicted risks are correct for every covariate pattern, is utopic; we should not aim for that
- At model development, focus on nonlinear effects and interaction terms only if a sufficiently large sample size is available; low sample sizes require simpler modeling strategies or that no model is developed at all
- Avoid the Hosmer–Lemeshow test to assess or prove calibration
- At internal validation, focus on the calibration slope as a part of the assessment of statistical overfitting
- At external validation, focus on the calibration curve, intercept and slope
- Model updating should be considered in case of poor calibration; re-estimating the model entirely requires sufficient data