Adaptive sample size determination for the development of clinical prediction models.

Adaptive design Clinical prediction models Events per variable Model development Model validation Sample size

Journal

Diagnostic and prognostic research
ISSN: 2397-7523
Titre abrégé: Diagn Progn Res
Pays: England
ID NLM: 101718985

Informations de publication

Date de publication:
22 Mar 2021
Historique:
received: 01 10 2020
accepted: 15 02 2021
entrez: 22 3 2021
pubmed: 23 3 2021
medline: 23 3 2021
Statut: epublish

Résumé

We suggest an adaptive sample size calculation method for developing clinical prediction models, in which model performance is monitored sequentially as new data comes in. We illustrate the approach using data for the diagnosis of ovarian cancer (n = 5914, 33% event fraction) and obstructive coronary artery disease (CAD; n = 4888, 44% event fraction). We used logistic regression to develop a prediction model consisting only of a priori selected predictors and assumed linear relations for continuous predictors. We mimicked prospective patient recruitment by developing the model on 100 randomly selected patients, and we used bootstrapping to internally validate the model. We sequentially added 50 random new patients until we reached a sample size of 3000 and re-estimated model performance at each step. We examined the required sample size for satisfying the following stopping rule: obtaining a calibration slope ≥ 0.9 and optimism in the c-statistic (or AUC) < = 0.02 at two consecutive sample sizes. This procedure was repeated 500 times. We also investigated the impact of alternative modeling strategies: modeling nonlinear relations for continuous predictors and correcting for bias on the model estimates (Firth's correction). Better discrimination was achieved in the ovarian cancer data (c-statistic 0.9 with 7 predictors) than in the CAD data (c-statistic 0.7 with 11 predictors). Adequate calibration and limited optimism in discrimination was achieved after a median of 450 patients (interquartile range 450-500) for the ovarian cancer data (22 events per parameter (EPP), 20-24) and 850 patients (750-900) for the CAD data (33 EPP, 30-35). A stricter criterion, requiring AUC optimism < = 0.01, was met with a median of 500 (23 EPP) and 1500 (59 EPP) patients, respectively. These sample sizes were much higher than the well-known 10 EPP rule of thumb and slightly higher than a recently published fixed sample size calculation method by Riley et al. Higher sample sizes were required when nonlinear relationships were modeled, and lower sample sizes when Firth's correction was used. Adaptive sample size determination can be a useful supplement to fixed a priori sample size calculations, because it allows to tailor the sample size to the specific prediction modeling context in a dynamic fashion.

Sections du résumé

BACKGROUND BACKGROUND
We suggest an adaptive sample size calculation method for developing clinical prediction models, in which model performance is monitored sequentially as new data comes in.
METHODS METHODS
We illustrate the approach using data for the diagnosis of ovarian cancer (n = 5914, 33% event fraction) and obstructive coronary artery disease (CAD; n = 4888, 44% event fraction). We used logistic regression to develop a prediction model consisting only of a priori selected predictors and assumed linear relations for continuous predictors. We mimicked prospective patient recruitment by developing the model on 100 randomly selected patients, and we used bootstrapping to internally validate the model. We sequentially added 50 random new patients until we reached a sample size of 3000 and re-estimated model performance at each step. We examined the required sample size for satisfying the following stopping rule: obtaining a calibration slope ≥ 0.9 and optimism in the c-statistic (or AUC) < = 0.02 at two consecutive sample sizes. This procedure was repeated 500 times. We also investigated the impact of alternative modeling strategies: modeling nonlinear relations for continuous predictors and correcting for bias on the model estimates (Firth's correction).
RESULTS RESULTS
Better discrimination was achieved in the ovarian cancer data (c-statistic 0.9 with 7 predictors) than in the CAD data (c-statistic 0.7 with 11 predictors). Adequate calibration and limited optimism in discrimination was achieved after a median of 450 patients (interquartile range 450-500) for the ovarian cancer data (22 events per parameter (EPP), 20-24) and 850 patients (750-900) for the CAD data (33 EPP, 30-35). A stricter criterion, requiring AUC optimism < = 0.01, was met with a median of 500 (23 EPP) and 1500 (59 EPP) patients, respectively. These sample sizes were much higher than the well-known 10 EPP rule of thumb and slightly higher than a recently published fixed sample size calculation method by Riley et al. Higher sample sizes were required when nonlinear relationships were modeled, and lower sample sizes when Firth's correction was used.
CONCLUSIONS CONCLUSIONS
Adaptive sample size determination can be a useful supplement to fixed a priori sample size calculations, because it allows to tailor the sample size to the specific prediction modeling context in a dynamic fashion.

Identifiants

pubmed: 33745449
doi: 10.1186/s41512-021-00096-5
pii: 10.1186/s41512-021-00096-5
pmc: PMC7983402
doi:

Types de publication

Journal Article

Langues

eng

Pagination

6

Subventions

Organisme : Research Foundation - Flanders (FWO)
ID : G0B4716N
Organisme : Internal Funds KU Leuven
ID : C24/15/037

Références

J Clin Epidemiol. 2011 Dec;64(12):1464-5; author reply 1463-4
pubmed: 22032755
BMC Med Res Methodol. 2016 Oct 26;16(1):144
pubmed: 27782817
BMJ. 2014 Oct 15;349:g5920
pubmed: 25320247
Stat Methods Med Res. 2019 Aug;28(8):2455-2474
pubmed: 29966490
BMC Med Res Methodol. 2016 Nov 24;16(1):163
pubmed: 27881078
Am J Obstet Gynecol. 2016 Jan;214(1):79-90.e36
pubmed: 26070707
Stat Methods Med Res. 2017 Apr;26(2):796-808
pubmed: 25411322
Stat Med. 2019 Mar 30;38(7):1276-1296
pubmed: 30357870
BMC Med Res Methodol. 2014 Oct 16;14:116
pubmed: 25323009
BMJ. 2020 Apr 7;369:m1328
pubmed: 32265220
J Clin Epidemiol. 2018 Jun;98:133-143
pubmed: 29174118
Stat Methods Med Res. 2020 Nov;29(11):3166-3178
pubmed: 32401702
Med Decis Making. 2001 Jan-Feb;21(1):45-56
pubmed: 11206946
BMC Med Res Methodol. 2014 Dec 22;14:137
pubmed: 25532820
Diagn Progn Res. 2017 Dec 21;1:20
pubmed: 31093549
Stat Med. 2019 Mar 30;38(7):1262-1275
pubmed: 30347470
J Clin Epidemiol. 1996 Dec;49(12):1373-9
pubmed: 8970487
J Clin Epidemiol. 2016 Jun;74:167-76
pubmed: 26772608
J Clin Epidemiol. 2011 Sep;64(9):993-1000
pubmed: 21411281
BMJ. 2020 Mar 18;368:m441
pubmed: 32188600
J Clin Epidemiol. 2001 Aug;54(8):774-81
pubmed: 11470385
Stat Med. 1990 Nov;9(11):1303-25
pubmed: 2277880
Stat Med. 2016 Oct 15;35(23):4124-35
pubmed: 27193918
BMJ Open. 2017 Apr 7;7(4):e014467
pubmed: 28389492

Auteurs

Evangelia Christodoulou (E)

Department of Development & Regeneration, KU Leuven, Leuven, Belgium.

Maarten van Smeden (M)

Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht, Netherlands.

Michael Edlinger (M)

Department of Development & Regeneration, KU Leuven, Leuven, Belgium.
Department of Medical Statistics, Informatics, and Health Economics, Medical University Innsbruck, Innsbruck, Austria.

Dirk Timmerman (D)

Department of Development & Regeneration, KU Leuven, Leuven, Belgium.
Department of Obstetrics and Gynecology, University Hospitals Leuven, Leuven, Belgium.

Maria Wanitschek (M)

University Clinic of Internal Medicine III - Cardiology and Angiology, Tirol Kliniken, Innsbruck, Austria.

Ewout W Steyerberg (EW)

Department of Biomedical Data Sciences, Leiden University Medical Center, Leiden, Netherlands.

Ben Van Calster (B)

Department of Development & Regeneration, KU Leuven, Leuven, Belgium. ben.vancalster@kuleuven.be.
Department of Biomedical Data Sciences, Leiden University Medical Center, Leiden, Netherlands. ben.vancalster@kuleuven.be.
EPI-centre, KU Leuven, Leuven, Belgium. ben.vancalster@kuleuven.be.

Classifications MeSH