Maintaining a National Acute Kidney Injury Risk Prediction Model to Support Local Quality Benchmarking.
acute kidney injury
benchmarking
models, statistical
supervised machine learning
veterans
Journal
Circulation. Cardiovascular quality and outcomes
ISSN: 1941-7705
Titre abrégé: Circ Cardiovasc Qual Outcomes
Pays: United States
ID NLM: 101489148
Informations de publication
Date de publication:
08 2022
08 2022
Historique:
pubmed:
13
8
2022
medline:
19
8
2022
entrez:
12
8
2022
Statut:
ppublish
Résumé
The utility of quality dashboards to inform decision-making and improve clinical outcomes is tightly linked to the accuracy of the information they provide and, in turn, accuracy of underlying prediction models. Despite recognition of the need to update prediction models to maintain accuracy over time, there is limited guidance on updating strategies. We compare predefined and surveillance-based updating strategies applied to a model supporting quality evaluations among US veterans. We evaluated the performance of a US Department of Veterans Affairs-specific model for postcardiac catheterization acute kidney injury using routinely collected observational data over the 6 years following model development (n=90 295 procedures in 2013-2019). Predicted probabilities were generated from the original model, an annually retrained model, and a surveillance-based approach that monitored performance to inform the timing and method of updates. We evaluated how updating the national model impacted regional quality profiles. We compared observed-to-expected outcome ratios, where values above and below 1 indicated more and fewer adverse outcomes than expected, respectively. The original model overpredicted risk at the national level (observed-to-expected outcome ratio, 0.75 [0.74-0.77]). Annual retraining updated the model 5×; surveillance-based updating retrained once and recalibrated twice. While both strategies improved performance, the surveillance-based approach provided superior calibration (observed-to-expected outcome ratio, 1.01 [0.99-1.03] versus 0.94 [0.92-0.96]). Overprediction by the original model led to optimistic quality assessments, incorrectly indicating most of the US Department of Veterans Affairs' 18 regions observed fewer acute kidney injury events than predicted. Both updating strategies revealed 16 regions performed as expected and 2 regions increasingly underperformed, having more acute kidney injury events than predicted. Miscalibrated clinical prediction models provide inaccurate pictures of performance across clinical units, and degrading calibration further complicates our understanding of quality. Updating strategies tailored to health system needs and capacity should be incorporated into model implementation plans to promote the utility and longevity of quality reporting tools.
Sections du résumé
BACKGROUND
The utility of quality dashboards to inform decision-making and improve clinical outcomes is tightly linked to the accuracy of the information they provide and, in turn, accuracy of underlying prediction models. Despite recognition of the need to update prediction models to maintain accuracy over time, there is limited guidance on updating strategies. We compare predefined and surveillance-based updating strategies applied to a model supporting quality evaluations among US veterans.
METHODS
We evaluated the performance of a US Department of Veterans Affairs-specific model for postcardiac catheterization acute kidney injury using routinely collected observational data over the 6 years following model development (n=90 295 procedures in 2013-2019). Predicted probabilities were generated from the original model, an annually retrained model, and a surveillance-based approach that monitored performance to inform the timing and method of updates. We evaluated how updating the national model impacted regional quality profiles. We compared observed-to-expected outcome ratios, where values above and below 1 indicated more and fewer adverse outcomes than expected, respectively.
RESULTS
The original model overpredicted risk at the national level (observed-to-expected outcome ratio, 0.75 [0.74-0.77]). Annual retraining updated the model 5×; surveillance-based updating retrained once and recalibrated twice. While both strategies improved performance, the surveillance-based approach provided superior calibration (observed-to-expected outcome ratio, 1.01 [0.99-1.03] versus 0.94 [0.92-0.96]). Overprediction by the original model led to optimistic quality assessments, incorrectly indicating most of the US Department of Veterans Affairs' 18 regions observed fewer acute kidney injury events than predicted. Both updating strategies revealed 16 regions performed as expected and 2 regions increasingly underperformed, having more acute kidney injury events than predicted.
CONCLUSIONS
Miscalibrated clinical prediction models provide inaccurate pictures of performance across clinical units, and degrading calibration further complicates our understanding of quality. Updating strategies tailored to health system needs and capacity should be incorporated into model implementation plans to promote the utility and longevity of quality reporting tools.
Identifiants
pubmed: 35959674
doi: 10.1161/CIRCOUTCOMES.121.008635
pmc: PMC9388604
mid: NIHMS1812490
doi:
Types de publication
Journal Article
Research Support, N.I.H., Extramural
Langues
eng
Sous-ensembles de citation
IM
Pagination
e008635Subventions
Organisme : NIDDK NIH HHS
ID : R01 DK113201
Pays : United States
Commentaires et corrections
Type : CommentIn
Références
Neurosurgery. 2014 Mar;74(3):235-43; discussion 243-4
pubmed: 24335812
Catheter Cardiovasc Interv. 2008 Sep 1;72(3):347-354
pubmed: 18729173
J Am Med Inform Assoc. 2019 Dec 1;26(12):1448-1457
pubmed: 31397478
J Am Soc Nephrol. 2005 Nov;16(11):3365-70
pubmed: 16177006
Int J Epidemiol. 2020 Aug 1;49(4):1316-1325
pubmed: 32243524
Circ Cardiovasc Qual Outcomes. 2014 Sep;7(5):693-700
pubmed: 25074372
Epidemiology. 2010 Jan;21(1):128-38
pubmed: 20010215
Circulation. 2017 Mar 7;135(10):e146-e603
pubmed: 28122885
Jt Comm J Qual Patient Saf. 2015 Sep;41(9):414-20
pubmed: 26289236
Circ Cardiovasc Interv. 2010 Aug;3(4):346-50
pubmed: 20587788
J Biomed Inform. 2015 Apr;54:283-93
pubmed: 25579635
Am J Cardiol. 2015 May 1;115(9):1174-8
pubmed: 25759106
J Biomed Inform. 2020 Dec;112:103611
pubmed: 33157313
BMC Med Inform Decis Mak. 2011 Dec 14;11:75
pubmed: 22168892
Eur J Cardiothorac Surg. 2013 Jun;43(6):1146-52
pubmed: 23152436
JACC Cardiovasc Interv. 2014 Jan;7(1):1-9
pubmed: 24456715
J Am Med Inform Assoc. 2017 Nov 01;24(6):1052-1061
pubmed: 28379439
Ann Thorac Surg. 2010 Mar;89(3):677-82
pubmed: 20172107
Heart. 2012 May;98(9):691-8
pubmed: 22397946
Am Heart J. 2008 Jan;155(1):114-20
pubmed: 18082501
J Am Heart Assoc. 2015 Dec 11;4(12):
pubmed: 26656858
Circulation. 2013 Jan 1;127(1):e6-e245
pubmed: 23239837
J Am Coll Cardiol. 2012 Jun 19;59(25):2309-16
pubmed: 22698487
Circ Cardiovasc Qual Outcomes. 2016 Mar;9(2):171-81
pubmed: 26933048
Med Decis Making. 2009 Mar-Apr;29(2):247-56
pubmed: 19015285
J Clin Epidemiol. 2008 Nov;61(11):1085-94
pubmed: 19208371
AMIA Annu Symp Proc. 2020 Mar 04;2019:1002-1010
pubmed: 32308897
Stat Med. 2017 Dec 10;36(28):4529-4539
pubmed: 27891652
Radiology. 1982 Apr;143(1):29-36
pubmed: 7063747
Crit Care. 2013 Feb 04;17(1):205
pubmed: 23394215
Intensive Care Med. 2012 Jan;38(1):40-6
pubmed: 22042520
Circ Cardiovasc Interv. 2013 Feb;6(1):37-43
pubmed: 23322741
BMJ Qual Saf. 2012 Jan;21(1):54-62
pubmed: 21890755
Diagn Progn Res. 2017 Feb 08;1:2
pubmed: 31093534