Quick Take

  • An interpretable XGBoost model using routinely available electronic health record (EHR
  • Epic/Clarity) vitals, laboratory, and microbiology data flagged 425 antibiotic orders with area under the receiver operating characteristic curve (AUROC) 0.89 for necessity and 0.80 for optimal selection.
  • The model can help antibiotic stewardship programs (ASPs) and pharmacists prioritize high-yield prospective audit-and-feedback (PAF) reviews, reducing manual workload and focusing interventions on unstable or inflammatory cases.

Why it Matters

  • Hospitalized patients frequently receive antibiotics, and a sizable fraction of in-hospital use is unnecessary or suboptimal, contributing to adverse drug events, Clostridioides difficile infection, and antimicrobial resistance while requiring empiric decisions with limited early data.
  • Prospective audit-and-feedback (PAF) improves appropriateness but is labor-intensive and cannot feasibly cover the volume of orders
  • many hospitals therefore perform limited, periodic point-prevalence reviews that leave many potentially inappropriate orders unreviewed.
  • Automated, EHR-based prescreening to triage antibiotic orders could enable stewardship teams and pharmacists to concentrate scarce resources on unstable or high-yield cases and inform clinical decision support and operational resource allocation.

What They Did

  • Single-center study at a tertiary academic hospital reviewing inpatient antibiotic orders on hospital medicine services between May 2021 and August 2022
  • intensive care, surgical, cardiology, malignant hematology services, and patients with infectious diseases consultation were excluded.
  • Three infectious-disease experts performed point-prevalence reviews using the National Antimicrobial Prescribing Survey (NAPS) to label individual orders as 'indicated' or 'optimal' for use.
  • EHR data were extracted from Epic/Clarity (vitals, laboratories, microbiology, medication administration record), censored to the day before review, and used to train an interpretable time-series XGBoost model with 10-fold cross-validation
  • performance was compared to simpler clinical scores to prioritize orders for stewardship review.

What They Found

  • XGBoost (n = 425 orders) achieved AUROC 0.89 (95% CI 0.87–0.91) for necessity and 0.80 (95% CI 0.77–0.83) for optimal antimicrobial selection
  • reviewers labeled 75% of orders as 'indicated' and 56% as 'optimal'.
  • Models using only vital signs performed strongly (AUROC 0.87/0.77 indicated/optimal)
  • laboratory-only models 0.76/0.76
  • microbiology/medication administration record (MAR) alone performed worse (0.74/0.66)
  • systemic inflammatory response syndrome (SIRS) criteria performed poorly (AUROC 0.42).
  • Performance was consistent across antibiotic spectra: broad-spectrum hospital onset (BSHO) 0.86/0.77, broad-spectrum community onset (BSCA) 0.87/0.81, MRSA 0.87/0.77, ESBL 0.89/0.73, Pseudomonas 0.82/0.78.
  • SHAP (SHapley Additive exPlanations) identified hemodynamic variability, inflammatory markers, and recent laboratory trends as the dominant drivers of predictions rather than microbiology or antibiotic class.
  • Practically, automated triage could prioritize high-yield PAF early in therapy (60% of orders < 48 hours), concentrating review on unstable or inflammatory cases driven mainly by vital sign and lab trends rather than early microbiology results.

Takeaways

  • Integrate the model into a daily Epic/Clarity feed and present prioritized antibiotic orders on a stewardship dashboard or workqueue to focus prospective audit-and-feedback on unstable/inflammatory cases and early therapy orders.
  • Operationalize via a nightly Clarity export, calibrate alert thresholds to match ASP and pharmacy capacity, and display SHAP-derived feature drivers in the user interface so pharmacists can rapidly interpret why an order was flagged.
  • Treat the tool as an AI 'triage nurse'—it sorts high-yield orders for human review, enabling stewardship teams to allocate time to complex cases while preserving final clinical judgment.
  • Maintain pharmacist oversight and governance: require pharmacist sign-off on flagged orders, exclude infectious diseases–consulted patients per model training, and perform periodic audits to monitor performance and guide refinements.

Strengths and Limitations

Strengths:

  • Uses an interpretable time-series XGBoost model with SHAP and reporting aligned to TRIPOD guidelines, enhancing transparency and reproducibility.
  • Relies on routinely available structured EHR data (vitals, labs) with automated extraction, supporting scalable, timely triage without manual chart review.

Limitations:

  • Single-center, hospital-medicine–only cohort and modest sample size limit external validity across intensive care units, surgical services, and community hospitals.
  • Model excludes unstructured clinical notes and patients with infectious diseases consultation
  • periodic NAPS point-prevalence labeling with inter-rater variability may miss contextual factors (for example, prehospital antibiotics or nuanced comorbidities) that affect label fidelity and deployment performance.

Bottom Line

This interpretable XGBoost triage model reliably prioritizes high-yield antibiotic orders and is ready for real-world pilot deployment within ASP and pharmacy workflows with local calibration and continued pharmacist governance.