Can LLMs turbo-charge IPC?

Can LLMs turbo-charge IPC?

From Manual Monitoring to Machine-Augmented Insights: How LLMs Are Reshaping Infection Detection and Prevention

Dr. Chang Ho Yoon

March 20, 2025

Large language models are poised to become valuable allies in infection prevention and control (IPC). The jury will be out for some time but the evidence to date is already hinting at an enticing, brave new world of increased efficiency (think real-time screening for HAIs and auto-generating IPC reports) and improved detection of cases otherwise missed or delayed by traditional – i.e., manual – methods​ (Rodriguez-Nava et al.2024; Boussina et al.2024) (Rodriguez-Nava et al. 2024; Boussina et al. 2024).

What’s the main innovation here?
The analysis of complex unstructured and structured data often centred around language. At a high-level, the basic idea is to serve as on-demand knowledge resources for busy IPC team members. When compared to existing tools, LLM-based approaches often match or surpass performance benchmarks, for instance, an LLM system matching expert reviewers in 90% of sepsis cases (Boussina et al. 2024), or an AI model beating standard early warning scores in predicting infections (Creagh et al.2024)​. If we free clinicians from tedious documentation and digging through stodgy interfaces to find the data they need, can they provide quicker, more comprehensive IPC needs? Can we predict and prevent outbreaks within hospitals before they happen?

Large language models are poised to become valuable allies in infection prevention and control (IPC). The jury will be out for some time but the evidence to date is already hinting at an enticing, brave new world of increased efficiency (think real-time screening for HAIs and auto-generating IPC reports) and improved detection of cases otherwise missed or delayed by traditional – i.e., manual – methods​ (Rodriguez-Nava et al. 2024; Boussina et al. 2024).

What’s the main innovation here?
The analysis of complex unstructured and structured data often centred around language. At a high-level, the basic idea is to serve as on-demand knowledge resources for busy IPC team members. When compared to existing tools, LLM-based approaches often match or surpass performance benchmarks, for instance, an LLM system matching expert reviewers in 90% of sepsis cases (Boussina et al. 2024), or an AI model beating standard early warning scores in predicting infections​ (Creagh et al. 2024). If we free clinicians from tedious documentation and digging through stodgy interfaces to find the data they need, can they provide quicker, more comprehensive IPC needs? Can we predict and prevent outbreaks within hospitals before they happen?

For instance, an LLM system [matches] expert reviewers in 90% of sepsis cases

Aaron Boussina (Boussina et al. 2024).

For instance, an LLM system [matches] expert reviewers in 90% of sepsis cases

Aaron Boussina (Boussina et al. 2024).

LLMs are not magic bullets
LLMs are looking especially promising, but definitive empirical validation isn’t out there yet. Additionally, enthusiasm must be tempered by realism about current limitations. LLMs are not magic bullet – they sometimes fail in the exact scenarios where IPC cannot afford errors (complex, high-stakes decisions). The literature urges an augmented approach: use LLMs to handle workload and crunch data, but keep humans in the loop for oversight, validation, and the nuanced decision-making that machines can’t replicate (Wiemken & Carrico 2024; Schwartz et al.2024)


With this approach, the IPC community can harness LLMs as “co-pilots” in infection control: the AI handles the mundane and provides preliminary analyses, while IPC teams remain the ultimate decision-makers ensuring patient safety. Going forward, we clearly can’t rest on our laurels – continued research, real-world evaluations, and interdisciplinary collaboration (between, e.g., AI developers, IPC experts, policy-makers, ethicists) will be crucial.

LLMs have arrived and, even if they are soon to be subsumed by more general artificial intelligence models, the linguistic component will remain front and centre. At every turn, however, it’s our prerogative to step back and question the co-pilot status of these “knowledge companions”:

LLMs are not magic bullets
LLMs are looking especially promising, but definitive empirical validation isn’t out there yet. Additionally, enthusiasm must be tempered by realism about current limitations. LLMs are not magic bullet – they sometimes fail in the exact scenarios where IPC cannot afford errors (complex, high-stakes decisions). The literature urges an augmented approach: use LLMs to handle workload and crunch data, but keep humans in the loop for oversight, validation, and the nuanced decision-making that machines can’t replicate (Wiemken & Carrico 2024; Schwartz et al.2024)


With this approach, the IPC community can harness LLMs as “co-pilots” in infection control: the AI handles the mundane and provides preliminary analyses, while IPC teams remain the ultimate decision-makers ensuring patient safety. Going forward, we clearly can’t rest on our laurels – continued research, real-world evaluations, and interdisciplinary collaboration (between, e.g., AI developers, IPC experts, policy-makers, ethicists) will be crucial.

LLMs have arrived and, even if they are soon to be subsumed by more general artificial intelligence models, the linguistic component will remain front and centre. At every turn, however, it’s our prerogative to step back and question the co-pilot status of these “knowledge companions”:

Would you trust your second-in-command with the lives of so many?

References

  1. Rodriguez-Nava, G., Egoryan, G., Goodman, K.E., Morgan, D.J. and Salinas, J.L., 2025. Performance of a large language model for identifying central line-associated bloodstream infections (CLABSI) using real clinical notes. Infection Control & Hospital Epidemiology, 46(3), pp.305-308.

  2. Boussina, A., Krishnamoorthy, R., Quintero, K., Joshi, S., Wardi, G., Pour, H., Hilbert, N., Malhotra, A., Hogarth, M., Sitapati, A.M. and VanDenBerg, C., 2024. Large language models for more efficient reporting of hospital quality measures. NEJM AI, 1(11), p.AIcs2400420.

  3. Creagh, A.P., Pease, T., Ashworth, P., Bradley, L. and Duport, S., 2024. Explainable machine learning to identify patients at risk of developing hospital acquired infections. medRxiv, pp.2024-11.

  4. Wiemken, T.L. and Carrico, R.M., 2024. Assisting the infection preventionist: Use of artificial intelligence for health care–associated infection surveillance. American Journal of Infection Control, 52(6), pp.625-629.

  5. Schwartz, I.S., Link, K.E., Daneshjou, R. and Cortés-Penfield, N., 2024. Black box warning: large language models and the future of infectious diseases consultation. Clinical infectious diseases, 78(4), pp.860-866.

Want to see NEX in action?

We’ll show you how we can enhance your hospitals' infection prevention.

Want to see NEX in action?

Enhance your hospital's infection prevention capabilities.

Contact Us

London, United Kingdom

©2025 NEX Health Intelligence

Contact Us

London, United Kingdom

©2025 NEX Health Intelligence

Contact Us

London, United Kingdom

©2025 NEX Health Intelligence