Enhancing privacy-preserving deployable large language models for perioperative complication detection: a targeted strategy with LoRA fine-tuning

增强用于围手术期并发症检测的可部署的隐私保护型大型语言模型:一种基于 LoRA 微调的针对性策略

阅读:1

Abstract

Perioperative complications are a major global concern, yet manual detection suffers from 27% under‑reporting and frequent misclassification. Clinical LLM deployment is constrained by data sovereignty, compute cost, and limited locally deployable model performance. We show targeted prompt engineering plus Low‑Rank Adaptation (LoRA) fine‑tuning converts smaller open‑source LLMs into expert‑level diagnostic tools. In dual‑center validation, we built a framework simultaneously identifying and grading 22 complication severities. State‑of‑the‑art models outperformed human experts; Chain‑of‑Thought prompting significantly improved general models (p < 0.001) while preserving reasoning models' performance. Across documentation length quartiles, AI models maintained F1 > 0.64, whereas human performance declined from 0.73 to 0.45, demonstrating superior robustness to documentation complexity. Our targeted strategy-decomposing detection into focused single‑complication assessments-improved small models, with further gains from LoRA. On external validation (Center 2), the optimized 4B model's micro‑F1 rose from 0.28 to 0.64, approaching human experts (F1 = 0.69), driven by the targeted strategy (ΔF1 = 0.256, 95% CI [0.181, 0.336]) and LoRA (ΔF1 = 0.103, 95% CI [0.023, 0.186]). Concurrently, the 8B model surpassed human experts (F1 > 0.70). Optimized small models enable expert‑level accuracy with local deployment and preserved data sovereignty, offering a practical path for resource‑limited healthcare.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。