Use of large language models to identify pseudo-information: Implications for health information

利用大型语言模型识别伪信息:对健康信息的启示

阅读:2

Abstract

BACKGROUND: Open-access scientific research is an essential source of health-related information and self-education. Artificial intelligence-based large language models (LMMs) may be used to identify erroneous health information. OBJECTIVE: To investigate to what extent LMMs can be used to identify pseudo-information. METHODS: Four common LMM applications (ChatGPT-4o, Claude 3.5 Sonnet, Gemini and Copilot) were used to investigate their capability to indicate erroneous information provided in an open-access article. RESULTS: Initially, ChatGPT-4o and Claude were able to mark the provided article as an unreliable information source, identifying most of the inaccuracy problems. The assessments provided by Gemini and Copilot were inaccurate, as several critical aspects were not identified or were misinterpreted. During the validation phase, the initially accurate assessment of ChatGPT-4o was not reproducible, and only Claude was able to detect several critical issues in this phase. The verdicts of Copilot and Gemini remained largely unaltered. DISCUSSION: Large heterogeneity exists between LMMs in identifying inaccurate pseudo-information. Replication in LMM output may constitute a significant hurdle in their application. CONCLUSION: The accuracy of LMMs needs to be further improved until they can be reliably used by patients for health-related online information and as assistant tools for health information and library services workers without restriction.

特别声明

1、本页面内容包含部分的内容是基于公开信息的合理引用;引用内容仅为补充信息,不代表本站立场。

2、若认为本页面引用内容涉及侵权,请及时与本站联系,我们将第一时间处理。

3、其他媒体/个人如需使用本页面原创内容,需注明“来源:[生知库]”并获得授权;使用引用内容的,需自行联系原作者获得许可。

4、投稿及合作请联系:info@biocloudy.com。