Abstract
BACKGROUND: Artificial intelligence (AI)-generated text to video is emerging in medical education, but its effectiveness, accuracy and safety remain uncertain. We aimed to synthesise empirical studies evaluating these tools in learner or patient education. METHODS: A comprehensive search was conducted in MEDLINE/PubMed, Google Scholar, Scopus, Cochrane Review and Web of Science for studies published up to January 2025. Eligible studies evaluated AI-generated text to video for medical or patient education, reporting both quantitative and qualitative outcomes. Two reviewers screened and extracted data. The review adhered to Preferred Reporting Items for Systematic Reviews and Meta-Analyses guidelines. RESULTS: Out of 103 identified studies, five studies met the inclusion criteria: four evaluated patient education and one evaluated physician training. Clinical areas were ophthalmology (2/5), plastic surgery (1/5), dysphagia rehabilitation (1/5) and neurosurgical training (1/5). In ophthalmology, control materials outperformed AI-generated videos on image/script accuracy (p<0.005), with similar script-image alignment. In dysphagia rehabilitation, a randomised trial reported improvements in swallowing function and related outcomes with an AI-assisted video game intervention (p<0.001). A plastic surgery study reported greater user preference for a video avatar tool compared with a text chatbot (63.5% vs 28.1%). Across the reviewed studies, samples were small and CIs were rarely reported. Outcome measures were heterogeneous, and meta-analysis was not feasible. CONCLUSION: AI-generated videos can enhance engagement or selected outcomes in certain contexts, yet concerns about accuracy and inconsistent measurement persist. Current evidence is sparse and mixed. Currently, these tools can complement, rather than replace, standard resources until non-inferiority is demonstrated for the prespecified primary outcomes. PROSPERO REGISTRATION NUMBER: CRD42025640042.