LLMs hallucinate when removing patient info from EPR, finds study
AI tools sometimes produce hallucinations when asked to remove personal patient information from electronic patient records (EPRs), a study has found.
Researchers from the University of Oxford evaluated the ability of large language models (LLMs)…
