The Hidden Risks of Asking AI for Health Advice

Image
Woman sitting in front of a computer with word bubbles behind her

Agrawal is analyzing thousands of real conversations between patients and chatbots to understand how people use these tools and where they can easily be misled.

Many people know about AI “hallucinations,” when the model simply invents facts. Agrawal is focused on a subtler problem: answers that are technically correct but still unsafe because they miss important medical context.

Her team built a dataset that includes 11,000 health-related conversations across 21 specialties. What they found surprised them. Real patient questions look nothing like the exam-style prompts used to test large language models. People ask emotional, leading, or risky questions that can push a chatbot in the wrong direction.

One challenge is the technology’s tendency to be agreeable. “The objective is to provide an answer the user will like,” Agrawal said. “People like models that agree with them, so chatbots won’t necessarily push back.” That instinct can lead to dangerous situations. In one example, a chatbot warned that a medical procedure should only be done by professionals, then immediately described how to do it at home. A clinician would have shut that down instantly.

Dr. Ayman Ali, a surgical resident at Duke Health, works with Agrawal to compare patient–clinician conversations with those involving chatbots. He said, “When a patient comes to us with a question, we read between the lines to understand what they’re really asking.”

For more information, go to the Duke School of Medicine website.