
AI chatbots sound utterly convincing when delivering health advice that’s dead wrong half the time, potentially steering millions into medical danger without a clue.
Story Snapshot
- 50% of AI responses to health questions are problematic, 20% highly so and harmful.
- Chatbots exude confidence that masks unreliability, fooling users effortlessly.
- One in four American adults rely on AI for health info, but only 58% check with doctors.
- Tested ChatGPT, Gemini, Grok, Meta AI, DeepSeek across 250 responses in five medical areas.
- BMJ Open peer-reviewed study warns of public health risks without oversight.
Study Reveals Alarming AI Inaccuracy in Health Queries
Researchers tested five major AI chatbots in February 2025, posing 10 questions each across five medical categories: vaccines, cancer, nutrition, athletic performance, and respiratory diseases. They evaluated 250 responses total. Two independent experts rated each for accuracy and harm potential. BMJ Group, tied to the British Medical Association, led this peer-reviewed effort published in BMJ Open in April 2026. Results exposed a stark reality: 49.6% of answers proved problematic.
Confidence Paradox Fuels Hidden Dangers
Chatbots delivered incorrect information with unwavering assertiveness. Lee Schwamm, MD from Yale School of Medicine, captured it: “Chatbots are sometimes wrong, but never in doubt.” Users struggle to spot flaws because responses lack caveats or uncertainty signals. Nearly 20%, or 19.6%, rated highly problematic, posing direct harm risk if followed. Lead researcher Nick Tiller stressed these could injure people who trust AI summaries blindly. Open-ended questions, common in real life, fared worst at 32% highly problematic versus 7% for yes/no types.
Performance Varies by Medical Topic
Vaccines and cancer questions yielded better results, reflecting abundant consensus data online. Nutrition and athletic performance domains struggled, mired in conflicting evidence. Grok performed consistently worst among ChatGPT, Gemini, Meta AI, and DeepSeek. Gemini edged ahead slightly. Researchers noted their adversarial testing might inflate issues, yet findings align with known AI hallucinations. This variability underscores why AI falters in nuanced health areas lacking clear facts.
Real-world data from Kaiser Family Foundation poll shows one in four American adults use AI for health advice. Alarmingly, only 58% follow up with doctors for physical health queries. Patients arrive in clinics armed with unvetted AI insights, forcing physicians to debunk misinformation. This burdens overworked systems and risks delayed care.
Stakeholders Face Mounting Pressure
AI firms like OpenAI, Google, Meta, xAI, and DeepSeek face no immediate rules on health outputs, holding deployment power. BMJ Group and doctors push for education and safeguards. Patients remain most vulnerable, craving quick answers without expertise to judge them. No company responses surfaced yet, but media buzz from ABC News, WJLA, and others amplifies calls for disclaimers or query refusals on medical topics.
Short-Term Risks Demand Immediate Action
Harm looms as users act on bad advice, from misguided diets to ignored symptoms. Doctors waste time correcting AI errors, straining practices. Awareness may dip casual use, but polls suggest habits persist. Long-term, expect regulatory pushes for oversight, eroding unchecked AI trust. Economic hits from poor outcomes could spur liability shifts. Health literacy gaps widen, hitting those least able to question AI hardest. Facts demand caution: verify with pros, not pixels.
Sources:
Poll shows 50% of AI health query answers are problematic
AI health answers are wrong even though they sound convincing
AI chatbots provide poor answers to medical questions half the time, study finds
Asking AI health questions? Use with caution, researchers say
Asking AI health questions? Use with caution, researchers say
Half of AI Health Answers Are Wrong Even Though They Sound Convincing: New Study













