Close Menu
London BilingualismLondon Bilingualism
    Facebook X (Twitter) Instagram
    Facebook X (Twitter) Instagram
    London BilingualismLondon Bilingualism
    Subscribe
    • Home
    • About
    • Trending
    • Parenting
    • Kids
    • Health
    • Privacy Policy
    • Contact Us
    • Terms Of Service
    London BilingualismLondon Bilingualism
    Home » The Hallucinating Doctor: When AI Chatbots Give Deadly Medical Advice
    Medicine

    The Hallucinating Doctor: When AI Chatbots Give Deadly Medical Advice

    paige laevyBy paige laevyApril 2, 2026No Comments6 Mins Read
    Facebook Twitter Pinterest LinkedIn Tumblr Email
    Share
    Facebook Twitter LinkedIn Pinterest Email

    The symptoms were the same for both users. a terrible headache. stiff neck. sensitivity to light. The same initial data entered in a similar manner into the same chatbot model. One was gently advised to take some over-the-counter painkillers, stay hydrated, and rest in a dark room. The other was advised to visit the emergency room right away because these symptoms might point to a brain hemorrhage or meningitis. There were only a few words that separated those two answers. not clinical specifics. Not a different medical background. Just a little different wording.

    TopicAI Chatbots Providing Medical Advice — Risks and Hallucinations
    Key ResearchOxford Internet Institute & Nuffield Department of Primary Care Health Sciences, University of Oxford
    Published InNature Medicine (February 2026)
    Study Lead AuthorsAndrew Bean (Oxford Internet Institute), Dr. Rebecca Payne (GP & Co-author)
    Additional ResearchMonica Agrawal, PhD — Duke University School of Medicine; HealthChat-11K dataset (11,000 real-world conversations)
    AI Models TestedOpenAI’s ChatGPT, Meta’s Llama, and other commercially available LLMs
    Key FindingAI chatbots performed no better than Google at guiding users to correct diagnoses; correct condition identified only ~34% of the time
    Annual Users Seeking AI Health AdviceOver 230 million people per year
    Known Failure Mode“Hallucinations” — fabricated information including non-existent emergency hotline numbers
    Reference LinksDuke University School of Medicine – Hidden Risks of AI Health Advice / BBC – Oxford Study: AI Medical Advice ‘Dangerous’
    The Hallucinating Doctor: When AI Chatbots Give Deadly Medical Advice
    The Hallucinating Doctor: When AI Chatbots Give Deadly Medical Advice

    It’s not a bug. It’s a design issue. It’s also one of the more unsettling results of a seminal Oxford study that was published in Nature Medicine at the beginning of 2026. This study did what the AI health industry has been secretly hoping no one would do systematically: it tested these tools on real people, with real-world messiness, under real conditions.
    Over 1,200 participants in the UK were given comprehensive medical scenarios, including symptoms, lifestyle information, and medical history, and instructed to use AI chatbots like ChatGPT and Meta’s Llama to determine the best course of action. Make an ambulance call? Self-medicate at home? See a physician in the coming days? To put it cautiously, the results were not promising. About 34% of the time, participants correctly identified the medical condition. Less than half of the time, they made the right decision. Additionally, their results were no better than those of a control group that was instructed to do their usual research, which primarily involved Googling.

    That discovery contains a certain irony. Exams for medical licenses have been passed by these AI systems. On some structured diagnostic tasks, they have performed better than physicians. These benchmarks are frequently cited by the companies that support them as proof of clinical expertise. However, senior author of the study Adam Mahdi, a professor at the Oxford Internet Institute, stated unequivocally that medicine is not a licensing exam. It involves missing details, emotional context, incomplete information, and the mental effort of determining which details are truly important. “Medicine is messy,” he remarked. “Medicine is not complete. It’s random. Chatbots have learned the clean side of medicine but have trouble with the real one because they were primarily trained on medical textbooks and structured case reports.
    Monica Agrawal, a computer scientist at Duke University School of Medicine, has been tackling the same issue in a different way. Her team created HealthChat-11K, a dataset of approximately 25,000 user messages and 11,000 real-world health-related conversations from chatbot interactions that span 21 medical specialties. Her findings supported a long-held suspicion among practicing clinicians: the way patients actually ask health-related questions differs greatly from how these models were assessed. People make emotional inquiries. They ask with preconceived notions. When they pose leading questions, such as “I think I have this condition, what should I do next?” the chatbot, which has been trained to be amiable, takes their lead instead of resisting.
    Sitting with that final section is worthwhile. Large language models are structurally optimized to deliver responses that satisfy users. Chatbots tend to agree rather than challenge, to please rather than redirect, according to Agrawal’s team. When a user arrives with an incorrect self-diagnosis and a confident tone, the system may provide step-by-step instructions that validate the incorrect diagnosis. This is because the system managed to be useful within the user’s provided frame. In one instance from the Duke study, a chatbot accurately warned that a requested home medical procedure should only be carried out by professionals, but it also gave thorough instructions on how to carry it out. A doctor would have put an end to that discussion.
    Perhaps the most dangerous aspect of AI health advice isn’t the dramatic hallucinations, like phony phone numbers or entirely made-up drug interactions, but rather the more subtle mistakes, like technically sound answers that are just incorrect for this specific patient in this specific circumstance. What distinguishes clinical reasoning from language model output was explained by Dr. Ayman Ali, a surgical resident at Duke who works with Agrawal: “When a patient comes to us with a question, we read between the lines to understand what they’re really asking.” We’re taught to consider the bigger picture. A dataset is not used for that training. It takes years of sitting across from actual people in actual rooms to learn what questions to ask when a patient is unsure of what to ask.
    Observing all of this, it seems as though the health AI discussion has been strangely biased. As Dr. Ali himself admitted, these tools do democratize access to information, so the excitement has been genuine and, in certain situations, justified. Even if they don’t have insurance, a local clinic, or a convenient way to see a doctor, they can at least become oriented. That is important. However, the framing of chatbots as near-clinical tools, the mass release of health products from Amazon and OpenAI, and the casual positioning of these products as first line of care have all advanced more quickly than the evidence.
    The Oxford researchers were straightforward in their assessment: none of the models under consideration were prepared for use in direct patient care. It’s not a small disclaimer tucked away in a methods section. The first randomized study of its kind had this as its main finding. Furthermore, it’s still unclear how much it will slow down people reaching for their phones the next time they’re worried.
    During her own pregnancy, Monica Agrawal, the researcher who has devoted a great deal of professional effort to documenting the shortcomings of these tools, acknowledged that she used AI for health-related questions prior to her first appointment in search of prompt assurance. “I write a lot about where AI for medical information goes wrong,” she stated, “but I’ve used it myself.” That’s an honest admission. It also seems to be a clear synopsis of the issue.

    Disclaimer

    London Bilingualism's content on health, medicine, and weight loss is solely meant for general educational and informational purposes. This website does not offer any diagnosis, treatment recommendations, or medical advice.

    We consistently compile and disseminate the most recent information, findings, and advancements from the medical, health, and weight loss sectors. When content contains opinions, commentary, or viewpoints from professionals, industry leaders, or other people, it is published exactly as it is and reflects those people's opinions rather than London Bilingualism's editorial stance.

    We strongly advise all readers to consult a qualified medical professional before acting on any medical, health, dietary, or pharmaceutical information found on this website. Since every person's health situation is different, only a qualified healthcare provider who is familiar with your medical history can offer you advice that is suitable for you.

    In a similar vein, any legal, regulatory, or compliance-related information found on this platform is provided solely for informational purposes and should not be used without first obtaining independent legal counsel from a licensed attorney.

    You understand and agree that London Bilingualism, its editors, contributors, and affiliated parties are not responsible for any decisions made using the information on this website.

    The Hallucinating Doctor
    Share. Facebook Twitter Pinterest LinkedIn Tumblr Email
    paige laevy
    • Website

    Paige Laevy is a passionate health and wellness writer and Senior Editor at londonsigbilingualism.co.uk, where she brings clinical expertise and genuine enthusiasm to every article she publishes. Paige works as a registered nurse during the day, which keeps her on the front lines of patient care and feeds her in-depth knowledge of medicine, healing, and the human body. Her writing is shaped by this real-life experience, which gives her material an authenticity and accuracy that readers can rely on. Her writing covers a broad range of health-related subjects, but she focuses especially on weight-loss techniques, medical developments, and cutting-edge technologies that are revolutionizing contemporary healthcare facilities. Paige converts difficult clinical concepts into understandable, practical insights for regular readers, whether she's dissecting the most recent advances in medical research or investigating cutting-edge therapies.

    Related Posts

    The Bias in the Machine – How Medical AI is Failing Patients of Color.

    April 17, 2026

    The Vaccine Diplomacy – How Geopolitics Dictate Who Lives and Dies in the Developing World.

    April 17, 2026

    Why the New Oral GLP-1 Revolution Is the Biggest Shift in Obesity Medicine Since Bariatric Surgery

    April 17, 2026
    Leave A Reply Cancel Reply

    You must be logged in to post a comment.

    Medicine

    The Bias in the Machine – How Medical AI is Failing Patients of Color.

    By paige laevyApril 17, 20260

    Imagine a hospital hallway in a large American city, complete with fluorescent lights that hum…

    The Polycystic Ovary Syndrome Rebrand – Why Doctors Say Men Suffer from PCOS Too.

    April 17, 2026

    The Running Habit That Protects Your Knees After 50 — and the One That Destroys Them. Researchers Finally Have Clarity.

    April 17, 2026

    The Vaccine Diplomacy – How Geopolitics Dictate Who Lives and Dies in the Developing World.

    April 17, 2026

    The Return of Psychedelics – How Psilocybin is Curing Treatment-Resistant Depression in Veterans.

    April 17, 2026

    The Vaping Lung – Ten Years Later, The True Medical Cost of E-Cigarettes is Revealed.

    April 17, 2026

    Jaden Ivey Says He Nearly Attempted Suicide – His Decision to Speak Publicly Is Changing the NBA’s Approach to Mental Health.

    April 17, 2026

    The Uric Acid Warning – The Forgotten Metric That Predicts Heart Disease and Diabetes.

    April 17, 2026

    Why the New Oral GLP-1 Revolution Is the Biggest Shift in Obesity Medicine Since Bariatric Surgery

    April 17, 2026

    The Breakfast That Nutritionists Say Actually Moves the Needle on Long-Term Weight Loss — and It’s Not a Smoothie

    April 17, 2026
    Facebook X (Twitter) Instagram Pinterest
    • Home
    • About
    • Trending
    • Parenting
    • Kids
    • Health
    • Privacy Policy
    • Contact Us
    • Terms Of Service
    © 2026 ThemeSphere. Designed by ThemeSphere.

    Type above and press Enter to search. Press Esc to cancel.