Artificial Intelligence

Oxford Study Warns AI Chatbots Are Unsafe for Medical Advice

OXFORD: AI chatbots pose risks to people seeking medical advice despite excelling at standardized medical knowledge tests, according to a study published in Nature Medicine by Oxford Internet Institute and Nuffield Department of Primary Care Health Sciences.

However, advancements in AI-powered medical diagnostics continue to show promise in specialized areas like prostate cancer detection.

The randomized trial involving 1,298 UK participants found those using LLMs, GPT-4o, Llama 3, and Command R+made no better medical decisions than control groups using internet searches or personal judgment.

Study co-author Dr. Rebecca Payne, a GP and Clarendon-Reuben Doctoral Scholar, stated “Despite all the hype, AI just isn’t ready to take on the role of physician. Patients need to be aware that asking a large language model about their symptoms can be dangerous, giving wrong diagnoses and failing to recognize when urgent help is needed.”

Participants assessed ten medical scenarios developed by doctors, ranging from severe headaches after nights out to new mothers feeling constantly exhausted. They identified potential conditions and recommended courses of action like visiting GPs or attending A&E.

Researchers identified three key challenges: users didn’t know what information LLMs needed, models provided vastly different answers to slight question variations, and users struggled to distinguish good from bad information when both appeared together. At the same time, AI interpreting MRI scans highlights how machine learning can perform effectively in structured medical imaging environments.

Lead author Andrew Bean noted “In this study, we show that interacting with humans poses a challenge even for top LLMs.”

Senior author Dr. Adam Mahdi emphasized, “The disconnect between benchmark scores and real-world performance should be a wake-up call. We cannot rely on standardized tests alone. AI systems need rigorous testing with diverse, real users.”

The research was supported by Prolific, Oxford’s AI Government and Policy Research Programme funded by Dieter Schwarz Stiftung, Royal Society, UKRI, and NIHR Oxford Biomedical Research Centre.

Amita Parul

Amita Parul is an Independent journalist with experience in reporting and commentary on current events and sociopolitical developments. She contributes original reporting and analysis that aligns with Tea4Tech’s editorial standards for accuracy, transparency, and context, focusing on business and technology trends. || Amita covers emerging news stories and provides explanatory insights that help readers understand both the events and their implications.

Recent Posts

MatX’s Series B Fuels a New Front in the AI Chip Wars

MOUNTAIN VIEW, Calif.: MatX has raised $500 million in Series B funding to scale its…

2 hours ago

Aidoc Raises $150M to Bring AI Radiology to Hospitals Worldwide

TEL AVIV: Aidoc has closed a $150 million Series E to expand its clinical AI…

4 hours ago

Ciridae Raises $20M Seed for Mid-Market AI Automation Platform

SAN FRANCISCO: Ciridae has raised $20 million in seed funding to deliver AI automation tools…

8 hours ago

Panthalassa Raises $140M for Wave-Powered AI Data Centers

OSLO: Panthalassa raises $140 million in Series B funding to build wave-powered floating data centers…

23 hours ago

David Silver’s Ineffable Lands $1.1B Seed to Build Superintelligence

LONDON: A London-based AI startup founded on the ambition of achieving "first contact with superintelligence"…

1 day ago

AI Insurtech Corgi Hits Unicorn Status with $160 Million Series B

NEW YORK: Corgi has raised $160 million in a Series B round, valuing the company…

2 days ago