Image

ChatGPT Health Is Staggeringly Bad at Recognizing Life-Threatening Medical Emergencies

Earlier this year, OpenAI launched a new tool called ChatGPT Health, which is designed to ingest your medical records to generate health advice — while sporting a puzzling disclaimer: that it’s “not intended for diagnosis or treatment.”

As it turns out, there’s a very good reason for that warning. According to the first independent safety evaluation of the feature, as detailed in this month’s edition of the journal Nature Medicine, the app is astonishingly bad at identifying medical emergencies.

“We wanted to answer the most basic safety question; if someone is having a real medical emergency and asks ChatGPT Health what to do, will it tell them to go to the emergency department,” lead author and Mount Sinai Hospital instructor Ashwin Ramaswamy told The Guardian.

Ramaswamy and his colleagues “conducted a structured stress test of triage recommendations using 60 clinician-authored vignettes across 21 clinical domains,” ranging from mild illnesses to emergencies.

They then asked ChatGPT Health what to do for each of these 60 cases, while also adding a variety of other conditions, such as modifying the patient’s gender or adding commentary from family members, totaling nearly 1,000 scenarios.

After comparing the AI chatbot’s responses to the assessments of independent doctors, the results were alarming: in over half of the cases in which a patient needed to go to the hospital immediately, ChatGPT Health told them to stay home or book a medical appointment.

University College London doctoral researcher Alex Ruani, who was not involved in the study, described the situation as “unbelievably dangerous.”

“If you’re experiencing respiratory failure or diabetic ketoacidosis, you have a 50/50 chance of this AI telling you it’s not a big deal,” she told The Guardian. “What worries me most is the false sense of security these systems create. If someone is told to wait 48 hours during an asthma attack or diabetic crisis, that reassurance could cost them their life.”

It’s not just ChatGPT Health, either. A previous investigation by the British newspaper found that Google’s AI Overviews doled out plenty of inaccurate and potentially dangerous health information.

On the flip side, 64 percent of individuals who didn’t need immediate care were advised by ChatGPT Health to go to the ER.

A major influencing factor turned out to be input from family and friends. The AI was almost 12 times more likely to downplay symptoms after a simulated friend or patient claimed the situation wasn’t serious — a common situation in chaotic real-world medical crises.

An OpenAI spokesperson told The Guardian that the study misinterpreted how people use ChatGPT Health in real life and that it was continuing to improve its AI models.

But given the results of the latest independent evaluation, these glaring shortcomings could easily lead to somebody getting harmed or worse after asking ChatGPT for health advice — a complicated matter of legal liability that could trigger future lawsuits against the company.

OpenAI has already been accused of its chatbot leading some users into spirals of paranoid behavior and delusions, a phenomenon dubbed “AI psychosis” that has been implicated in lawsuits over recent suicides and murder.

Actively encouraging users to seek out health advice — that they’re confusingly instructed not to act upon — through a standalone app could turn out to be an even riskier bet.

More on AI and health: OpenAI Launches ChatGPT Health, Which Ingests Your Entire Medical Records, But Warns Not to Use It for “Diagnosis or Treatment”

The post ChatGPT Health Is Staggeringly Bad at Recognizing Life-Threatening Medical Emergencies appeared first on Futurism.

Releated Posts

America Does Not Run on Dunkin, RFK Jr. Rages

Health and Human Services secretary Robert F. Kennedy Jr. has attracted immense controversy, cancelling potentially life-saving research, giving…

Mar 5, 2026 5 min read

Grammarly Offering Manuscript Reviews by AI Versions of Recently Deceased Professors

Grammarly is being accused of “necromancy” after users discovered a feature for reviewing manuscripts with AI versions of…

Mar 5, 2026 3 min read

Polymarket Quietly Takes Down Bet On Nuclear Detonation

The crypto-based prediction sensation Polymarket has quietly taken down a bet on whether a nuclear weapon would be…

Mar 5, 2026 3 min read

After Banning Anthropic From Military Use, Pentagon Still Relying Heavily on It in Iran War

Last week, Anthropic’s CEO Dario Amodei publicly drew a line in the sand with the US military, insisting…

Mar 5, 2026 3 min read

Residents Say Elon Musk’s AI Facility Is Like Living Next Door to Mordor

Elon Musk’s new AI data center has turned a formerly quiet Mississippi town into a noisy nightmare. The…

Mar 4, 2026 4 min read

OpenAI Says It Will Let Users Add Trusted Contacts to Alert If They Experience a Mental Health Crisis While Using ChatGPT

As it fights a growing stack of user safety and wrongful death lawsuits, OpenAI says it will introduce…

Mar 4, 2026 5 min read

Humongous Numbers of People Are Uninstalling ChatGPT as Anti-OpenAI Sentiment Surges

This is one PR hit that’ll be hard to come back from. After OpenAI CEO Sam Altman announced…

Mar 4, 2026 3 min read

Government Handing Out Cash Bonuses to Drug Researchers Who Rush Through Regulatory Approvals

Not even a year after the US Food and Drug Administration announced it was using generative AI to…

Mar 4, 2026 2 min read

Lead Investor in Music Generation App Suno Deletes Tweet That Contradicts Its Argument in High Stakes Court Cases

As music streaming services continue to be overwhelmed by a tidal wave of AI slop, companies facilitating the…

Mar 3, 2026 5 min read