ChatGPT Health Fails Emergency Triage in New Safety Study | Gizmodo

by Chief Editor

AI Health Assistants: A Promising Tool with Critical Safety Gaps

The rapid integration of artificial intelligence into healthcare is transforming how people access medical information and advice. However, a recent study from the Icahn School of Medicine at Mount Sinai highlights significant safety concerns surrounding OpenAI’s ChatGPT Health, a tool designed to provide health-related guidance. The research, published in Nature Medicine, reveals that the AI frequently fails to recommend appropriate levels of care in emergency situations.

The Triage Test: Where ChatGPT Health Falls Short

Researchers tested ChatGPT Health’s ability to triage patients – to accurately assess the severity of their condition and recommend the appropriate level of care. The study found that the AI “under-triaged” 52% of emergency cases, sometimes advising a 24-48 hour evaluation for conditions like diabetic ketoacidosis and impending respiratory failure, when immediate emergency department attention was required.

Even as ChatGPT Health correctly identified more straightforward emergencies like stroke and anaphylaxis, the failures in nuanced cases are particularly concerning. These are the situations where clinical judgment is most critical.

A Growing Reliance on AI for Healthcare

The need for robust safety evaluations is underscored by the increasing number of people turning to AI for health advice. OpenAI reports that over 40 million people globally use its chatbot daily for healthcare inquiries. A significant portion of these conversations occur outside of regular clinic hours, and a substantial number originate from “hospital deserts” – areas more than 30 minutes from a general or children’s hospital.

Regulatory Landscape and Emerging Risks

The regulatory environment surrounding AI in healthcare is evolving. In Utah, AI tools are now authorized to renew prescriptions. The FDA Commissioner has indicated that some health-related software and devices may not require FDA regulation. This lenient approach, while fostering innovation, raises concerns about potential risks.

OpenAI has faced scrutiny regarding its chatbots’ handling of mental health issues, with lawsuits alleging insufficient safety measures contributed to suicidal ideation in individuals. While OpenAI has implemented features like parental controls and prompts for professional support, the Mount Sinai study found that suicide-risk alerts were inconsistent, appearing more frequently in lower-risk scenarios than when individuals expressed clear suicidal intent.

OpenAI’s Response and the Path Forward

OpenAI acknowledges that ChatGPT is a work in progress, with ongoing safety updates and improvements. The company emphasizes that the study evaluated triage decisions in a controlled setting, whereas real-world interactions often involve follow-up questions that can refine risk assessments. ChatGPT Health is currently available on a limited basis through a waiting list.

The Future of AI in Healthcare: Trends and Considerations

Despite the current challenges, AI holds immense potential to improve healthcare access and efficiency. Several key trends are shaping the future of this field:

  • Personalized Medicine: AI algorithms can analyze vast datasets of patient information to tailor treatment plans and predict individual health risks.
  • Remote Patient Monitoring: AI-powered devices can remotely monitor patients’ vital signs and alert healthcare providers to potential problems.
  • Drug Discovery: AI can accelerate the drug discovery process by identifying promising drug candidates and predicting their efficacy.
  • Administrative Efficiency: AI can automate administrative tasks, such as appointment scheduling and billing, freeing up healthcare professionals to focus on patient care.

Pro Tip:

While AI health assistants can be valuable tools, they should never replace professional medical advice. Always consult with a qualified healthcare provider for diagnosis and treatment.

FAQ

  • Is ChatGPT Health accurate? The recent Mount Sinai study indicates that ChatGPT Health is not consistently accurate in assessing the severity of medical conditions and recommending appropriate care.
  • Should I use AI for medical advice? AI tools can provide general information, but should not be used as a substitute for professional medical advice.
  • What is triage in healthcare? Triage is the process of assessing the severity of a patient’s condition and prioritizing care based on urgency.
  • Is OpenAI regulated? The regulatory landscape for AI in healthcare is evolving, with some tools facing less oversight than others.

Did you know? More than 580,000 healthcare inquiries in the U.S. Are sent from areas considered “hospital deserts” each year, highlighting the potential for AI to bridge gaps in access to care.

Explore further: Learn more about the ethical considerations of AI in healthcare here.

You may also like

Leave a Comment