As AI chatbots become more popular in mental healthcare, the American Medical Association is urging Congress to strengthen safeguards.
The organization sent letters to the Congressional Artificial Intelligence Caucus, the Congressional Digital Health Caucus and the Senate Artificial Intelligence Caucus. The letters follow numerous reports of AI chatbots encouraging suicide and self-harm among vulnerable populations.
Congress held hearings on the role of AI in mental health last year, which “emphasized several critical mental health concerns, including emotional dependency on AI systems, the potential distortion of reality through prolonged interaction with chatbots, and the current lack of consistent safety protocols,” the AMA said in the letters.
These hearings showed the need for “immediate attention” to ensure AI tools don’t harm those seeking mental health support, the letter added.
That said, the AMA acknowledged that AI tools could be potentially valuable in mental health care if used safely.
“Across the country, patients persistently struggle to access mental health care, either for reasons of access or affordability,” the AMA said. “Well-designed AI-enabled tools may serve as supportive resources that expand access to evidence-based information, facilitate early identification of mental health concerns, and connect individuals with appropriate clinical services. When developed and deployed within clear regulatory guardrails, these technologies have the potential to complement, not replace, clinicians and help mitigate persistent workforce shortages and other access issues.”
The AMA provided several recommendations for AI chatbot safeguards, including:
Improve transparency: Require chatbots to clearly disclose that users are communicating with AI, as well as ban systems from presenting themselves as licensed clinicians.
Create clear regulatory boundaries: Prevent chatbots from diagnosing or treating mental health conditions without the right regulatory review. The AMA calls on Congress to direct agencies to create a “modern, risk-based oversight framework and clarify when AI tools qualify as medical devices.”
Improve oversight: Require ongoing safety monitoring, reporting of adverse events and strict standards for technology used by children and adolescents.
Protect privacy and security: The AMA called for rigid data protection standards, such as limits on data collection and retention and clear user consent for data use.
Limit commercial use: Ban advertising on mental health chatbots, especially for minors.
“AI-enabled tools may help expand access to mental health resources and support innovation in health care delivery, but they lack consistent safeguards against serious risks, including emotional dependency, misinformation, and inadequate crisis response,” said Dr. John Whyte, AMA CEO, in a statement. “With thoughtful oversight and accountability, policymakers can support innovation and ensure technologies prioritize patient safety, strengthen public trust, and responsibly complement—not replace—clinical care.”
Photo: Witthaya Prasongsin, Getty Images