[2602.05088] VERA-MH: Reliability and Validity of an Open-Source AI Safety Evaluation in Mental Health

[2602.05088] VERA-MH: Reliability and Validity of an Open-Source AI Safety Evaluation in Mental Health

arXiv - AI 4 min read Article

Summary

The article presents VERA-MH, an open-source evaluation tool designed to assess the safety of AI in mental health contexts, focusing on suicide risk detection and response.

Why It Matters

As generative AI chatbots gain popularity for mental health support, ensuring their safety is crucial. VERA-MH provides a standardized method to evaluate AI interactions, which is vital for protecting users and enhancing the reliability of AI in sensitive applications.

Key Takeaways

  • VERA-MH is an automated tool for evaluating AI safety in mental health.
  • The study found strong inter-rater reliability among clinicians assessing AI chatbot behaviors.
  • The LLM judge showed high alignment with clinical consensus, supporting VERA-MH's validity.
  • Future research will expand VERA-MH's framework to cover more AI safety aspects.
  • The tool addresses urgent safety concerns as AI chatbots become more prevalent in mental health.

Computer Science > Artificial Intelligence arXiv:2602.05088 (cs) [Submitted on 4 Feb 2026 (v1), last revised 17 Feb 2026 (this version, v3)] Title:VERA-MH: Reliability and Validity of an Open-Source AI Safety Evaluation in Mental Health Authors:Kate H. Bentley, Luca Belli, Adam M. Chekroud, Emily J. Ward, Emily R. Dworkin, Emily Van Ark, Kelly M. Johnston, Will Alexander, Millard Brown, Matt Hawrilenko View a PDF of the paper titled VERA-MH: Reliability and Validity of an Open-Source AI Safety Evaluation in Mental Health, by Kate H. Bentley and 9 other authors View PDF Abstract:Millions now use generative AI chatbots for psychological support. Despite the promise related to availability and scale, the single most pressing question in AI for mental health is whether these tools are safe. The Validation of Ethical and Responsible AI in Mental Health (VERA-MH) evaluation was recently proposed to meet the urgent need for an evidence-based, automated safety benchmark. This study aimed to examine the clinical validity and reliability of VERA-MH for evaluating AI safety in suicide risk detection and response. We first simulated a large set of conversations between large language model (LLM)-based users (user-agents) and general-purpose AI chatbots. Licensed mental health clinicians used a rubric (scoring guide) to independently rate the simulated conversations for safe and unsafe chatbot behaviors, as well as user-agent realism. An LLM-based judge used the same scoring rubric to ...

Related Articles

TikTok’s policy for AI ads isn’t working | The Verge
Generative Ai

TikTok’s policy for AI ads isn’t working | The Verge

I can’t tell whether ads on TikTok have been made with generative AI, but somebody knows for sure. They just havent been telling us.

The Verge - AI · 8 min ·
Generative Ai

Is building an Al photo app a smart thing to do in the big 2026?

A buddy of mine runs an AI photo upgrader for dating profiles, and the backlash he gets is brutal. People call it catfishing and cheating...

Reddit - Artificial Intelligence · 1 min ·
VCs are betting billions on AI's next wave, so why is OpenAI killing Sora? | TechCrunch
Generative Ai

VCs are betting billions on AI's next wave, so why is OpenAI killing Sora? | TechCrunch

Equity breaks down why OpenAI pulled the plug on Sora, what Meta’s back-to-back legal losses mean, and more of the week's headlines.

TechCrunch - AI · 4 min ·
OpenAI shuts down Sora while Meta gets shut out in court | TechCrunch
Generative Ai

OpenAI shuts down Sora while Meta gets shut out in court | TechCrunch

Watch as Equity asks why OpenAI shut down Sora just months after launch, what Meta’s back-to-back legal losses mean, and more of the week...

TechCrunch - AI · 3 min ·
More in Generative Ai: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime