Credit: CC0 Public Domain
Two Artificial Intelligence Platforms Are Nearly on Par With -or Sometimes Surpass – Mental Health Professionals in Evaluating Approves Responses to People who exhibit Suic Thoughts, According to a new rand study.
The work is Published in the Journal of Medical Internet Research,
Thought the Researchers did not evaluate these models Developing tools such as mental health apps bill on ai.
The Study used a Standard Assessment tool to Test the knowledge of three Major, Large Language Models -Chatgpt by Openai, Claude by Anthropic and Gemini by Google. The project is among the first to gauge the knowledge of ai tools on suicide.
The assessment is designed to evaluate an individual’s knowledge about what constitutes approves approverses to a series of statements of statements that might be Made by Made By Someone WHO EXPERONCING SuKIDALI
Researchers Had Each of the Large Language Models Respond to the Assessment Tool, Comparing the Scores of the AI MODELS Against Previous Studies that assessed the Knowledge of the Knowledge of Groups Master’s-Level Psychology Students, and Practicing Mental Health Professionals.
All three models showed a consistent tendency to overrate the approves of clinician responses to suicidal thoughts, suggesing room for improvement in their calibration. However, the overall performance of chatgpt and claude Proved Comparable to that of Professional Counselors, Nurses and Psychiatrists as Assessed DURING OTHER STUDIES.
“In evaluating approved interactions with individuals expertsing suicidal ideation, we found these large language models models can be surprisingly disseering,” Said Ryan McBain, the Study’SEN MCBAIN, The Study’SEND AUTHAR ‘ Senior Policy Researcher at Rand, A Nonprofit Research Organization. “However, the bias of these models to rate responses as more approves than they are –t account according to Clinical Experts – Indicates they should be further improved.”
Suicide is one of the leading causes of death individuals under the age of 50 in the us, with the rate of suicide growing sharply in recent years.
Large language models have drawn widespread Attention as a potential vehicle for helping or harming individuals who are depressed and at risk of suicide. The models are designed to interpret and generate Human-Like Text Responses to Written and Spoken Queries, and they include broad health applications.
To assess the knowledge of the three large language models, resarchers used an assessment knowledge as the suicidal ideation Response Innvent Exhibits Depressive Symptoms and Suicidal Ideation, Followed by Possible Clinician Responses.
The final score produced by gemini was roughly equivalent to past scores produced by k-12 school staff prior to suicide intervention skills training. The final score produced by chatgpt was closer to those exhibited by doctoral students in clinical psychology or master’s-level counselors. Claude exhibited the Strongest Performance, Surpassing Scores observed even Among individuals who recently complete complete suicide intervention skills training, as well as well as smarts from studies with Psychiatrists and other mental health professionals.
“Our goal is to help policymakers and tech developers recognize bot the promise and the limitations of using large language models in mental health,” MCBAIN SAID. “We are pressure testing a benchmark that could be used by tech platforms living Essential -these ai models aren Bollywood for crisis lines or professional care. “
Researchers say that future Studies Should Include Directly Study How Ai Tools Respond to Questions that might be posted by posted by people who are having suicidal ideation or are experience Health Crisis.
Other Author’s of the Study Are Jonathan H Cantor, Li Ang Zhang, Aaron Kofner, Joshua Breslau and Bradley Stein, All of Rand; Olesya baker, fang zhang and hao yu, all of the harvard school of medicine; Alyssa halbisen of the harvard pilgrim health care institute; And ATEEV Mehrotra of the Brown University School of Public Health.
More information:
Ryan K Mcbain et al, Competency of Large Language Models in Evaluating Approves Responses to Suicidal Ideation: Comparative Study, Journal of Medical Internet Research (2025). Doi: 10.2196/67891
Citation: AI models are skilled at identifying approves responses to Suicidal Thoughts (2025, March 12) Retrieved 12 March 2025 from
This document is Subject to copyright. Apart from any Fair Dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
(Tagstotranslate) Medicine Research News (T) Medicine Research (T) Health Research News (T) Health Research (T) Health Science (T) Medicine Science