Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review

BackgroundThe rapid evolution of large language models (LLMs), such as Bidirectional Encoder Representations from Transformers (BERT; Google) and GPT (OpenAI), has introduced significant advancements in natural language processing. These models are increasingly integrated int...

Full description

Saved in:
Bibliographic Details
Main Authors: Dipak Gautam, Philipp Kellmeyer
Format: Article
Language:English
Published: JMIR Publications 2025-01-01
Series:JMIR Research Protocols
Online Access:https://www.researchprotocols.org/2025/1/e62865
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1832582363885862912
author Dipak Gautam
Philipp Kellmeyer
author_facet Dipak Gautam
Philipp Kellmeyer
author_sort Dipak Gautam
collection DOAJ
description BackgroundThe rapid evolution of large language models (LLMs), such as Bidirectional Encoder Representations from Transformers (BERT; Google) and GPT (OpenAI), has introduced significant advancements in natural language processing. These models are increasingly integrated into various applications, including mental health support. However, the credibility of LLMs in providing reliable and explainable mental health information and support remains underexplored. ObjectiveThis scoping review systematically maps the factors influencing the credibility of LLMs in mental health support, including reliability, explainability, and ethical considerations. The review is expected to offer critical insights for practitioners, researchers, and policy makers, guiding future research and policy development. These findings will contribute to the responsible integration of LLMs into mental health care, with a focus on maintaining ethical standards and user trust. MethodsThis review follows PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines and the Joanna Briggs Institute (JBI) methodology. Eligibility criteria include studies that apply transformer-based generative language models in mental health support, such as BERT and GPT. Sources include PsycINFO, MEDLINE via PubMed, Web of Science, IEEE Xplore, and ACM Digital Library. A systematic search of studies from 2019 onward will be conducted and updated until October 2024. Data will be synthesized qualitatively. The Population, Concept, and Context framework will guide the inclusion criteria. Two independent reviewers will screen and extract data, resolving discrepancies through discussion. Data will be synthesized and presented descriptively. ResultsAs of September 2024, this study is currently in progress, with the systematic search completed and the screening phase ongoing. We expect to complete data extraction by early November 2024 and synthesis by late November 2024. ConclusionsThis scoping review will map the current evidence on the credibility of LLMs in mental health support. It will identify factors influencing the reliability, explainability, and ethical considerations of these models, providing insights for practitioners, researchers, policy makers, and users. These findings will fill a critical gap in the literature and inform future research, practice, and policy development, ensuring the responsible integration of LLMs in mental health services. International Registered Report Identifier (IRRID)DERR1-10.2196/62865
format Article
id doaj-art-31aa4ea2f4f442d294182c7edf2357c0
institution Kabale University
issn 1929-0748
language English
publishDate 2025-01-01
publisher JMIR Publications
record_format Article
series JMIR Research Protocols
spelling doaj-art-31aa4ea2f4f442d294182c7edf2357c02025-01-29T21:00:30ZengJMIR PublicationsJMIR Research Protocols1929-07482025-01-0114e6286510.2196/62865Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping ReviewDipak Gautamhttps://orcid.org/0009-0008-9809-2199Philipp Kellmeyerhttps://orcid.org/0000-0001-5538-373X BackgroundThe rapid evolution of large language models (LLMs), such as Bidirectional Encoder Representations from Transformers (BERT; Google) and GPT (OpenAI), has introduced significant advancements in natural language processing. These models are increasingly integrated into various applications, including mental health support. However, the credibility of LLMs in providing reliable and explainable mental health information and support remains underexplored. ObjectiveThis scoping review systematically maps the factors influencing the credibility of LLMs in mental health support, including reliability, explainability, and ethical considerations. The review is expected to offer critical insights for practitioners, researchers, and policy makers, guiding future research and policy development. These findings will contribute to the responsible integration of LLMs into mental health care, with a focus on maintaining ethical standards and user trust. MethodsThis review follows PRISMA-ScR (Preferred Reporting Items for Systematic Reviews and Meta-Analyses extension for Scoping Reviews) guidelines and the Joanna Briggs Institute (JBI) methodology. Eligibility criteria include studies that apply transformer-based generative language models in mental health support, such as BERT and GPT. Sources include PsycINFO, MEDLINE via PubMed, Web of Science, IEEE Xplore, and ACM Digital Library. A systematic search of studies from 2019 onward will be conducted and updated until October 2024. Data will be synthesized qualitatively. The Population, Concept, and Context framework will guide the inclusion criteria. Two independent reviewers will screen and extract data, resolving discrepancies through discussion. Data will be synthesized and presented descriptively. ResultsAs of September 2024, this study is currently in progress, with the systematic search completed and the screening phase ongoing. We expect to complete data extraction by early November 2024 and synthesis by late November 2024. ConclusionsThis scoping review will map the current evidence on the credibility of LLMs in mental health support. It will identify factors influencing the reliability, explainability, and ethical considerations of these models, providing insights for practitioners, researchers, policy makers, and users. These findings will fill a critical gap in the literature and inform future research, practice, and policy development, ensuring the responsible integration of LLMs in mental health services. International Registered Report Identifier (IRRID)DERR1-10.2196/62865https://www.researchprotocols.org/2025/1/e62865
spellingShingle Dipak Gautam
Philipp Kellmeyer
Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review
JMIR Research Protocols
title Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review
title_full Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review
title_fullStr Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review
title_full_unstemmed Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review
title_short Exploring the Credibility of Large Language Models for Mental Health Support: Protocol for a Scoping Review
title_sort exploring the credibility of large language models for mental health support protocol for a scoping review
url https://www.researchprotocols.org/2025/1/e62865
work_keys_str_mv AT dipakgautam exploringthecredibilityoflargelanguagemodelsformentalhealthsupportprotocolforascopingreview
AT philippkellmeyer exploringthecredibilityoflargelanguagemodelsformentalhealthsupportprotocolforascopingreview