Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101

Abstract BackgroundIn recent years, large language models (LLMs) have shown a remarkable ability to generate human-like text. One potential application of this capability is using LLMs to simulate clients in a mental health context. This research presents the development and e...

Full description

Saved in:
Bibliographic Details
Main Authors: Daniel Cabrera Lozoya, Mike Conway, Edoardo Sebastiano De Duro, Simon D'Alfonso
Format: Article
Language:English
Published: JMIR Publications 2025-07-01
Series:JMIR Medical Education
Online Access:https://mededu.jmir.org/2025/1/e68056
Tags: Add Tag
No Tags, Be the first to tag this record!
_version_ 1849239188923219968
author Daniel Cabrera Lozoya
Mike Conway
Edoardo Sebastiano De Duro
Simon D'Alfonso
author_facet Daniel Cabrera Lozoya
Mike Conway
Edoardo Sebastiano De Duro
Simon D'Alfonso
author_sort Daniel Cabrera Lozoya
collection DOAJ
description Abstract BackgroundIn recent years, large language models (LLMs) have shown a remarkable ability to generate human-like text. One potential application of this capability is using LLMs to simulate clients in a mental health context. This research presents the development and evaluation of Client101, a web conversational platform featuring LLM-driven chatbots designed to simulate mental health clients. ObjectiveWe aim to develop and test a web-based conversational psychotherapy training tool designed to closely resemble clients with mental health issues. MethodsWe used GPT-4 and prompt engineering techniques to develop chatbots that simulate realistic client conversations. Two chatbots were created based on clinical vignette cases: one representing a person with depression and the other, a person with generalized anxiety disorder. A total of 16 mental health professionals were instructed to conduct single sessions with the chatbots using a cognitive behavioral therapy framework; a total of 15 sessions with the anxiety chatbot and 14 with the depression chatbot were completed. After each session, participants completed a 19-question survey assessing the chatbot’s ability to simulate the mental health condition and its potential as a training tool. Additionally, we used the LIWC (Linguistic Inquiry and Word Count) tool to analyze the psycholinguistic features of the chatbot conversations related to anxiety and depression. These features were compared to those in a set of webchat psychotherapy sessions with human clients—42 sessions related to anxiety and 47 related to depression—using an independent samples t ResultsParticipants’ survey responses were predominantly positive regarding the chatbots’ realism and portrayal of mental health conditions. For instance, 93% (14/15) considered that the chatbot provided a coherent and convincing narrative typical of someone with an anxiety condition. The statistical analysis of LIWC psycholinguistic features revealed significant differences between chatbot and human therapy transcripts for 3 of 8 anxiety-related features: negations (t56Pt56Pt56Pt60Pt60Pt60Pt60P ConclusionsThis research underscores both the strengths and limitations of using GPT-4-powered chatbots as tools for psychotherapy training. Participant feedback suggests that the chatbots effectively portray mental health conditions and are generally perceived as valuable training aids. However, differences in specific psycholinguistic features suggest targeted areas for enhancement, helping refine Client101’s effectiveness as a tool for training mental health professionals.
format Article
id doaj-art-ae40303dbd3d42fa82f210ad90335e20
institution Kabale University
issn 2369-3762
language English
publishDate 2025-07-01
publisher JMIR Publications
record_format Article
series JMIR Medical Education
spelling doaj-art-ae40303dbd3d42fa82f210ad90335e202025-08-20T04:01:09ZengJMIR PublicationsJMIR Medical Education2369-37622025-07-0111e68056e6805610.2196/68056Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101Daniel Cabrera Lozoyahttp://orcid.org/0009-0001-1501-7804Mike Conwayhttp://orcid.org/0000-0002-3209-8108Edoardo Sebastiano De Durohttp://orcid.org/0009-0009-9204-3203Simon D'Alfonsohttp://orcid.org/0000-0001-7407-8730 Abstract BackgroundIn recent years, large language models (LLMs) have shown a remarkable ability to generate human-like text. One potential application of this capability is using LLMs to simulate clients in a mental health context. This research presents the development and evaluation of Client101, a web conversational platform featuring LLM-driven chatbots designed to simulate mental health clients. ObjectiveWe aim to develop and test a web-based conversational psychotherapy training tool designed to closely resemble clients with mental health issues. MethodsWe used GPT-4 and prompt engineering techniques to develop chatbots that simulate realistic client conversations. Two chatbots were created based on clinical vignette cases: one representing a person with depression and the other, a person with generalized anxiety disorder. A total of 16 mental health professionals were instructed to conduct single sessions with the chatbots using a cognitive behavioral therapy framework; a total of 15 sessions with the anxiety chatbot and 14 with the depression chatbot were completed. After each session, participants completed a 19-question survey assessing the chatbot’s ability to simulate the mental health condition and its potential as a training tool. Additionally, we used the LIWC (Linguistic Inquiry and Word Count) tool to analyze the psycholinguistic features of the chatbot conversations related to anxiety and depression. These features were compared to those in a set of webchat psychotherapy sessions with human clients—42 sessions related to anxiety and 47 related to depression—using an independent samples t ResultsParticipants’ survey responses were predominantly positive regarding the chatbots’ realism and portrayal of mental health conditions. For instance, 93% (14/15) considered that the chatbot provided a coherent and convincing narrative typical of someone with an anxiety condition. The statistical analysis of LIWC psycholinguistic features revealed significant differences between chatbot and human therapy transcripts for 3 of 8 anxiety-related features: negations (t56Pt56Pt56Pt60Pt60Pt60Pt60P ConclusionsThis research underscores both the strengths and limitations of using GPT-4-powered chatbots as tools for psychotherapy training. Participant feedback suggests that the chatbots effectively portray mental health conditions and are generally perceived as valuable training aids. However, differences in specific psycholinguistic features suggest targeted areas for enhancement, helping refine Client101’s effectiveness as a tool for training mental health professionals.https://mededu.jmir.org/2025/1/e68056
spellingShingle Daniel Cabrera Lozoya
Mike Conway
Edoardo Sebastiano De Duro
Simon D'Alfonso
Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101
JMIR Medical Education
title Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101
title_full Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101
title_fullStr Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101
title_full_unstemmed Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101
title_short Leveraging Large Language Models for Simulated Psychotherapy Client Interactions: Development and Usability Study of Client101
title_sort leveraging large language models for simulated psychotherapy client interactions development and usability study of client101
url https://mededu.jmir.org/2025/1/e68056
work_keys_str_mv AT danielcabreralozoya leveraginglargelanguagemodelsforsimulatedpsychotherapyclientinteractionsdevelopmentandusabilitystudyofclient101
AT mikeconway leveraginglargelanguagemodelsforsimulatedpsychotherapyclientinteractionsdevelopmentandusabilitystudyofclient101
AT edoardosebastianodeduro leveraginglargelanguagemodelsforsimulatedpsychotherapyclientinteractionsdevelopmentandusabilitystudyofclient101
AT simondalfonso leveraginglargelanguagemodelsforsimulatedpsychotherapyclientinteractionsdevelopmentandusabilitystudyofclient101