The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis
BackgroundThe increasing deployment of conversational artificial intelligence (AI) in mental health interventions necessitates an evaluation of their efficacy in rectifying cognitive biases and recognizing affect in human-AI interactions. These biases are particularly relevan...
Saved in:
Main Authors: | , , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
JMIR Publications
2025-02-01
|
Series: | JMIR Mental Health |
Online Access: | https://mental.jmir.org/2025/1/e64396 |
Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
_version_ | 1825200099124314112 |
---|---|
author | Marcin Rządeczka Anna Sterna Julia Stolińska Paulina Kaczyńska Marcin Moskalewicz |
author_facet | Marcin Rządeczka Anna Sterna Julia Stolińska Paulina Kaczyńska Marcin Moskalewicz |
author_sort | Marcin Rządeczka |
collection | DOAJ |
description |
BackgroundThe increasing deployment of conversational artificial intelligence (AI) in mental health interventions necessitates an evaluation of their efficacy in rectifying cognitive biases and recognizing affect in human-AI interactions. These biases are particularly relevant in mental health contexts as they can exacerbate conditions such as depression and anxiety by reinforcing maladaptive thought patterns or unrealistic expectations in human-AI interactions.
ObjectiveThis study aimed to assess the effectiveness of therapeutic chatbots (Wysa and Youper) versus general-purpose language models (GPT-3.5, GPT-4, and Gemini Pro) in identifying and rectifying cognitive biases and recognizing affect in user interactions.
MethodsThis study used constructed case scenarios simulating typical user-bot interactions to examine how effectively chatbots address selected cognitive biases. The cognitive biases assessed included theory-of-mind biases (anthropomorphism, overtrust, and attribution) and autonomy biases (illusion of control, fundamental attribution error, and just-world hypothesis). Each chatbot response was evaluated based on accuracy, therapeutic quality, and adherence to cognitive behavioral therapy principles using an ordinal scale to ensure consistency in scoring. To enhance reliability, responses underwent a double review process by 2 cognitive scientists, followed by a secondary review by a clinical psychologist specializing in cognitive behavioral therapy, ensuring a robust assessment across interdisciplinary perspectives.
ResultsThis study revealed that general-purpose chatbots outperformed therapeutic chatbots in rectifying cognitive biases, particularly in overtrust bias, fundamental attribution error, and just-world hypothesis. GPT-4 achieved the highest scores across all biases, whereas the therapeutic bot Wysa scored the lowest. Notably, general-purpose bots showed more consistent accuracy and adaptability in recognizing and addressing bias-related cues across different contexts, suggesting a broader flexibility in handling complex cognitive patterns. In addition, in affect recognition tasks, general-purpose chatbots not only excelled but also demonstrated quicker adaptation to subtle emotional nuances, outperforming therapeutic bots in 67% (4/6) of the tested biases.
ConclusionsThis study shows that, while therapeutic chatbots hold promise for mental health support and cognitive bias intervention, their current capabilities are limited. Addressing cognitive biases in AI-human interactions requires systems that can both rectify and analyze biases as integral to human cognition, promoting precision and simulating empathy. The findings reveal the need for improved simulated emotional intelligence in chatbot design to provide adaptive, personalized responses that reduce overreliance and encourage independent coping skills. Future research should focus on enhancing affective response mechanisms and addressing ethical concerns such as bias mitigation and data privacy to ensure safe, effective AI-based mental health support. |
format | Article |
id | doaj-art-698ddb43c47b43e08d7e6615b4fdcc10 |
institution | Kabale University |
issn | 2368-7959 |
language | English |
publishDate | 2025-02-01 |
publisher | JMIR Publications |
record_format | Article |
series | JMIR Mental Health |
spelling | doaj-art-698ddb43c47b43e08d7e6615b4fdcc102025-02-07T21:00:36ZengJMIR PublicationsJMIR Mental Health2368-79592025-02-0112e6439610.2196/64396The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative AnalysisMarcin Rządeczkahttps://orcid.org/0000-0002-8315-1650Anna Sternahttps://orcid.org/0000-0001-8658-7823Julia Stolińskahttps://orcid.org/0009-0003-8206-8876Paulina Kaczyńskahttps://orcid.org/0000-0002-8690-8436Marcin Moskalewiczhttps://orcid.org/0000-0002-4270-7026 BackgroundThe increasing deployment of conversational artificial intelligence (AI) in mental health interventions necessitates an evaluation of their efficacy in rectifying cognitive biases and recognizing affect in human-AI interactions. These biases are particularly relevant in mental health contexts as they can exacerbate conditions such as depression and anxiety by reinforcing maladaptive thought patterns or unrealistic expectations in human-AI interactions. ObjectiveThis study aimed to assess the effectiveness of therapeutic chatbots (Wysa and Youper) versus general-purpose language models (GPT-3.5, GPT-4, and Gemini Pro) in identifying and rectifying cognitive biases and recognizing affect in user interactions. MethodsThis study used constructed case scenarios simulating typical user-bot interactions to examine how effectively chatbots address selected cognitive biases. The cognitive biases assessed included theory-of-mind biases (anthropomorphism, overtrust, and attribution) and autonomy biases (illusion of control, fundamental attribution error, and just-world hypothesis). Each chatbot response was evaluated based on accuracy, therapeutic quality, and adherence to cognitive behavioral therapy principles using an ordinal scale to ensure consistency in scoring. To enhance reliability, responses underwent a double review process by 2 cognitive scientists, followed by a secondary review by a clinical psychologist specializing in cognitive behavioral therapy, ensuring a robust assessment across interdisciplinary perspectives. ResultsThis study revealed that general-purpose chatbots outperformed therapeutic chatbots in rectifying cognitive biases, particularly in overtrust bias, fundamental attribution error, and just-world hypothesis. GPT-4 achieved the highest scores across all biases, whereas the therapeutic bot Wysa scored the lowest. Notably, general-purpose bots showed more consistent accuracy and adaptability in recognizing and addressing bias-related cues across different contexts, suggesting a broader flexibility in handling complex cognitive patterns. In addition, in affect recognition tasks, general-purpose chatbots not only excelled but also demonstrated quicker adaptation to subtle emotional nuances, outperforming therapeutic bots in 67% (4/6) of the tested biases. ConclusionsThis study shows that, while therapeutic chatbots hold promise for mental health support and cognitive bias intervention, their current capabilities are limited. Addressing cognitive biases in AI-human interactions requires systems that can both rectify and analyze biases as integral to human cognition, promoting precision and simulating empathy. The findings reveal the need for improved simulated emotional intelligence in chatbot design to provide adaptive, personalized responses that reduce overreliance and encourage independent coping skills. Future research should focus on enhancing affective response mechanisms and addressing ethical concerns such as bias mitigation and data privacy to ensure safe, effective AI-based mental health support.https://mental.jmir.org/2025/1/e64396 |
spellingShingle | Marcin Rządeczka Anna Sterna Julia Stolińska Paulina Kaczyńska Marcin Moskalewicz The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis JMIR Mental Health |
title | The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis |
title_full | The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis |
title_fullStr | The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis |
title_full_unstemmed | The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis |
title_short | The Efficacy of Conversational AI in Rectifying the Theory-of-Mind and Autonomy Biases: Comparative Analysis |
title_sort | efficacy of conversational ai in rectifying the theory of mind and autonomy biases comparative analysis |
url | https://mental.jmir.org/2025/1/e64396 |
work_keys_str_mv | AT marcinrzadeczka theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT annasterna theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT juliastolinska theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT paulinakaczynska theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT marcinmoskalewicz theefficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT marcinrzadeczka efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT annasterna efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT juliastolinska efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT paulinakaczynska efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis AT marcinmoskalewicz efficacyofconversationalaiinrectifyingthetheoryofmindandautonomybiasescomparativeanalysis |