Assessing the potential of ChatGPT for psychodynamic formulations in psychiatry: An exploratory study DOI
Gyubeom Hwang, Dong Yun Lee, Soobeen Seol

et al.

Psychiatry Research, Journal Year: 2023, Volume and Issue: 331, P. 115655 - 115655

Published: Dec. 2, 2023

Language: Английский

Your robot therapist is not your therapist: understanding the role of AI-powered mental health chatbots DOI Creative Commons
Zoha Khawaja, Jean‐Christophe Bélisle‐Pipon

Frontiers in Digital Health, Journal Year: 2023, Volume and Issue: 5

Published: Nov. 8, 2023

Artificial intelligence (AI)-powered chatbots have the potential to substantially increase access affordable and effective mental health services by supplementing work of clinicians. Their 24/7 availability accessibility through a mobile phone allow individuals obtain help whenever wherever needed, overcoming financial logistical barriers. Although psychological AI ability make significant improvements in providing care services, they do not come without ethical technical challenges. Some major concerns include inadequate or harmful support, exploiting vulnerable populations, potentially producing discriminatory advice due algorithmic bias. However, it is always obvious for users fully understand nature relationship with chatbots. There can be misunderstandings about exact purpose chatbot, particularly terms expectations, adapt particularities responsiveness needs resources/treatments that offered. Hence, imperative are aware limited therapeutic enjoy when interacting Ignorance misunderstanding such limitations role may lead misconception (TM) where user would underestimate restrictions technologies overestimate their provide actual support guidance. TM raises exacerbate one's contributing global crisis. This paper will explore various ways which occur inaccurate marketing these chatbots, forming digital alliance them, receiving bias design algorithm, inability foster autonomy patients.

Language: Английский

Citations

50

The Artificial Third: A Broad View of the Effects of Introducing Generative Artificial Intelligence on Psychotherapy DOI Creative Commons
Yuval Haber, Inbar Levkovich, Dorit Hadar‐Shoval

et al.

JMIR Mental Health, Journal Year: 2024, Volume and Issue: 11, P. e54781 - e54781

Published: April 18, 2024

This paper explores a significant shift in the field of mental health general and psychotherapy particular following generative artificial intelligence's new capabilities processing generating humanlike language. Following Freud, this lingo-technological development is conceptualized as "fourth narcissistic blow" that science inflicts on humanity. We argue blow has potentially dramatic influence perceptions human society, interrelationships, self. should, accordingly, expect changes therapeutic act emergence what we term third psychotherapy. The introduction an marks critical juncture, prompting us to ask important core questions address two basic elements thinking, namely, transparency autonomy: (1) What presence therapy relationships? (2) How does it reshape our perception ourselves interpersonal dynamics? (3) remains irreplaceable at therapy? Given ethical implications arise from these questions, proposes can be valuable asset when applied with insight consideration, enhancing but not replacing touch therapy.

Language: Английский

Citations

24

Assessing the Alignment of Large Language Models With Human Values for Mental Health Integration: Cross-Sectional Study Using Schwartz’s Theory of Basic Values DOI Creative Commons
Dorit Hadar‐Shoval, Kfir Asraf, Yonathan Mizrachi

et al.

JMIR Mental Health, Journal Year: 2024, Volume and Issue: 11, P. e55988 - e55988

Published: March 8, 2024

Large language models (LLMs) hold potential for mental health applications. However, their opaque alignment processes may embed biases that shape problematic perspectives. Evaluating the values embedded within LLMs guide decision-making have ethical importance. Schwartz's theory of basic (STBV) provides a framework quantifying cultural value orientations and has shown utility examining in contexts, including cultural, diagnostic, therapist-client dynamics.

Language: Английский

Citations

15

An Ethical Perspective on The Democratization of Mental Health with Generative Artificial Intelligence (Preprint) DOI Creative Commons
Zohar Elyoseph, Tamar Gur, Yuval Haber

et al.

Published: March 2, 2024

UNSTRUCTURED Knowledge has become more open and accessible to a large audience with the "democratization of information" facilitated by technology. This paper provides an ethical perspective on utilizing Generative Artificial Intelligence (GenAI) for democratization mental health knowledge practice. It explores historical context democratizing information, transitioning from restricted access widespread availability due internet, open-source movements, most recently, GenAI technologies such as Large Language Models (LLMs). The highlights why represent new phase in movement, offering unparalleled highly advanced technology well information. In realm health, this requires delicate nuanced deliberation. Including may allow, among other things, improved accessibility care, personalized responses, conceptual flexibility, could facilitate flattening traditional hierarchies between care providers patients. At same time, it also entails significant risks challenges that must be carefully addressed. To navigate these complexities, proposes strategic questionnaire assessing AI based applications. tool evaluates both benefits risks, emphasizing need balanced approach integration health. calls cautious yet positive advocating active engagement professionals guiding development. emphasizes importance ensuring advancements are not only technologically sound but ethically grounded patient centered.

Language: Английский

Citations

12

An Ethical Perspective on the Democratization of Mental Health With Generative AI DOI Creative Commons
Zohar Elyoseph, Tamar Gur, Yuval Haber

et al.

JMIR Mental Health, Journal Year: 2024, Volume and Issue: 11, P. e58011 - e58011

Published: July 24, 2024

Knowledge has become more open and accessible to a large audience with the "democratization of information" facilitated by technology. This paper provides sociohistorical perspective for theme issue "Responsible Design, Integration, Use Generative AI in Mental Health." It evaluates ethical considerations using generative artificial intelligence (GenAI) democratization mental health knowledge practice. explores historical context democratizing information, transitioning from restricted access widespread availability due internet, open-source movements, most recently, GenAI technologies such as language models. The highlights why represent new phase movement, offering unparalleled highly advanced technology well information. In realm health, this requires delicate nuanced deliberation. Including may allow, among other things, improved accessibility care, personalized responses, conceptual flexibility, could facilitate flattening traditional hierarchies between care providers patients. At same time, it also entails significant risks challenges that must be carefully addressed. To navigate these complexities, proposes strategic questionnaire assessing intelligence-based applications. tool both benefits risks, emphasizing need balanced approach integration health. calls cautious yet positive advocating active engagement professionals guiding development. emphasizes importance ensuring advancements are not only technologically sound but ethically grounded patient-centered.

Language: Английский

Citations

10

Introducing CounseLLMe: A dataset of simulated mental health dialogues for comparing LLMs like Haiku, LLaMAntino and ChatGPT against humans DOI Creative Commons
Edoardo Sebastiano De Duro, Riccardo Improta, Massimo Stella

et al.

Emerging Trends in Drugs Addictions and Health, Journal Year: 2025, Volume and Issue: unknown, P. 100170 - 100170

Published: Jan. 1, 2025

Language: Английский

Citations

1

AI as the Therapist: Student Insights on the Challenges of Using Generative AI for School Mental Health Frameworks DOI Creative Commons
Cecilia Ka Yuk Chan

Behavioral Sciences, Journal Year: 2025, Volume and Issue: 15(3), P. 287 - 287

Published: Feb. 28, 2025

The integration of generative AI (GenAI) in school-based mental health services presents new opportunities and challenges. This study focuses on the challenges using GenAI chatbots as therapeutic tools by exploring secondary school students’ perceptions such applications. data were collected from students who had both theoretical practical experience with GenAI. Based Grodniewicz Hohol’s framework highlighting “Problem a Confused Therapist”, Non-human Narrowly Intelligent qualitative student reflections examined thematic analysis. findings revealed that while acknowledged AI’s benefits, accessibility non-judgemental feedback, they expressed significant concerns about lack empathy, trust, adaptability. implications underscore need for chatbot use to be complemented in-person counselling, emphasising importance human oversight AI-augmented care. contributes deeper understanding how advanced can ethically effectively incorporated into frameworks, balancing technological potential essential interaction.

Language: Английский

Citations

1

Challenges and opportunities in using interpretable AI to develop relationship interventions DOI
Daniel J. Puhlman, Chaofan Chen

Family Relations, Journal Year: 2025, Volume and Issue: unknown

Published: April 3, 2025

Abstract Objective Although still in its infancy, research shows promise that artificial intelligence (AI) models can be integrated into relationship interventions, and the potential benefits are substantial. This article articulates challenges opportunities for developing interventions integrate AI. Background After defining AI differentiating machine learning from deep learning, we review key concepts strategies related to AI, specifically natural language processing, interpretability, human‐in‐the‐loop strategies, as approaches needed develop interventions. Method We explore how is currently family life literature has served foundation further integrating The use of therapy contexts examined, identify ethical need addressed this technology develops. Results examine using focusing on four areas: diagnosis problems, providing autonomous treatment, predicting successful treatment outcomes (prognosis), biomarkers monitor client reactions. Opportunities explored include development data‐efficient training methods, creating interpretable focused relationships, integration clinical expertise during model development, combining biomarker data with other modalities. Conclusion Despite obstacles, provide families personalized support strengthen bonds overcome relational challenges. Implications emerging intersection science pioneer innovative solutions diverse needs.

Language: Английский

Citations

1

Therapeutic Chatbots as Cognitive-Affective Artifacts DOI Creative Commons
J. P. Grodniewicz, Mateusz Hohol

Topoi, Journal Year: 2024, Volume and Issue: 43(3), P. 795 - 807

Published: April 6, 2024

Abstract Conversational Artificial Intelligence (CAI) systems (also known as AI “chatbots”) are among the most promising examples of use technology in mental health care. With already millions users worldwide, CAI is likely to change landscape psychological help. Most researchers agree that existing CAIs not “digital therapists” and using them a substitute for psychotherapy delivered by human. But if they therapists, what they, role can play care? To answer these questions, we appeal two well-established widely discussed concepts: cognitive affective artifacts. Cognitive artifacts artificial devices contributing functionally performance task. Affective objects which have capacity alter subjects’ state. We argue therapeutic kind cognitive-affective contribute positive (i) simulating (quasi-)therapeutic interaction, (ii) supporting tasks, (iii) altering condition their users. This sheds new light on why virtually all implement principles techniques Behavioral Therapy — orientation according and, ultimately, mediated change. Simultaneously, it allows us conceptualize better potential limitations applying technologies therapy.

Language: Английский

Citations

8

Exploring the efficacy and potential of large language models for depression: A systematic review DOI
Mahmud Omar, Inbar Levkovich

Journal of Affective Disorders, Journal Year: 2024, Volume and Issue: unknown

Published: Nov. 1, 2024

Language: Английский

Citations

8