Impact of Large Language Model–Based AI Tools on Physician–Patient Communication: A Systematic Review and Meta-Analysis (Preprint) DOI
Sven Richter,

Clara Buszello,

Markus Prem

et al.

Published: May 11, 2025

BACKGROUND Recent advances in large language models (LLMs) such as GPT-3/4 have spurred development of AI chatbots and advisory tools medicine. These systems are posited to assist or augment physician–patient communication, potentially improving empathy, clarity, responsiveness. However, their actual impact on communication outcomes remains uncertain. OBJECTIVE To systematically review meta-analyze peer-reviewed studies (2020–2025) evaluating how LLM-based interventions affect including trust, patient understanding. METHODS Following PRISMA 2020 guidelines, we searched PubMed/MEDLINE for published from 2025 examining LLM chatbot applications clinical contexts. Eligible designs included randomized, observational, cross-sectional, qualitative studies. Two reviewers independently screened titles/abstracts, assessed full texts, extracted data study design, population, type, measures, outcomes. We conducted a synthesis random-effects meta-analysis, reporting pooled standardized mean differences (SMD) odds ratios (OR) with 95% confidence intervals (CI). RESULTS From 312 records, 10 (N=10) were included, all quantitative predominantly cross-sectional. Populations ranged patients chronic conditions healthcare professionals laypersons. Outcomes empathy (7 studies), clarity/information quality (6), satisfaction usefulness (4), trust perceptions (2). In six direct comparisons AI- versus physician-generated responses, LLMs rated significantly higher five One found replies judged empathetic 45.1% cases 4.6% physician (OR ~9.8, P<.001). Similarly, ChatGPT-4 answers scored 5-point scale than human-written responses (mean 4.18 vs 2.70, neurology showed scores (CARE +1.38, P<.01) ChatGPT answers. Only one no significant difference. content was also longer more information-rich, patient-perceived clarity On the other hand, GPT-4 simplified pathology reports, increasing comprehension (7.98 5.23/10, P<.001) reducing consultation time by 70%. sometimes less concise readable low-literacy patients. analyses (4 studies, n=2,604), positive effect (SMD +1.05, CI 0.45–1.65) improved understanding +0.82, 0.30–1.34). Patient results mixed. No directly long-term trust. CONCLUSIONS Current evidence suggests can enhance producing empathetic, detailed, understandable responses. improvements may positively influence experience engagement. generate overly lengthy occasionally inaccurate advice, emphasizing need oversight. While meta-analytic findings promising, robust controlled trials needed confirm benefits, assess outcomes, define optimal integration strategies.

Language: Английский

Cognitive vs. emotional empathy: exploring their impact on user outcomes in health-assistant chatbots DOI
Tingting Jiang, Chuanhe Huang, Yanrun Xu

et al.

Behaviour and Information Technology, Journal Year: 2025, Volume and Issue: unknown, P. 1 - 16

Published: March 6, 2025

Language: Английский

Citations

0

Artificial intelligence and psychoanalysis: is it time for psychoanalyst.AI? DOI Creative Commons
Thomas Rabeyron

Frontiers in Psychiatry, Journal Year: 2025, Volume and Issue: 16

Published: April 7, 2025

The current development of artificial intelligences (AI) is leading to major transformations within society. In this context, we observe how some these AIs are spontaneously used by individuals as confidants, and even romantic partners. emergence such relationships with raises questions about their integration in psychiatry the possibility developing "digital therapists". regard, highlight four key elements (accessibility availability; confidentiality; knowledge; memory) compare what an AI offers comparison a human therapist. We also discuss results studies that have already investigated use psychotherapy, particularly fields depression anxiety. then propose reflect more specifically on creating "psychoanalyst.AI," which leads us examine therapeutic relationship (transference, free association, play, dreams, reflexivity, narrativity) AI. conclusion, offer reflections relevance considering "therapeutic artifact," while taking into account ethical issues raised settings.

Language: Английский

Citations

0

ChatGPT's role in sleep health: Informative or misleading DOI
Mila Yunita, Palasara Brahmani Laras,

Fiki Prayogi

et al.

Sleep Health, Journal Year: 2025, Volume and Issue: unknown

Published: April 1, 2025

Language: Английский

Citations

0

Leveraging AI-Generated Emotional Self-Voice to Nudge People towards their Ideal Selves DOI
Cathy Mengying Fang, Phoebe Chua, Samantha Chan

et al.

Published: April 24, 2025

Language: Английский

Citations

0

Impact of Large Language Model–Based AI Tools on Physician–Patient Communication: A Systematic Review and Meta-Analysis (Preprint) DOI
Sven Richter,

Clara Buszello,

Markus Prem

et al.

Published: May 11, 2025

BACKGROUND Recent advances in large language models (LLMs) such as GPT-3/4 have spurred development of AI chatbots and advisory tools medicine. These systems are posited to assist or augment physician–patient communication, potentially improving empathy, clarity, responsiveness. However, their actual impact on communication outcomes remains uncertain. OBJECTIVE To systematically review meta-analyze peer-reviewed studies (2020–2025) evaluating how LLM-based interventions affect including trust, patient understanding. METHODS Following PRISMA 2020 guidelines, we searched PubMed/MEDLINE for published from 2025 examining LLM chatbot applications clinical contexts. Eligible designs included randomized, observational, cross-sectional, qualitative studies. Two reviewers independently screened titles/abstracts, assessed full texts, extracted data study design, population, type, measures, outcomes. We conducted a synthesis random-effects meta-analysis, reporting pooled standardized mean differences (SMD) odds ratios (OR) with 95% confidence intervals (CI). RESULTS From 312 records, 10 (N=10) were included, all quantitative predominantly cross-sectional. Populations ranged patients chronic conditions healthcare professionals laypersons. Outcomes empathy (7 studies), clarity/information quality (6), satisfaction usefulness (4), trust perceptions (2). In six direct comparisons AI- versus physician-generated responses, LLMs rated significantly higher five One found replies judged empathetic 45.1% cases 4.6% physician (OR ~9.8, P<.001). Similarly, ChatGPT-4 answers scored 5-point scale than human-written responses (mean 4.18 vs 2.70, neurology showed scores (CARE +1.38, P<.01) ChatGPT answers. Only one no significant difference. content was also longer more information-rich, patient-perceived clarity On the other hand, GPT-4 simplified pathology reports, increasing comprehension (7.98 5.23/10, P<.001) reducing consultation time by 70%. sometimes less concise readable low-literacy patients. analyses (4 studies, n=2,604), positive effect (SMD +1.05, CI 0.45–1.65) improved understanding +0.82, 0.30–1.34). Patient results mixed. No directly long-term trust. CONCLUSIONS Current evidence suggests can enhance producing empathetic, detailed, understandable responses. improvements may positively influence experience engagement. generate overly lengthy occasionally inaccurate advice, emphasizing need oversight. While meta-analytic findings promising, robust controlled trials needed confirm benefits, assess outcomes, define optimal integration strategies.

Language: Английский

Citations

0