How Our AI-assisted Qualitative Analysis Failed DOI Open Access
Luke Thominet, Kristine Acosta,

Jacqueline Amorim

et al.

Published: Oct. 20, 2024

The authors describe how their Generative Artificial Intelligence (GAI)-assisted qualitative research project failed to produce publishable results. Based on this experience, they argue for the value of embracing and reflecting failure in GAI-assisted research. To frame argument, draw two theories generative failure: failing forward, which uses failures iterate designs meet existing criteria, sideways, reconsiders criteria success. Using a fail-forward perspective, might revise methods data preparation, process documentation, task delegation create more reliable Then, using fail-sideways reexamine results reimagine study fundamentally.

Language: Английский

Can Large Language Models Serve as Data Analysts? A Multi-Agent Assisted Approach for Qualitative Data Analysis DOI
Zeeshan Rasheed, Muhammad Waseem, Aakash Ahmad

et al.

Published: Jan. 1, 2025

Language: Английский

Citations

3

Harnessing AI for Understanding Scientific Literature: Innovations and Applications of Chat-Agent System in Battery Recycling Research DOI

Rongfan Liu,

Zhi Zou,

Sihui Chen

et al.

Materials Today Energy, Journal Year: 2025, Volume and Issue: unknown, P. 101818 - 101818

Published: Jan. 1, 2025

Language: Английский

Citations

1

Deductively coding psychosocial autopsy interview data using a few-shot learning large language model DOI Creative Commons
Elias Balt, Salim Salmi, Sandjai Bhulai

et al.

Frontiers in Public Health, Journal Year: 2025, Volume and Issue: 13

Published: Feb. 19, 2025

Psychosocial autopsy is a retrospective study of suicide, aimed to identify emerging themes and psychosocial risk factors. It typically relies heavily on qualitative data from interviews or medical documentation. However, research has often been scrutinized for being prone bias notoriously time- cost-intensive. Therefore, the current investigate if Large Language Model (LLM) can be feasibly integrated with procedures, by evaluating performance model in deductively coding coherently summarizing interview obtained autopsy. Data 38 semi-structured conducted individuals bereaved suicide loved one was coded researchers server-installed LLAMA3 large language model. The evaluated three tasks: (1) binary classification segments, (2) independent using sliding window approach, (3) summarization data. Intercoder agreement scores were calculated Cohen's Kappa, LLM's summaries qualitatively assessed Constant Comparative Method. results showed that LLM achieved substantial (accuracy: 0.84) task 0.67). had variability across codes. rich enough subsequent analysis researcher, around 80% rated independently two as 'adequate' 'good.' Emerging assessment included unsolicited elaboration hallucination. State-of-the-art LLMs show great potential support complex data, which would alleviate investment time resources. Integrating models procedures facilitate near real-time monitoring. Based findings, we recommend collaborative model, whereby deductive complemented review, inductive further interpretation researcher. Future may aim replicate findings different contexts evaluate larger context size.

Language: Английский

Citations

0

Validating the use of large language models for psychological text classification DOI Creative Commons
Harry Bunt, Alex Goddard, Tom W. Reader

et al.

Frontiers in Social Psychology, Journal Year: 2025, Volume and Issue: 3

Published: Feb. 21, 2025

Large language models (LLMs) are being used to classify texts into categories informed by psychological theory (“psychological text classification”). However, the use of LLMs in classification requires validation, and it remains unclear exactly how psychologists should prompt validate for this purpose. To address gap, we examined potential using classification, focusing on ways ensure validity. We employed OpenAI's GPT-4o (1) reported speech online diaries, (2) other-initiations conversational repair Reddit dialogues, (3) harm healthcare complaints submitted NHS hospitals trusts. Employing a two-stage methodology, developed tested validity prompts instruct manually labeled data ( N = 1,500 each task). First, iteratively three types one-third coded dataset, examining their semantic validity, exploratory predictive content Second, performed confirmatory test final remaining two-thirds dataset. Our findings contribute literature demonstrating that can serve as valid coders phenomena text, condition researchers work with LLM secure semantic, predictive, They also demonstrate rapid cost-effective iterations over big qualitative datasets, enabling explore refine concepts operationalizations during manual coding classifier development. Accordingly, secondary contribution, enable an intellectual partnership researcher, defined synergistic recursive process where LLM's generative nature facilitates checks. argue may signify paradigm shift toward novel, iterative approach improve operationalizations.

Language: Английский

Citations

0

What Ethical Issues do ChatGPT Face: A Bibliometrics Based Study DOI Creative Commons
Bo Wang, Rozaini Binti Rosli

Published: March 17, 2025

ChatGPT represents a groundbreaking AI application that has garnered significant attention since its inception. However, despite promising potential, ethical implications have sparked considerable debate. This study aims to examine the key concerns surrounding governance of by conducting bibliometric analysis and cluster-based content relevant scientific literature. The identifies influential authors, countries, pivotal publications, revealing three primary categories issues associated with ChatGPT: human-related ethics, academic integrity technical literacy, artificial intelligence (AI) technology ethics derived concerns. Additionally, further refines these synthesizing frequently occurring keywords. Building on this framework, provides comprehensive discussion major challenges faced ChatGPT, as well outlining future research priorities. Furthermore, investigates knowledge base underlying ChatGPT's governance, exploring high-citation high-link-strength literature through co-citation analysis, thereby mapping landscape highlighting areas growing scholarly interest. offers valuable insights for policymakers, researchers, practitioners, emphasizing need more stringent policies, guidelines, robust design in development similar technologies.

Language: Английский

Citations

0

The Ethical Implications of Using AI in Qualitative Research DOI
Danielle Hitch, Kieva Richards, Adyya Gupta

et al.

CABI eBooks, Journal Year: 2025, Volume and Issue: unknown, P. 137 - 149

Published: April 8, 2025

Language: Английский

Citations

0

AI Augmented Qualitative Analysis: Is it the Right Choice for Your Study? DOI
Danielle Hitch, Kieva Richards,

Richard Knight

et al.

CABI eBooks, Journal Year: 2025, Volume and Issue: unknown, P. 59 - 70

Published: April 8, 2025

Language: Английский

Citations

0

A Guide for Qualitative Researchers Using Large Language Models with Representative Examples Using ChatGPT DOI

Ann Armstrong,

Albert J. Gale

CABI eBooks, Journal Year: 2025, Volume and Issue: unknown, P. 84 - 96

Published: April 8, 2025

Language: Английский

Citations

0

On the juggernaut of artificial intelligence in organizations, research and society DOI
Yves Gendron, Jane Andrew, Christine Cooper

et al.

Critical Perspectives on Accounting, Journal Year: 2024, Volume and Issue: 100, P. 102759 - 102759

Published: July 14, 2024

Language: Английский

Citations

3

Role Play: Conversational Roles as a Framework for Reflexive Practice in AI-Assisted Qualitative Research DOI
Luke Thominet,

Jacqueline Amorim,

Kristine Acosta

et al.

Journal of Technical Writing and Communication, Journal Year: 2024, Volume and Issue: unknown

Published: June 11, 2024

Previous literature has shown that generative artificial intelligence (GAI) software, including large language model (LLM) chatbots, might contribute to qualitative research studies. However, there is still a need examine the relationships between researchers, GAI technologies, data, and findings. To address this need, our team conducted thematic analysis of reflexive journals from an LLM chatbot-assisted project. We identified four roles researchers adopted: managers closely monitored LLM's work, teachers instructed on theories methods, colleagues openly discussed data with LLM, advocates worked improve user experiences. Planning for playing multiple also helped enrich process. This study underscores potential using conversational as framework support reflexivity when working technologies research.

Language: Английский

Citations

1