Evaluating AI Excellence: A Comparative Analysis of Generative Models in Library and Information Science DOI

Raiyan Bin Reza,

Md. Rifat Mahmud, S.M. Zabed Ahmed

et al.

Science & Technology Libraries, Journal Year: 2024, Volume and Issue: unknown, P. 1 - 14

Published: Oct. 7, 2024

This study compares the performance of GPT-3.5, GPT-4, Bard, and Gemini in answering Library Information Science (LIS) questions. Sixteen questions were used for assessment, with two independent examiners scoring initial successive responses from each AI system. Statistical analyses, including one-way Analysis Variance (ANOVA), sample t-test, one-sample employed to identify differences. The results revealed consistency generated across iterations all systems. Significant differences observed among models, Bard consistently underperforming compared Gemini. uncovered variability examiners' emphasized need multiple evaluators assessment.

Language: Английский

ChatGPT-4o's Performance on Pediatric Vesicoureteral Reflux DOI
Esra Nagehan Akyol Önder, Esra Ensari, Pelin Ertan

et al.

Journal of Pediatric Urology, Journal Year: 2024, Volume and Issue: unknown

Published: Dec. 1, 2024

Language: Английский

Citations

3

Cognitive Domain Assessment of Artificial Intelligence Chatbots: A Comparative Study Between ChatGPT and Gemini’s Understanding of Anatomy Education DOI
Arthi Ganapathy, Parul Kaushal

Medical Science Educator, Journal Year: 2025, Volume and Issue: unknown

Published: Feb. 15, 2025

Language: Английский

Citations

0

Artificial Intelligence’s Performance on the Japanese National Dental Examination DOI Open Access

Tatsuya Akitomo,

Masakazu Hamada,

Yasuko Tsuge

et al.

Cureus, Journal Year: 2024, Volume and Issue: unknown

Published: Nov. 5, 2024

Background/purpose: Artificial intelligence (AI) has developed rapidly and is applied to many fields including dental education. In this study, we evaluated AI performance on the Japanese National Dental Examination. Materials methods: We extracted 349 of 400 compulsory questions from Examinations over past five years. Questions were presented ChatGPT 3.5, 4o mini, Gemini, their was across 13 topic categories. Results: mini achieved passing criteria for exams two years had highest total score three AIs. The scores "Society dentistry" "Cardinal signs" significantly higher than those 3.5 (P<0.05). Conclusions: high indicates potential value tool in Further improving its may lead future clinical applications dentistry.

Language: Английский

Citations

2

Evaluating AI Excellence: A Comparative Analysis of Generative Models in Library and Information Science DOI

Raiyan Bin Reza,

Md. Rifat Mahmud, S.M. Zabed Ahmed

et al.

Science & Technology Libraries, Journal Year: 2024, Volume and Issue: unknown, P. 1 - 14

Published: Oct. 7, 2024

This study compares the performance of GPT-3.5, GPT-4, Bard, and Gemini in answering Library Information Science (LIS) questions. Sixteen questions were used for assessment, with two independent examiners scoring initial successive responses from each AI system. Statistical analyses, including one-way Analysis Variance (ANOVA), sample t-test, one-sample employed to identify differences. The results revealed consistency generated across iterations all systems. Significant differences observed among models, Bard consistently underperforming compared Gemini. uncovered variability examiners' emphasized need multiple evaluators assessment.

Language: Английский

Citations

0