Bratislavské lekárske listy/Bratislava medical journal, Journal Year: 2025, Volume and Issue: unknown
Published: April 14, 2025
Language: Английский
Bratislavské lekárske listy/Bratislava medical journal, Journal Year: 2025, Volume and Issue: unknown
Published: April 14, 2025
Language: Английский
BMC Medical Informatics and Decision Making, Journal Year: 2025, Volume and Issue: 25(1)
Published: March 26, 2025
This study evaluates and compares ChatGPT-4.0, Gemini Pro 1.5(0801), Claude 3 Opus, Qwen 2.0 72B in answering dental implant questions. The aim is to help doctors underserved areas choose the best LLMs(Large Language Model) for their procedures, improving care accessibility clinical decision-making. Two specialists with over twenty years of experience evaluated models. Questions were categorized into simple true/false, complex short-answer, real-life case analyses. Performance was measured using precision, recall, Bayesian inference-based evaluation metrics. ChatGPT-4 exhibited most stable consistent performance on both 1.5(0801)performed well questions but less tasks. provided high-quality answers specific cases showed variability. opus had lowest across various Statistical analysis indicated significant differences between models diagnostic not treatment planning. reliable model handling medical questions, followed by 1.5(0801). shows potential lacks consistency, Opus performs poorly overall. Combining multiple recommended comprehensive
Language: Английский
Citations
0Bratislavské lekárske listy/Bratislava medical journal, Journal Year: 2025, Volume and Issue: unknown
Published: April 14, 2025
Language: Английский
Citations
0