Robotics & Machine Learning Daily News2024,Issue(Jun.24) :31-31.

Selcuk University Reports Findings in Artificial Intelligence (Comparative Analy sis of Artificial Intelligence Chatbot Recommendations for Urolithiasis Manageme nt: A Study of EAU Guideline Compliance)

塞尔库克大学报告人工智能的发现(人工智能聊天机器人推荐泌尿系结石治疗的比较分析:EAU指南依从性研究)

Robotics & Machine Learning Daily News2024,Issue(Jun.24) :31-31.

Selcuk University Reports Findings in Artificial Intelligence (Comparative Analy sis of Artificial Intelligence Chatbot Recommendations for Urolithiasis Manageme nt: A Study of EAU Guideline Compliance)

塞尔库克大学报告人工智能的发现(人工智能聊天机器人推荐泌尿系结石治疗的比较分析:EAU指南依从性研究)

扫码查看

摘要

一位新闻记者-机器人与机器学习的工作人员新闻编辑每日新闻-人工智能的新研究是一篇报道的主题。根据NewsRx记者在土耳其Konya的新闻报道,研究表明:“人工智能(AI)AP应用越来越多地被患者和医生用来获取医疗信息。这项研究集中在欧洲泌尿外科协会(EAU)指南的泌尿系结石部分(涉及肾脏和输尿管结石),这是泌尿科医生的关键参考。”新闻记者从塞尔库克大学的研究中获得了一句话:“我们向四个不同的人工智能聊天机器人发出询问,以评估他们对指南遵守情况的反应。总共115条建议被转化为问题,回答由两名至少有5年经验的泌尿科医生使用5点李克特量表(1-错误,2-不充分,3-有效,4-正确)进行评估。”困惑和Cha tGPT 4.0的平均分分别为4.68(SD:0.80)和4.80(SD:0.47),Bing和Bard的得分都有显著性差异(Bing对困惑,P<0.001;Bard对困惑,P<0.001;Bing对ChatGPT,P<0.001;Bard对ChatGPT,P<0.001;Bard对ChatGPT,PBard在所有聊天机器人中得分最低(分别为47.3%和30%)。我们的研究结果表明,ChatGPT 4.0和Permission对EAU指南的推荐效果最好。

Abstract

By a News Reporter-Staff News Editor at Robotics & Machine Learning Daily News Daily News-New research on Artificial Intelligenc e is the subject of a report. According to news reporting originating in Konya, Turkey, by NewsRx journalists, research stated, "Artificial intelligence (AI) ap plications are increasingly being utilized by both patients and physicians for a ccessing medical information. This study focused on the urolithiasis section (pe rtaining to kidney and ureteral stones) of the European Association of Urology ( EAU) guideline, a key reference for urologists." The news reporters obtained a quote from the research from Selcuk University, "W e directed inquiries to four distinct AI chatbots to assess their responses in r elation to guideline adherence. A total of 115 recommendations were transformed into questions, and responses were evaluated by two urologists with a minimum of 5 years of experience using a 5-point Likert scale (1-False, 2-Inadequate, 3-Su fficient, 4- Correct, and 5-Very Correct). The mean scores for Perplexity and Cha tGPT 4.0 were 4.68 (SD: 0.80) and 4.80 (SD: 0.47), respectively, both significan tly differed the scores of Bing and Bard (Bing vs. Perplexity, p<.001; Bard vs. Perplexity, p<.001; Bing vs. ChatGPT, p<.001; Bard vs. ChatGPT, p<.001). Bing had a mean score of 4 .21 (SD: 0.96), while Bard scored 3.56 (SD: 1.14), with a significant difference (Bing vs. Bard, p<.001). Bard exhibited the lowest score a mong all chatbots. Analysis of references revealed that Perplexity and Bing cite d the guideline most frequently (47.3% and 30%, respe ctively). Our findings demonstrate that ChatGPT 4.0 and, notably, Perplexity ali gn well with EAU guideline recommendations."

Key words

Konya/Turkey/Eurasia/Artificial Intel ligence/Emerging Technologies/Machine Learning

引用本文复制引用

出版年

2024
Robotics & Machine Learning Daily News

Robotics & Machine Learning Daily News

ISSN:
段落导航相关论文