Please use this identifier to cite or link to this item: http://hdl.handle.net/1942/43057
Full metadata record
DC FieldValueLanguage
dc.contributor.authorD'Anna, Gennaro-
dc.contributor.authorVAN CAUTER, Sofie-
dc.contributor.authorThurnher, Majda-
dc.contributor.authorVan Goethem, Johan-
dc.contributor.authorHaller, Sven-
dc.date.accessioned2024-06-05T06:44:39Z-
dc.date.available2024-06-05T06:44:39Z-
dc.date.issued2024-
dc.date.submitted2024-06-05T06:40:13Z-
dc.identifier.citationNEURORADIOLOGY,-
dc.identifier.urihttp://hdl.handle.net/1942/43057-
dc.description.abstractWe compared different LLMs, notably chatGPT, GPT4, and Google Bard and we tested whether their performance differs in subspeciality domains, in executing examinations from four different courses of the European Society of Neuroradiology (ESNR) notably anatomy/embryology, neuro-oncology, head and neck and pediatrics. Written exams of ESNR were used as input data, related to anatomy/embryology (30 questions), neuro-oncology (50 questions), head and neck (50 questions), and pediatrics (50 questions). All exams together, and each exam separately were introduced to the three LLMs: chatGPT 3.5, GPT4, and Google Bard. Statistical analyses included a group-wise Friedman test followed by a pair-wise Wilcoxon test with multiple comparison corrections. Overall, there was a significant difference between the 3 LLMs (p < 0.0001), with GPT4 having the highest accuracy (70%), followed by chatGPT 3.5 (54%) and Google Bard (36%). The pair-wise comparison showed significant differences between chatGPT vs GPT 4 (p < 0.0001), chatGPT vs Bard (p < 0. 0023), and GPT4 vs Bard (p < 0.0001). Analyses per subspecialty showed the highest difference between the best LLM (GPT4, 70%) versus the worst LLM (Google Bard, 24%) in the head and neck exam, while the difference was least pronounced in neuro-oncology (GPT4, 62% vs Google Bard, 48%). We observed significant differences in the performance of the three different LLMs in the running of official exams organized by ESNR. Overall GPT 4 performed best, and Google Bard performed worst. This difference varied depending on subspeciality and was most pronounced in head and neck subspeciality.-
dc.description.sponsorshipWe would like to thank Sara Fullone of European Society of Neuroradiology Central Ofce for the support provided.-
dc.language.isoen-
dc.publisherSPRINGER-
dc.rightsThe Author(s), under exclusive licence to Springer-Verlag GmbH Germany, part of Springer Nature 2024-
dc.subject.otherAI-
dc.subject.otherLLM-
dc.subject.otherchatGPT-
dc.subject.otherNeuroradiology-
dc.subject.otherGPT4-
dc.titleCan large language models pass official high-grade exams of the European Society of Neuroradiology courses? A direct comparison between OpenAI chatGPT 3.5, OpenAI GPT4 and Google Bard-
dc.typeJournal Contribution-
local.format.pages6-
local.bibliographicCitation.jcatA1-
dc.description.notesD'Anna, G (corresponding author), ASST Ovest Milanese, Neuroimaging Unit, Legnano, Milan, Italy.-
dc.description.notesgennaro.danna@gmail.com-
local.publisher.placeONE NEW YORK PLAZA, SUITE 4600, NEW YORK, NY, UNITED STATES-
local.type.refereedRefereed-
local.type.specifiedArticle-
local.bibliographicCitation.statusEarly view-
dc.identifier.doi10.1007/s00234-024-03371-6-
dc.identifier.pmid38705899-
dc.identifier.isi001216135700001-
dc.contributor.orcidD'Anna, Gennaro/0000-0001-9890-9359; Haller, Sven/0000-0001-7433-0203-
local.provider.typewosris-
local.description.affiliation[D'Anna, Gennaro] ASST Ovest Milanese, Neuroimaging Unit, Legnano, Milan, Italy.-
local.description.affiliation[Van Cauter, Sofie] Ziekenhuis Oost Limburg, Dept Med Imaging, Genk, Belgium.-
local.description.affiliation[Van Cauter, Sofie] Hasselt Univ, Dept Med & Life Sci, Hasselt, Belgium.-
local.description.affiliation[Thurnher, Majda] Med Univ Vienna, Dept Biomed Imaging & Image Guided Therapy, Vienna, Austria.-
local.description.affiliation[Van Goethem, Johan] VITAZ, Dept Med & Mol Imaging, St Niklaas, Belgium.-
local.description.affiliation[Van Goethem, Johan] Univ Hosp Antwerp, Dept Radiol, Antwerp, Belgium.-
local.description.affiliation[Haller, Sven] CIMC Ctr Imagerie Med Cornavin, Geneva, Switzerland.-
local.description.affiliation[Haller, Sven] Uppsala Univ, Dept Surg Sci, Radiol, Uppsala, Sweden.-
local.description.affiliation[Haller, Sven] Univ Geneva, Fac Med, Geneva, Switzerland.-
local.description.affiliation[Haller, Sven] Capital Med Univ, Beijing Tiantan Hosp, Dept Radiol, Beijing, Peoples R China.-
local.uhasselt.internationalyes-
item.fulltextWith Fulltext-
item.contributorD'Anna, Gennaro-
item.contributorVAN CAUTER, Sofie-
item.contributorThurnher, Majda-
item.contributorVan Goethem, Johan-
item.contributorHaller, Sven-
item.accessRightsRestricted Access-
item.fullcitationD'Anna, Gennaro; VAN CAUTER, Sofie; Thurnher, Majda; Van Goethem, Johan & Haller, Sven (2024) Can large language models pass official high-grade exams of the European Society of Neuroradiology courses? A direct comparison between OpenAI chatGPT 3.5, OpenAI GPT4 and Google Bard. In: NEURORADIOLOGY,.-
crisitem.journal.issn0028-3940-
crisitem.journal.eissn1432-1920-
Appears in Collections:Research publications
Files in This Item:
File Description SizeFormat 
Can large language models pass official high-grade exams of the European Society .pdf
  Restricted Access
Early view950.95 kBAdobe PDFView/Open    Request a copy
Show simple item record

Google ScholarTM

Check

Altmetric


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.