Comparison of ChatGPT knowledge against 2020 consensus statement on ankyloglossia in children

Int J Pediatr Otorhinolaryngol. 2024 May:180:111957. doi: 10.1016/j.ijporl.2024.111957. Epub 2024 Apr 16.

Abstract

Objective: This paper evaluates ChatGPT's accuracy and consistency in providing information on ankyloglossia, a congenital oral condition. Assessing alignment with expert consensus, the study explores potential implications for patients relying on AI for medical information.

Methods: Statements from the 2020 clinical consensus statement on ankyloglossia were presented to ChatGPT, and its responses were scored using a 9-point Likert scale. The study analyzed the mean and standard deviation of ChatGPT scores for each statement. Statistical analysis was conducted using Excel.

Results: Among the 63 statements assessed, 67 % of ChatGPT responses closely aligned with expert consensus mean scores. However, 17 % (11/63) were statements in which the ChatGPT mean response was different from the CCS mean by 2.0 or greater, raising concerns about ChatGPT's potential influence in disseminating uncertain or debated medical information. Variations in mean scores highlighted discrepancies, with some statements showing significant deviations from expert opinions.

Conclusion: While ChatGPT mirrored medical viewpoints on ankyloglossia, alignment with non-consensus statements raises caution in relying on it for medical advice. Future research should refine AI models, address inaccuracies, and explore diverse user queries for safe integration into medical decision-making. Despite potential benefits, ongoing examination of ChatGPT's power and limitations is crucial, considering its impact on health equity and information access.

Keywords: Ankyloglossia; Artificial intelligence; ChatGPT; Consensus statement.

Publication types

  • Comparative Study

MeSH terms

  • Ankyloglossia*
  • Child
  • Consensus*
  • Humans