BMC Oral Health, cilt.25, sa.1, 2025 (SCI-Expanded, Scopus)
Background: To evaluate the credibility of large language models (LLMs) compared to American Association of Orthodontists (AAO) and British Orthodontic Society (BOS) guides regarding nutritional guidelines for orthodontic patients. Methods: The responses offered by ChatGPT 4.0, Copilot and Gemini were assessed for information credibility about tooth decay, food, beverages, oral care, and further assistance as; compatible, compatible but insufficient, partially incompatible and incompatible compared to the guidelines. Reliability was analyzed by the DISCERN tool. The readability of the sources was assessed using the Flesch Reading Ease Score and the Flesch–Kincaid Grade Level. The Friedman Test was conducted to compare DISCERN scores. Results: Responses of LLMs were understandable and compatible with the guidelines, but detailed information on tooth decay, dental plaque and risks of acidic environments were inadequate. ChatGPT 4.0, Copilot and Gemini provided detailed lists of foods to avoid and include. Only AAO suggested being aware of extreme temperatures and usage of sugar-free gums. All sources mentioned the necessity of good oral hygiene, but oral hygiene tools were not mentioned in Copilot. All, except ChatGPT 4.0, recommended orthodontist consultation for personalized advice. BOS leaflet had the highest mean DISCERN score (4.70 ± 0.27), followed by Gemini (4.54 ± 1.03), AAO web-source (4.45 ± 0.75), Copilot (3.87 ± 1.64) and ChatGPT 4.0 (3.08 ± 1.65), revealing no significant difference. BOS and AAO were more readable than the LLMs. ChatGPT 4.0 was more readable among the LLMs but was still found to be difficult for the readers. Conclusion: Guidelines have a superior narrative in terms of their detailed content and especially the justifications for the recommendations. Artificial Intelligence (AI)-supported LLMs provided understandable, simple and accurate information, despite lack of some details on certain topics. The readability of the responses from LLMs was difficult. Overall, patients should be advised that pre-trained algorithms should be used with caution as a source of information and that they should receive individual information from their orthodontists.