EVALUATION OF READABILITY INDICES OF CHATGPT-4 AND GOOGLE GEMINI IN PATIENT EDUCATION ABOUT INTRACRANIAL HEMORRHAGES
DOI:
https://doi.org/10.34689/k5j79m06Keywords:
Subarachnoid haemorrhage, intracranial aneurysms, Coleman-Liau ReadabilityAbstract
Aim. Subarachnoid haemorrhage (SAH) and intracranial aneurysms are critical neurological conditions with significant implications for patient morbidity and mortality. The intersection of readability indices and artificial intelligence (AI) is an emerging field that aims to improve the accessibility and understanding of written material in different areas. Readability indices, such as the Automated Readability Index (ARI) and the Flesch–Kincaid grade level, provide quantitative measures of text complexity that are crucial for tailoring content to specific audiences . Therefore, in this study, we aimed to examine the answers given to questions asked by patients with intracranial haemorrhage using readability indices. Materials and Methods. In this study, questions directly posed by patients and their relatives concerning subarachnoid haemorrhage and intracranial aneurysms were compiled. The collated questions were then divided into subcategories, including definition, diagnosis, treatment options, surgical procedures, complications, and impact on daily life. Flesch Reading Ease (FRE) Formula, Fog Scale (Gunning FOG Formula), SMOG Index, Automated Readability Index (ARI), Coleman-Liau Index, Linsear Write Formula, Dale-Chall Readability Score, Spache Readability Formula. AI technologies were compared across groups. Results. The results indicate that for most readability indices, there is no statistically significant difference between the two models, with one notable exception; Coleman-Liau Readability Index: Gemini: 10.59 ± 0.98 vs. ChatGPT: 11.80 ± 1.64; p-Value: 0.014 The only exception is the Coleman-Liau Readability Index, where a statistically significant difference was found, with ChatGPT showing a slightly higher score, implying potentially greater complexity according to that specific measure. Conclusion. Our article provides valuable quantitative data on the readability of texts from ChatGPT and Gemini, its scope is narrow. A more comprehensive study would ideally include qualitative assessments, a broader range of text types, and detailed information on the methodology and model versions to provide a more holistic understanding of the models' performance.
References
Mutlucan U.O., Bedel C., Zortuk Ö., Selvi F. Evaluation of readability indices of ChatGPT-4 and Google Gemini in patient education about intracranial hemorrhages // Nauka i Zdravookhranenie [Science & Healthcare]. 2025. Vol.27 (4), pp. 107- 112. doi 10.34689/SH.2025.27.4.014
Мутлукан У.О., Бедел Дж., Зортук О., Сельви Ф. Оценка показателей читаемости CHATGPT-4 и Google Gemini в обучении пациентов по вопросам внутричерепных кровоизлияний // Наука и Здравоохранение. 2025. Vol.27 (4), С.107-112. doi 10.34689/SH.2025.27.4.014
Мутлукан У.О., Бедел Дж., Зортук О., Сельви Ф. Пациенттердің бас сүйек ішіне қан құйылу бойынша білім алуында CHATGPT-4 және Google Gemini оқу көрсеткіштерін бағалау // Ғылым және Денсаулық сақтау. 2025. Vol.27 (4), Б. 107-112. doi 10.34689/SH.2025.27.4.014
Downloads
Published
License
Copyright (c) 2025 Рецензируемый медицинский научно-практический журнал «Наука и здравоохранение»

This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.