Instytut Podstawowych Problemów Techniki
Polskiej Akademii Nauk

Partnerzy

Klaudia Watros


Ostatnie publikacje
1.  Olszewski R., Watros K., Mańczak M., Owoc J., Jeziorski K., Brzeziński J., Assessing the response quality and readability of chatbots in cardiovascular health, oncology, and psoriasis: A comparative study, International Journal of Medical Informatics, ISSN: 1386-5056, DOI: 10.1016/j.ijmedinf.2024.105562, Vol.190, No.105562, pp.1-7, 2024

Streszczenie:
Background: Chatbots using the Large Language Model (LLM) generate human responses to questions from all
categories. Due to staff shortages in healthcare systems, patients waiting for an appointment increasingly use
chatbots to get information about their condition. Given the number of chatbots currently available, assessing the
responses they generate is essential.
Methods: Five chatbots with free access were selected (Gemini, Microsoft Copilot, PiAI, ChatGPT, ChatSpot) and
blinded using letters (A, B, C, D, E). Each chatbot was asked questions about cardiology, oncology, and psoriasis.
Responses were compared to guidelines from the European Society of Cardiology, American Academy of
Dermatology and American Society of Clinical Oncology. All answers were assessed using readability scales
(Flesch Reading Scale, Gunning Fog Scale Level, Flesch-Kincaid Grade Level and Dale-Chall Score). Using a 3-
point Likert scale, two independent medical professionals assessed the compliance of the responses with the
guidelines.
Results: A total of 45 questions were asked of all chatbots. Chatbot C gave the shortest answers, 7.0 (6.0 – 8.0), and Chatbot A the longest 17.5 (13.0 – 24.5). The Flesch Reading Ease Scale ranged from 16.3 (12.2 – 21.9)
(Chatbot D) to 39.8 (29.0 – 50.4) (Chatbot A). Flesch-Kincaid Grade Level ranged from 12.5 (10.6 – 14.6) (Chatbot A) to 15.9 (15.1 – 17.1) (Chatbot D). Gunning Fog Scale Level ranged from 15.77 (Chatbot A) to 19.73 (Chatbot D). Dale-Chall Score ranged from 10.3 (9.3 – 11.3) (Chatbot A) to 11.9 (11.5 – 12.4) (Chatbot D).
Conclusion: This study indicates that chatbots vary in length, quality, and readability. They answer each question
in their own way, based on the data they have pulled from the web. Reliability of the responses generated by
chatbots is high. This suggests that people who want information from a chatbot need to be careful and verify the answers they receive, particularly when they ask about medical and health aspects.

Słowa kluczowe:
Chatbots,Readability,Cardiovascular health,Oncology

Afiliacje autorów:
Olszewski R. - IPPT PAN
Watros K. - inna afiliacja
Mańczak M. - National Institute of Geriatrics Rheumatology and Rehabilitation (PL)
Owoc J. - National Institute of Geriatrics Rheumatology and Rehabilitation (PL)
Jeziorski K. - National Institute of Geriatrics Rheumatology and Rehabilitation (PL)
Brzeziński J. - inna afiliacja
140p.

Kategoria A Plus

IPPT PAN

logo ippt            ul. Pawińskiego 5B, 02-106 Warszawa
  +48 22 826 12 81 (centrala)
  +48 22 826 98 15
 

Znajdź nas

mapka
© Instytut Podstawowych Problemów Techniki Polskiej Akademii Nauk 2024