Hostname: page-component-cd9895bd7-mkpzs Total loading time: 0 Render date: 2024-12-24T18:17:51.708Z Has data issue: false hasContentIssue false

Large Language Model (LLM)-Powered Chatbots Fail to Generate Guideline-Consistent Content on Resuscitation and May Provide Potentially Harmful Advice

Published online by Cambridge University Press:  06 November 2023

Alexei A. Birkun*
Affiliation:
Department of General Surgery, Anaesthesiology, Resuscitation and Emergency Medicine, Medical Academy named after S.I. Georgievsky of V.I. Vernadsky Crimean Federal University, Simferopol, 295051, Russian Federation
Adhish Gautam
Affiliation:
Regional Government Hospital, Una (H.P.), 174303, India
*
Correspondence: Alexei A. Birkun, MD, DMedSc Medical Academy named after S.I. Georgievsky of V.I. Vernadsky Crimean Federal University Lenin Blvd, 5/7, Simferopol, 295051, Russian Federation E-mail: [email protected]
Rights & Permissions [Opens in a new window]

Abstract

Introduction:

Innovative large language model (LLM)-powered chatbots, which are extremely popular nowadays, represent potential sources of information on resuscitation for the general public. For instance, the chatbot-generated advice could be used for purposes of community resuscitation education or for just-in-time informational support of untrained lay rescuers in a real-life emergency.

Study Objective:

This study focused on assessing performance of two prominent LLM-based chatbots, particularly in terms of quality of the chatbot-generated advice on how to give help to a non-breathing victim.

Methods:

In May 2023, the new Bing (Microsoft Corporation, USA) and Bard (Google LLC, USA) chatbots were inquired (n = 20 each): “What to do if someone is not breathing?” Content of the chatbots’ responses was evaluated for compliance with the 2021 Resuscitation Council United Kingdom guidelines using a pre-developed checklist.

Results:

Both chatbots provided context-dependent textual responses to the query. However, coverage of the guideline-consistent instructions on help to a non-breathing victim within the responses was poor: mean percentage of the responses completely satisfying the checklist criteria was 9.5% for Bing and 11.4% for Bard (P >.05). Essential elements of the bystander action, including early start and uninterrupted performance of chest compressions with adequate depth, rate, and chest recoil, as well as request for and use of an automated external defibrillator (AED), were missing as a rule. Moreover, 55.0% of Bard’s responses contained plausible sounding, but nonsensical guidance, called artificial hallucinations, that create risk for inadequate care and harm to a victim.

Conclusion:

The LLM-powered chatbots’ advice on help to a non-breathing victim omits essential details of resuscitation technique and occasionally contains deceptive, potentially harmful directives. Further research and regulatory measures are required to mitigate risks related to the chatbot-generated misinformation of public on resuscitation.

Type
Original Research
Copyright
© The Author(s), 2023. Published by Cambridge University Press on behalf of the World Association for Disaster and Emergency Medicine

Introduction

Recent public release of novel conversational bots powered by artificial intelligence (AI) algorithms have resulted in rapid and continued growth of academic interest and ignited wide debates concerning the possible impact of these tools on society and research. Reference Haleem, Javaid and Singh1,Reference De Angelis, Baglivo and Arzilli2 These cutting-edge chatbots utilize AI technology called large language models (LLMs). These LLMs are trained on massive amounts of text data to produce new, fluent, human-like text in response to a user input by predicting and repeatedly generating the next word in a sentence based on the preceding words. Reference Hassani and Silva3 By means of the LLM, the chatbots offer unprecedented opportunities to handle a wide range of natural language processing tasks, including text writing, content summarization, and question answering.

Except for several exploratory studies, Reference Ahn4Reference Sarbay, Berikol and Özturan9 the LLM-based chatbots currently lack evaluation in terms of perspective application in emergency medicine. In relation to resuscitation research and practice, where implementation of contemporary digital technologies is encouraged, Reference Berg, Cheng and Panchal10,Reference Semeraro, Greif and Böttiger11 it seems important and well-timed to examine the practicability of utilizing the LLM-powered chatbots in two directions: (1) to generate guideline-consistent advice on help in cardiac arrest (for purposes of public resuscitation education or for just-in-time informational support of untrained lay rescuers in a real-life emergency), and thus to contribute towards the promotion of community response to out-of-hospital cardiac arrest; and (2) to evaluate the quality of information on resuscitation available online (that is known to be generally low Reference Liu, Haukoos and Sasson12Reference Birkun, Gautam, Trunkwala and Böttiger14 ) and suggest how to enhance the content. The latter could help to establish systematic quality surveillance and assurance for publicly available resources on resuscitation and reduce potential harm from misinformation.

Accordingly, this study was commenced to assess the quality of advice on how to give help to a non-breathing victim generated by two prominent LLM-powered chatbots, as well as to test the ability of the chatbots to perform self-rating of their advice and improve quality of the content.

Methods

Study Design

This was a cross-sectional, analytical study based on open-source online services’ data. The study design was informed by previous related research. Reference Birkun and Gautam6,Reference Birkun and Dr15 The chatbots were interrogated in English using the Microsoft Edge web browser (Microsoft Corporation; Redmond, Washington USA) for the new Bing, and Google Chrome web browser (Google LLC; Mountain View, California USA) for Bard, on an Apple macOS Big Sur (Apple Inc.; Cupertino, California USA) operated personal computer. In the chatbots’ settings, the region of search was set as the United Kingdom (UK), and a Virtual Private Network (VPN) was used to simulate search from this country with location set to London. In order to avoid impact of previous user activity on the chatbots’ responses, before each search query, all browsing history, download history, search history, cache, and cookies were cleared from the browsers, Microsoft, and Google accounts. For Bing, the search was made under “More Precise” conversation style.

In May 2023, the chatbots were sequentially inquired (20 times per each chatbot): (1) “What to do if someone is not breathing?”; (2) to rate content of the chatbot’s own response to the first query for compliance with the Resuscitation Council UK (London, England) Guidelines on a 10-point scale (one being very low compliance, ten being very high compliance); (3) to indicate whether the response contains any guideline-noncompliant instructions; and (4) to correct the response to make it fully compliant with the guidelines (Appendix Table A shows literal prompts; available online only). Original and self-corrected chatbot responses containing instructions on help to a non-breathing victim were tabulated and independently manually assessed by the authors for compliance with the 2021 Resuscitation Council UK Guidelines on adult Basic Life Support Reference Perkins, Colquhoun and Deakin16 using an author-developed checklist (Dataset Reference Birkun and Gautam17 ). For each item of the checklist, congruence of the chatbot-generated instructions with the guidelines was rated as True (when checklist item wording was satisfied completely), Partially True (when checklist item wording was satisfied in part), or Not True (when corresponding instruction was missing in the chatbot response). Results of the evaluation provided by both authors were compared, and in case of discrepancies, the authors resolved them by consensus. When a chatbot provided links to the source web articles, the articles’ content was evaluated using the same methodology. Also, the authors independently rated original chatbot responses for compliance with the guidelines using the 10-point scale, and the median expert rating was calculated.

Additionally, original and self-corrected chatbot responses were evaluated for length (number of sentences) and checked for readability based on the Flesch-Kincaid Grade Level (FKGL) Reference Kincaid, Fishburne, Rogers and Chissom18 metric using an open online readability analyzer Datayze. 19 The FKGL formula utilizes the average number of syllables per word and average number of words per sentence to conclude how easy a passage of English text is to read and understand. Reference Kincaid, Fishburne, Rogers and Chissom18 The FKGL values correspond with a United States grade level of education. Lower FKGL values entail greater readability.

The New Bing

The new Bing is an AI-powered web search engine by Microsoft Corporation made available for the public in February 2023. The chatbot functionality of the new Bing allows users to perform web search in a conversational way. It searches for relevant content across the web and consolidates what it finds to generate a summarized answer using a LLM from OpenAI (San Francisco, California USA) known as Generative Pre-Trained Transformer 4 (GPT-4). Reference Peters20 Bing centers its response to a user’s query on high-ranking content from the web. It ranks the content by weighing a set of features, including relevance, quality and credibility, and freshness. 21 To determine quality and credibility of a website, it evaluates clarity of purpose of the site, its usability, presentation, and authoritativeness. The latter includes such factors as author’s or site’s reputation, completeness of the content, and transparency of the authorship. Higher quality is considered for a website containing citations and references to data sources. Bing accompanies its responses with links to search results that were used to ground the response.

Bard

Bard is an AI chatbot launched by Google LLC in March 2023. Similar to the new Bing, to respond to user’s inquiries, it retrieves information from the internet. To produce the responses, Bard utilizes Google’s conversational AI language model called Language Model for Dialogue Applications (LaMDA). 22 The mechanism how Bard ranks its web search results to generate answers is undisclosed. Unlike the new Bing, Bard does not routinely cite sources of information for its responses. 23

The study results were analyzed descriptively. Mann Whitney U Test and Wilcoxon signed-rank test were used to determine differences.

All data that support the findings of this study are openly available in Mendeley Data repository. Reference Birkun and Gautam17

Because the study did not involve human participants, it did not require ethical approval.

Results

Both chatbots comprehended all user queries and provided context-consistent textual responses.

Bing’s responses were considerably shorter than Bard’s responses (Table 1). Readability was higher for Bard’s responses, requiring approximately a sixth-grade level of education to understand the text compared with seventh-eighth-grade level for Bing.

Table 1. Length and Readability of the Chatbot Responses

Abbreviation: IQR, interquartile range.

a Bing vs Bard, P <.001;

b Bing vs Bard, P <.050.

Original chatbot responses showed poor coverage of the guideline-consistent instructions on help to a non-breathing victim (Table 2). Essential elements of the bystander action, including assurance of safety, request for and use of an automated external defibrillator (AED), early start, and uninterrupted performance of chest compressions following the recommended technique, were for the most part omitted. Mean percentage of the chatbots’ responses completely satisfying the checklist criteria was 9.5% for Bing and 11.4% for Bard (P >.050).

Table 2. Compliance of Original Chatbot Responses Containing Instructions on Help to a Non-Breathing Victim with the Checklist Criteria

Abbreviations: AED, automated external defibrillator; EMS, Emergency Medical Services.

The chatbots over-estimated the quality of their responses in terms of compliance with the resuscitation guidelines. Median (interquartile range) self-rating of the original responses amounted 7.0 (7.0–7.0) points for Bing and 9.0 (9.0–9.0) points for Bard, whereas the expert rating was significantly lower (P <.001) — 4.0 (2.0–4.5) and 3.0 (2.6–4.0) points, respectively.

Bing’s original responses were more accurate in terms of suggestion of the search-region-specific Emergency Medical Services (EMS) telephone number. Bing recommended to call the UK national emergency number 9-9-9 in 95.0% (n = 19) of cases, whereas Bard’s advice was always to call the United States national emergency number 9-1-1 or a local (unspecified) emergency number.

When inquired about whether the responses contain any guidelines-inconsistent instructions, both chatbots denied this on all occasions. However, the manual assessment revealed that all Bing and Bard responses included some superfluous instructions which either were inappropriate for an untrained lay rescuer or contradicted current resuscitation guidelines (Table 3). Whereas for Bing, the excessive instructions were limited to unnecessary breathing check and suggestion to give rescue breaths, Bard in 55.0% responses (n = 11) presented one or more seemingly plausible but factually incorrect and commonly potentially harmful statements, representing the phenomenon of “artificial hallucination.” Reference Alkaissi and McFarlane24

Table 3. Instructions Contained in Original and Self-Corrected Chatbot Responses to the Query: “What to do if someone is not breathing?” – which were Considered Guideline-Inconsistent or Inappropriate for an Untrained Lay Rescuer

Abbreviation: CPR, cardiopulmonary resuscitation.

As for the sources of information contained in the chatbots’ responses, Bing on all occasions cited the same two web articles which demonstrated incomplete adherence with the resuscitation guidelines, omitting important aspects of the life-saving approach (percentage of the checklist items completely or partially satisfied by the content of these web articles was 36.4% and 72.7%; Dataset Reference Birkun and Gautam17 ). Bard did not cite any sources for its responses.

In reply to the request to correct the original responses to ensure full compliance with the guidelines and applicability of the instructions on cardiopulmonary resuscitation (CPR) for untrained rescuers only, both chatbots made adjustments to their responses. Despite some enhancement, quality of the responses did not improve significantly (Table 4). Mean percentage of the chatbots’ responses having complete compliance with the checklist criteria remained low (14.5% for Bing and 24.1% for Bard, P >.050), and superfluous guidelines-inconsistent instructions on many occasions remained in place (Table 3). Bard improved its advice in terms of accuracy of suggestion of the search-region-specific EMS number: the UK emergency number 9-9-9 was recommended in 80.0% (n = 16) self-corrected responses (versus 95.0%, n = 19 for Bing).

Table 4. Compliance of Self-Corrected Chatbot Responses Containing Instructions on Help to a Non-Breathing Victim with the Checklist Criteria

Abbreviations: AED, automated external defibrillator; EMS, Emergency Medical Services.

Discussion

Despite the innovative AI-powered question-answering systems seeming to constitute a promising opportunity to engage lay people in provision of help and to improve health outcomes in emergencies, there are little published data on the effectiveness of such systems. Previous studies tested capabilities of voice-based conversational digital assistants (Alexa [Amazon; Seattle, Washington USA], Cortana [Cortana Corp.; Falls Church, Virginia USA], Google Assistant [Google LLC; Mountain View, California USA], and Siri [Apple Inc.; Cupertino, California USA]) Reference Bickmore, Trinh and Olafsson25,Reference Picard, Smith, Picard and Can Alexa26 and Google web search engine’s question-answering system Reference Birkun and Dr15 in responding to inquiries related to first aid in a range of emergency conditions. The studies showed that the AI assistants frequently failed to recommend how to give help, or suggested to take inappropriate actions that could have resulted in harm to a victim. Such poor performance in particular was explained by limitations of the search engine’s AI algorithms, that seem to generate and present responses as literal quotations automatically extracted from a search-engine-indexed webpage that most closely resemble the user’s query. Reference Birkun and Dr15

Current research focused on evaluation of performance of the two flagship LLM-powered chatbots — Bing and Bard — which exercise a fundamentally new approach to question answering. Instead of using the quote-offering as is done by conventional search engine question-answering systems, the LLM chatbots search information online, perform ranking of the information, and utilize a neural network to generate summarized responses based on the high-ranking content. 21,22

The study found that both chatbots at all times correctly recognized user inquiries and provided easily comprehensible responses containing some advice on how to give help to a non-breathing victim. However, quality of the responses’ content in terms of compliance with the resuscitation guidelines was low. Both Bing and Bard omitted essential characteristics of the life-saving help in all responses. In fact, the mean percentage of the chatbots’ responses completely satisfying the guidelines-based checklist criteria was less than 10% for Bing and less than 12% for Bard. For instance, the chatbots never suggested to request an AED, to begin chest compressions as early as possible, or to perform compressions with minimal interruptions. Where the guideline-consistent instructions were given, the chatbots usually did not provide sufficient details on the life-saving technique. In particular, important characteristics of chest compressions, including compression depth and rate, as well as the need to release pressure on the chest after each compression, were missing as a rule. Lack of sufficient details in LLM-powered chatbots’ responses to user inquiries on help in emergencies, although much less prominent than in the current study, was reported in previous related research. Reference Birkun and Gautam6,Reference Dahdah, Kassab, Helou, Gaballa, Sayles and Phelan7

Along with that, the chatbots’ responses commonly included directions which were guidelines-compliant but inappropriate for an untrained rescuer (eg, advice to give rescue breaths), or contained AI hallucinations — incorrect and nonsensical guidance that represent risk of harm, since it may sound believable for an unfamiliar user. All the hallucinations were generated by Bard. These findings are contrasting with results of previous exploratory studies Reference Birkun and Gautam6,Reference Dahdah, Kassab, Helou, Gaballa, Sayles and Phelan7 which reported that LLM-based chatbots (Bing and ChatGPT [OpenAI; San Francisco, California USA]) did not instruct to perform harmful actions in a range of health emergencies.

Further, this study showed that the chatbots substantially over-estimated the quality of their advice on help for a non-breathing victim in terms of compliance with the resuscitation guidelines. Also, when being asked to enhance the responses’ content to make the advice fully guideline-concordant and applicable for an untrained rescuer, the chatbots corrected their responses, but the improvement was negligible and quality of the instructions remained low. Potentially harmful guideline-inconsistent advice and instructions inappropriate for an untrained bystander were mostly kept in place.

Taken together, these observations indicate that currently neither Bing nor Bard should be considered as a source of reliable guideline-consistent information on resuscitation, and the chatbots cannot be utilized to detect quality flaws or enhance quality of such information. Moreover, the artificial hallucinations generated by Bard may sound convincing for an incompetent user and therefore create an apparent risk of causing harm in case the user will take action following the chatbot advice.

Although the developers of Bing and Bard give up responsibility by asserting that the chatbots can make mistakes, provide incomplete, inaccurate, or inappropriate responses, 22,27 one should consider that a large portion of users may neglect the disclaimers, whereas the ever-increasing popularity of the LLM-powered chatbots along with their integration into the search engines and mobile devices would probably greatly intensify public use of these tools as an everyday source of informational support, including in real-life health emergencies. This stipulates the need on the one hand to enhance laypeople’s awareness of potential risks related with reliance on the chatbots’ advice in health crises instead of seeking professional help, and on the other hand, to develop regulatory procedures aimed at elimination of potential harm from the chatbot-generated misinformation by replacing the uncontrollable LLM-mediated question answering to the health-related questions with reliable human expert-developed advice. Both tasks would require commitment and close collaboration of the AI chatbot developers with recognized public health organizations.

Limitations

This study has limitations. Both tested chatbots currently run in a pilot version. Performance of the chatbots could change as a result of evolution of the question-answering AI algorithms. Repeated investigation carried out at a later point in time, with different search queries, languages, or search regions, may produce different results. Reproducibility of the research findings is further limited by the dynamic nature of the internet utilized by the chatbots as a source of information.

Conclusions

The LLM-powered chatbots readily respond to user inquiries concerning advice on help to a non-breathing victim by generating clearly understandable summarized answers containing instructions on resuscitation. However, the responses always omit essential details on the life-saving technique and occasionally contain deceptive, nonsensical directives which create risk for inadequate care and harm to a victim. The chatbots over-estimate the quality of their responses and were unable to improve their advice to achieve congruence with the current resuscitation guidelines. Along with further research aimed at better understanding possible use of the LLM-based chatbots in emergency medicine, regulatory actions are required to mitigate risks related to the AI-generated misinformation.

Conflicts of interest

A.A.B. and A.G. have no conflicts of interest.

Supplementary Materials

To view supplementary material for this article, please visit https://doi.org/10.1017/S1049023X23006568

References

Haleem, A, Javaid, M, Singh, RP. An era of ChatGPT as a significant futuristic support tool: a study on features, abilities, and challenges. BenchCouncil Transactions on Benchmarks, Standards, and Evaluations. 2022;2(4):100089.CrossRefGoogle Scholar
De Angelis, L, Baglivo, F, Arzilli, G, et al. ChatGPT and the rise of large language models: the new AI-driven infodemic threat in public health. Front Public Health. 2023;11:1166120.CrossRefGoogle ScholarPubMed
Hassani, H, Silva, ES. The role of ChatGPT in data science: how AI-assisted conversational interfaces are revolutionizing the field. Big Data Cogn Comput. 2023;7(2):62.CrossRefGoogle Scholar
Ahn, C. Exploring ChatGPT for information of cardiopulmonary resuscitation. Resuscitation. 2023;185:109729.CrossRefGoogle ScholarPubMed
Altamimi, I, Altamimi, A, Alhumimidi, AS, Altamimi, A, Temsah, MH. Snakebite advice and counseling from artificial intelligence: an acute venomous snakebite consultation with ChatGPT. Cureus. 2023;15(6):e40351.Google ScholarPubMed
Birkun, AA, Gautam, A. Instructional support on first aid in choking by an artificial intelligence-powered chatbot. Am J Emerg Med. 2023;70:200202.CrossRefGoogle ScholarPubMed
Dahdah, JE, Kassab, J, Helou, MCE, Gaballa, A, Sayles, S 3rd, Phelan, MP. ChatGPT: a valuable tool for emergency medical assistance. Ann Emerg Med. 2023;82(3):411413.CrossRefGoogle ScholarPubMed
Fijačko, N, Gosak, L, Štiglic, G, Picard, CT, John Douma, M. Can ChatGPT pass the life support exams without entering the American Heart Association course? Resuscitation. 2023;185:109732.CrossRefGoogle ScholarPubMed
Sarbay, İ, Berikol, GB, Özturan, İU. Performance of emergency triage prediction of an open access natural language processing based chatbot application (ChatGPT): a preliminary, scenario-based cross-sectional study. Turkish J Emerg Med. 2023;23(3):156.CrossRefGoogle ScholarPubMed
Berg, KM, Cheng, A, Panchal, AR, et al. Part 7: Systems of Care: 2020 American Heart Association Guidelines for Cardiopulmonary Resuscitation and Emergency Cardiovascular Care. Circulation. 2020;142(16_suppl_2):S580S604.CrossRefGoogle ScholarPubMed
Semeraro, F, Greif, R, Böttiger, BW, et al. European Resuscitation Council Guidelines 2021: systems saving lives. Resuscitation. 2021;161:8097.CrossRefGoogle ScholarPubMed
Liu, KY, Haukoos, JS, Sasson, C. Availability and quality of cardiopulmonary resuscitation information for Spanish-speaking population on the Internet. Resuscitation. 2014;85(1):131137.CrossRefGoogle ScholarPubMed
Metelmann, B, Metelmann, C, Schuffert, L, Hahnenkamp, K, Brinkrolf, P. Medical correctness and user friendliness of available apps for cardiopulmonary resuscitation: systematic search combined with guideline adherence and usability evaluation. JMIR Mhealth Uhealth. 2018;6:e190.CrossRefGoogle ScholarPubMed
Birkun, A, Gautam, A, Trunkwala, F, Böttiger, BW. Open online courses on basic life support: availability and resuscitation guidelines compliance. Am J Emerg Med. 2022;62:102107.CrossRefGoogle ScholarPubMed
Birkun, AA, Dr, Gautam A.. Google’s advice on first aid: evaluation of the search engine’s question-answering system responses to queries seeking help in health emergencies. Prehosp Disaster Med. 2023;38(3):345351.CrossRefGoogle ScholarPubMed
Perkins, GD, Colquhoun, M, Deakin, CD, et al. Resuscitation Council UK. 2021 Resuscitation Guidelines. Adult Basic Life Support Guidelines, 2021. https://www.resus.org.uk/library/2021-resuscitation-guidelines/adult-basic-life-support-guidelines. Accessed August 15, 2023.Google Scholar
Birkun, A, Gautam, A. Dataset of analysis of the large language model-powered chatbots’ advice on help to a non-breathing victim. Mendeley Data. 2023;V1.CrossRefGoogle Scholar
Kincaid, JP, Fishburne, RP Jr, Rogers, RL, Chissom, BS. Derivation of New Readability Formulas (Automated Readability Index, Fog Count, and Flesch Reading Ease Formula) for Navy Enlisted Personnel. Millington, Tennessee USA: Naval Technical Training Command, Millington TN Research Branch; 1975.CrossRefGoogle Scholar
Datayze. Readability Analyzer. https://datayze.com/readability-analyzer. Accessed August 15, 2023.Google Scholar
Peters, J. The Bing AI bot has been secretly running GPT-4. The Verge. https://www.theverge.com/2023/3/14/23639928/microsoft-bing-chatbot-ai-gpt-4-llm. Accessed August 15, 2023.Google Scholar
Microsoft Bing. Bing Webmaster Guidelines. https://www.bing.com/webmasters/help/webmasters-guidelines-30fba23a. Accessed August 15, 2023.Google Scholar
Bard. Bard FAQ. https://bard.google.com/faq. Accessed August 15, 2023.Google Scholar
Search Engine Land. SEO. Breaking Bard: Google’s AI chatbot lacks sources, hallucinates, gives bad SEO advice. https://searchengineland.com/google-bard-first-looks-394583. Accessed August 15, 2023.Google Scholar
Alkaissi, H, McFarlane, SI. Artificial hallucinations in ChatGPT: implications in scientific writing. Cureus. 2023;15:e35179.Google ScholarPubMed
Bickmore, TW, Trinh, H, Olafsson, S, et al. Patient and consumer safety risks when using conversational assistants for medical information: an observational study of Siri, Alexa, and Google Assistant. J Med Internet Res. 2018;20(9):e11510.CrossRefGoogle ScholarPubMed
Picard, C, Smith, KE, Picard, K, Can Alexa, Douma MJ., Cortana, Google Assistant and Siri save your life? A mixed-methods analysis of virtual digital assistants and their responses to first aid and basic life support queries. BMJ Innovations. 2020;6.CrossRefGoogle Scholar
Bing. Introducing the new Bing. https://www.bing.com/new. Accessed August 15, 2023.Google Scholar
Figure 0

Table 1. Length and Readability of the Chatbot Responses

Figure 1

Table 2. Compliance of Original Chatbot Responses Containing Instructions on Help to a Non-Breathing Victim with the Checklist Criteria

Figure 2

Table 3. Instructions Contained in Original and Self-Corrected Chatbot Responses to the Query: “What to do if someone is not breathing?” – which were Considered Guideline-Inconsistent or Inappropriate for an Untrained Lay Rescuer

Figure 3

Table 4. Compliance of Self-Corrected Chatbot Responses Containing Instructions on Help to a Non-Breathing Victim with the Checklist Criteria

Supplementary material: File

Birkun and Gautam supplementary material

Birkun and Gautam supplementary material

Download Birkun and Gautam supplementary material(File)
File 12.9 KB