Preprint
Review

Harnessing the Power of ChatGPT in Cardiovascular Medicine: Innovations, Challenges, and Future Directions

Altmetrics

Downloads

178

Views

116

Comments

0

A peer-reviewed article of this preprint also exists.

Submitted:

28 August 2024

Posted:

28 August 2024

You are already at the latest version

Alerts
Abstract
Cardiovascular diseases (CVDs) remain the leading cause of morbidity and mortality globally, posing significant challenges to public health. The rapid development of artificial intelligence, particularly Large Language Models like ChatGPT, has revolutionized various fields, including cardiovascular medicine. ChatGPT, recognized for its broad applicability, efficiency, and public acceptance, has been integrated into multiple facets of CVD management. This comprehensive review examines the current applications of ChatGPT in cardiovascular medicine, emphasizing its innovative and significant role in enhancing diagnosis and clinical decision-making, improving medical education, and facilitating research and academic communication. Furthermore, the review article addresses the significant challenges associated with ChatGPT and proposes directions for future development to overcome these obstacles and boost its effectiveness. By exploring the vast potential of ChatGPT to meet the needs of the large CVD patient population and the extensive community of cardiovascular professionals, this review aims to provide valuable insights for optimizing its use in cardiovascular medicine, ultimately improving patient outcomes and advancing the field.
Keywords: 
Subject: Medicine and Pharmacology  -   Cardiac and Cardiovascular Systems

1. Introduction

The rapid advancement of artificial intelligence (AI) is profoundly transforming the medical field, revolutionizing clinical practices, patient care, medical education, and research methodologies [1,2]. Among various AI technologies, Large Language Models (LLMs) such as ChatGPT are particularly noteworthy due to their broad applicability, extensive public recognition, and widespread adoption [3,4,5]. Cardiovascular diseases (CVDs) represent the most prevalent and severe health threat globally. In the United States, the prevalence of CVDs in adults was 48.6% in 2020, affecting approximately 127.9 million people [6]. Globally, CVDs caused an estimated 19.05 million deaths in 2020, a 19% increase since 2010, with 607.64 million cases reported worldwide [6]. Heart disease and stroke now claim more lives each year than cancer and chronic lower respiratory disease combined, highlighting the significant burden and severity of CVDs [6]. Given the extensive impact of CVDs and the large number of professionals dedicated to their research and treatment, exploring the applications of ChatGPT in this field is both timely and crucial. This comprehensive review aims to systematically evaluate the innovative applications of ChatGPT in the CVD field, addressing current challenges, discussing future directions for development (Figure 1), and providing valuable insights for the vast patient population and the extensive community of professionals committed to cardiovascular health.

2. ChatGPT’s Rapid Rise and Predominance in Public Medical AI Applications

Large Language Models utilize sophisticated AI algorithms to produce text that mimics human language. Trained on vast datasets sourced from the internet, these models excel in answering questions, summarizing information, translating text, and generating creative content [7]. By inputting specific keywords or queries, users can prompt LLMs to generate text on a wide range of topics in various styles. Notable advancements in LLMs include the development of Google’s Bidirectional Encoder Representations from Transformers and OpenAI’s Generative Pre-trained Transformer (GPT) series. OpenAI released GPT-1 in 2018, followed by GPT-2 and GPT-3 in 2020, each iteration showcasing enhanced capabilities. In November 2022, OpenAI introduced ChatGPT, an updated LLM that quickly gained significant attention due to its public accessibility, ease of use, and human-like responses achieved through reinforcement learning from human feedback (RLHF) [8]. ChatGPT’s rapid user base growth, exceeding 100 million active users within two months of launch, underscores its widespread adoption. Its popularity as the most widely used LLM today is due to its ability to provide detailed and accurate responses, appealing to students, professionals, and the general public. A study on public perception attributes ChatGPT’s high acceptance rate to its intuitive interface and seamless integration into everyday tasks [9]. Furthermore, the rapid development of ChatGPT, particularly with the advent of GPT-4 and GPT-4o, has significantly augmented its capabilities. These iterations enable multi-modal interactions (text, audio, image, video) with near-human response times and improved performance across various languages. These technological advancements have further established ChatGPT as a versatile and efficient AI tool, widely embraced for its ability to enhance user experience across multiple domains, including medicine [10].
In the medical field, ChatGPT distinguishes itself for several notable reasons. Its performance on medical examinations, such as the United States Medical Licensing Exam (USMLE), has been rigorously assessed, showing results at or near the passing threshold even without specialized training. A recent study demonstrates that ChatGPT-4 significantly outperforms its predecessor, ChatGPT-3.5, in answering USMLE Step 2 Clinical Knowledge questions, achieving an accuracy of 87.2% compared to 47.7% [11]. Furthermore, it showcases a 74.6% success rate in generating accurate differential diagnoses from clinical vignettes [11]. The European Exam in Core Cardiology (EECC) is a rigorous postgraduate examination required for the completion of specialty training in cardiology. It consists of 120 multiple-choice questions (MCQs) covering pathophysiology, clinical reasoning, and guideline-recommended medical management, with a typical pass mark of around 60%. In a study, ChatGPT was tested on the EECC using 362 publicly available MCQs, achieving an overall accuracy of 58.8%, suggesting that it has acquired a broad knowledge base in cardiovascular medicine similar to cardiologists in-training [12]. These findings highlight ChatGPT’s potential as both an educational tool and a decision-support system in clinical settings [13]. Comparative analyses with other LLMs, such as Google’s Bard and Meta’s Large Language Model Meta AI (LLaMA), indicate that ChatGPT provides more comprehensive and accurate responses to medical inquiries and tasks [5,14]. Research also shows that ChatGPT provides more accurate, complete, and reproducible answers to home blood pressure monitoring knowledge compared to Bing [15], indicating its superior reliability and value for patients seeking information. This superiority is largely attributed to ChatGPT’s extensive training on diverse datasets and the implementation of RLHF, which enhances the model’s ability to generate credible and contextually appropriate outputs [5]. Furthermore, fine-tuning models such as ClinicalGPT with domain-specific data has exhibited markedly enhanced performance in specialized medical tasks, reflecting a higher level of expertise and precision [16]. Additionally, there is a significant increase in the number of medical professionals using ChatGPT, paralleled by a growing trend of patients turning from traditional search engines like Google to ChatGPT [17]. Compared to traditional search engines, ChatGPT offers a superior interactive experience, employs easily understandable language, saves time, and reduces information overload [18].
In conclusion, the evolution of LLMs has greatly advanced multiple fields, with ChatGPT leading due to its exceptional performance, wide applicability, low entry barriers, and high acceptance. Its impact is particularly significant in medicine [19]. As CVDs remain the most widespread and severe health threat globally, ChatGPT’s ability to enhance health education and clinical management in this area is increasingly acknowledged.

3. Enhancement of Clinical Decision-Making and Diagnostics in Cardiovascular Medicine with ChatGPT

ChatGPT demonstrates considerable potential in revolutionizing clinical decision-making and diagnostics in cardiovascular medicine by leveraging its ability to analyze patient-reported symptoms, assess risk factors, and support differential diagnoses through extensive medical knowledge databases. Its capabilities in real-time data collection and emergency decision-making can enhance diagnostic accuracy and expedite treatment plans. By facilitating early identification of issues, suggesting specialist referrals, and improving remote consultations, ChatGPT has the potential to significantly elevate patient care outcomes. These attributes underscore its promise in augmenting the diagnostic process and clinical decision-making in cardiovascular diseases [20].
A study evaluated ChatGPT’s accuracy in answering cardiovascular trivia questions and providing recommendations for clinical case vignettes [21]. The research involved testing ChatGPT with 50 trivia questions and 20 clinical cases, comparing its responses to expert opinions. ChatGPT correctly answered 92% of trivia questions and matched the actual clinical advice in 85% of case vignettes, with substantial improvement over earlier versions. The findings suggest that ChatGPT has substantial potential as an AI-assisted decision support tool, particularly for straightforward medical questions. Adil Salihu et al. evaluated the ability of ChatGPT-4 to improve clinical decision-making and enhance the efficiency of Heart Teams (HTs) in managing severe aortic stenosis [22]. Data from 150 patients presented at HT meetings were retrospectively analyzed using a standardized multiple-choice questionnaire with 14 key variables. The AI-generated decisions were compared to those made by the HT, with ChatGPT showing a 77% overall agreement. Specifically, agreement rates were 90% for transcatheter valve implantation, and 65% for both surgical valve replacement and medical treatment. The study concluded that ChatGPT could enhance HT decision-making by acting as a failsafe and identifying potential discrepancies [22]. By potentially serving as a second opinion, ChatGPT may improve decision-making efficiency, ensuring comprehensive patient evaluations and enhancing overall care quality in the management of valvular heart diseases. Another study employed a simulated cardiology consultation by asking ChatGPT ten hypothetical questions related to typical clinical scenarios of cardiovascular disorders [23]. The responses were evaluated by medical specialists for accuracy and applicability. Out of the ten scenarios, ChatGPT correctly diagnosed eight and provided appropriate, albeit general, management plans for all. This suggests that ChatGPT can assist in symptom analysis, risk assessment, and preliminary management planning. However, the study also highlighted limitations, such as the need for detailed clinical information and the inability to personalize treatment plans fully [23]. A retrospective study compared the diagnostic accuracy of ChatGPT to that of resident physicians in an emergency department (ED) setting [24]. The study involved 100 adults admitted to the ED with internal medicine issues. Diagnoses made by ED resident physicians and ChatGPT models were compared against final hospital discharge diagnoses using a point system for grading accuracy. Results showed that GPT-4 outperformed both GPT-3.5 (P<.001) and ED resident physicians (P=.01) in diagnostic accuracy. GPT-4 demonstrated significant superiority in diagnosing cardiovascular (P=.03) and endocrine or gastrointestinal diseases (P=.01) compared to resident physicians and GPT-3.5, respectively. Although the study was retrospective and had a limited sample size, the findings underscore the potential of GPT-4 as a supportive diagnostic tool in ED settings. Moreover, a study investigated the potential of ChatGPT in complementing patient-centered care for heart failure (HF) [25]. Using 30 questions related to HF diagnosis, management, and prognosis derived from online patient forums and physician experience, the responses from ChatGPT were evaluated by two independent HF experts. ChatGPT provided accurate answers in 90% of cases and exhibited a high consistency rate of 93%. Concordance between the two experts’ evaluations was 83%. The study concludes that ChatGPT holds significant promise in enhancing HF diagnosis and patient empowerment.
ChatGPT’s capabilities have extended to providing significant support in clinical pharmacy practices, demonstrating its potential to assist clinicians in medication decision-making and management. A study aimed to evaluate ChatGPT’s performance in key domains of clinical pharmacy practice, including prescription review, patient medication education, adverse drug reaction (ADR) recognition, ADR causality assessment, and drug counseling [26]. Real clinical cases and clinical pharmacist competency assessments were used to collect questions and answers, which were then inputted into ChatGPT. Five licensed clinical pharmacists independently rated the responses on a scale from 0 (completely incorrect) to 10 (completely correct). The mean scores of ChatGPT and clinical pharmacists were compared. ChatGPT performed well in drug counseling (8.77 vs. 9.50, P = .0791) but showed weaknesses in prescription review (5.23 vs. 9.90, P = .0089), patient medication education (6.20 vs. 9.07, P = .0032), ADR recognition (5.07 vs. 9.70, P = .0483), and ADR causality assessment (4.03 vs. 9.73, P = .023) compared to the pharmacists. Overall, while ChatGPT shows potential as a supplementary tool in drug counseling, it requires substantial improvements to handle complex tasks effectively. Another study further supports ChatGPT’s utility in polypharmacy management for geriatric care. The study evaluated ChatGPT’s performance in deprescribing decisions using standardized clinical vignettes [27]. ChatGPT’s responses demonstrated a consistent internal logic similar to general practitioners, effectively managing polypharmacy by recommending deprescribing based on patients’ activities of daily living status and CVD history. These findings suggest that ChatGPT can provide valuable clinical support in polypharmacy management, highlighting its potential to assist primary care physicians in managing complex medication regimens.
ChatGPT has demonstrated significant potential to enhance diagnostic processes across key cardiovascular diagnostic methods, including echocardiography, electrocardiography (ECG), and cardiovascular magnetic resonance (CMR). ChatGPT has been employed to enhance diagnostic accuracy and interpretation in generating echocardiographic reports [28]. It significantly accelerates the reporting process by producing detailed and concise descriptions of cardiac conditions, summarizing data, and recommending further tests or treatments. By translating complex medical findings into layman’s terms for patients while maintaining technical accuracy for healthcare professionals, ChatGPT improves communication. Furthermore, it ensures consistency and standardization in report generation, reducing variability and subjectivity, which is crucial for high-quality patient care and clinical outcomes [28]. A study evaluated the diagnostic accuracy of ChatGPT-4 in assessing ECG data, comparing its performance to that of emergency medicine specialists and cardiologists [29]. Using 40 ECG cases, ChatGPT-4 demonstrated superior performance in everyday ECG questions compared to both emergency medicine specialists and cardiologists. In more challenging ECG questions, ChatGPT-4 outperformed emergency medicine specialists but performed similarly to cardiologists. Overall, ChatGPT-4 was more successful than both groups in total ECG question accuracy. These findings demonstrate ChatGPT-4’s superior diagnostic accuracy in ECG assessment compared to medical specialists and underscore its potential to enhance clinical decision-making. ChatGPT can also enhance clinical diagnostics by transforming complex CMR reports into more understandable text, thereby improving communication between clinicians and patients. A study evaluated the performance of ChatGPT-4 in simplifying 20 CMR reports, generating three versions of each report for layperson comprehension [30]. Two cardiovascular radiologists assessed the factual correctness, completeness, and lack of potential harm, while 13 laypersons evaluated understandability using a Likert scale. The results showed that ChatGPT-4-generated reports were significantly easier to understand with a lower Automated Readability Index score compared to original reports. Radiologists rated the ChatGPT-4 reports highly for correctness, completeness, and safety. The study concluded that ChatGPT-4 can reliably simplify CMR reports, ensuring patients receive clear and accurate information. This improved clarity can enhance patient understanding and engagement, leading to more informed clinical decisions and better adherence to treatment plans, ultimately improving the cardiovascular care.
While ChatGPT demonstrates significant potential in clinical decision-making and diagnostics, it also encounters several challenges, including inaccuracies, hallucinations, and limitations influenced by language and context. A study evaluating ChatGPT’s performance in diagnosing retinal vascular diseases using 1226 fundus fluorescein angiography reports in Chinese highlighted the impact of language on diagnostic effectiveness [31]. Results indicated higher performance with English prompts (F1-score of 80.05%) compared to Chinese prompts (F1-score of 70.47%), though both were lower than ophthalmologists (89.35%) but close to interns (82.69%). English prompts also produced more accurate and complete reasoning steps with fewer errors, underscoring the need for robust models in non-English clinical applications. Moreover, ChatGPT’s application in cardiovascular diagnostics carries significant risks. A case report detailed a delayed diagnosis of a transient ischemic attack due to reliance on ChatGPT’s erroneous evaluation [32]. A 63-year-old man consulted ChatGPT after pulmonary vein isolation, leading to misclassification of symptoms and a 24-hour delay in seeking urgent medical attention. This case emphasizes the critical need for caution and further scrutiny when integrating AI into clinical practice, as errors may result in life-threatening consequences. Additionally, a study on ChatGPT’s performance in complex clinical scenarios involving the American Heart Association’s Advanced Cardiovascular Life Support guidelines found suboptimal accuracy in following the guidelines for bradycardia and cardiac arrest [33]. Across 20 simulation attempts, ChatGPT’s median accuracy was 69% for cardiac arrest and 42% for bradycardia. The study identified variability in the outputs of recommended steps, persistent omissions of critical steps, repetitive overemphasis of specific actions like checking heart rhythm and resuming cardiopulmonary resuscitation (CPR), and erroneous medication information. These findings highlight the need for improvements to ensure consistent and reliable guidance from ChatGPT in clinical settings.
In conclusion, while ChatGPT shows considerable promise in enhancing clinical decision-making and diagnostics in cardiovascular medicine, it also faces challenges such as inaccuracies and language-dependent limitations. Addressing these issues is crucial to fully realize its potential as a reliable and effective tool in improving patient outcomes and supporting clinicians in complex medical scenarios.

4. Innovative Medical Education for Cardiovascular Professionals and Patients with ChatGPT

The integration of advanced ChatGPT in the CVD field has the potential to revolutionize both professional training and patient education. By providing real-time, accurate, and easily understandable information, ChatGPT bridges the gap between complex medical knowledge and practical application. Its capacity to interpret extensive guidelines, deliver precise clinical answers, and simplify medical jargon ensures that healthcare professionals are well-informed and patients are better equipped to manage their health.
ChatGPT has already played a significant role in medical professional education and has gained considerable acceptance. An exploratory survey involving 844 participants from various academic and healthcare roles, with a response rate of 49.7%, revealed that 40% of the audience had tried ChatGPT, with more trainees having used it compared to faculty [34]. Those who had used ChatGPT expressed greater interest in utilizing it across a wider range of contexts in the future. The survey highlighted differing perspectives based on roles, with trainees showing more enthusiasm for its use in education, healthcare, and research. Despite some uncertainty, particularly in educational contexts, the broad acceptance and interest indicate that ChatGPT is already making a substantial impact on medical professional education and training. ChatGPT can significantly aid in the education and training of medical professionals by providing accurate and concise explanations of complex and extensive guidelines, enhancing their understanding and management. A study evaluated ChatGPT’s ability to answer frequently asked questions (FAQs) and guideline-based questions about acute coronary syndromes (ACS) using the 2023 European Society of Cardiology Guidelines [35]. The accuracy and reproducibility of ChatGPT’s responses were assessed by two cardiologists with ten years of experience. The results showed that 90.3% of ChatGPT’s answers to FAQs scored the highest accuracy and proficiency, with no responses scoring the lowest. For guideline-based questions, 88.3% of responses achieved the highest score. ChatGPT’s answers were also highly reproducible, with 94.4% for FAQs and 90.9% for guideline questions. A cross-sectional study assessed the accuracy and comprehensiveness of chatbot-generated responses to physician-developed medical queries across 17 specialties [36]. Thirty-three physicians generated 284 questions, classified as easy, medium, or hard. ChatGPT responses were scored for accuracy and completeness using Likert scales. The median accuracy score was 5.5 out of 6, and the median completeness score was 3 out of 3. Accuracy scores improved significantly over time and between different versions of the chatbot (GPT-3.5 and GPT-4). This study demonstrated ChatGPT’s high accuracy and capability to provide largely accurate and comprehensive answers to diverse medical queries, underscoring its potential as a valuable educational tool for medical professionals. Given these findings, ChatGPT can be integrated into medical education to provide precise answers to clinical questions, thereby enhancing learning, and facilitating continuous education and professional development. Furthermore, ChatGPT has the potential to provide information on CPR by generating tailored responses based on CPR guidelines [37]. ChatGPT’s responses were compared to traditional methods of obtaining CPR information. The AI demonstrated the ability to quickly provide accurate information, aiding healthcare professionals in emergencies. This study underscores ChatGPT’s potential as an educational tool for enhancing CPR training and emergency response techniques, crucial for preparing healthcare professionals to handle cardiac emergencies effectively.
ChatGPT can enhance patient education by providing accurate, accessible, and empathetic health information. By simplifying medical jargon, it helps patients better understand their conditions and treatment options, leading to improved health literacy, adherence to treatment plans, and informed decision-making. A study evaluated ChatGPT’s ability to provide quality and empathetic responses to patient questions on a public social media forum, comparing its performance to that of verified physicians [38]. Using 195 randomly selected patient-physician exchanges, ChatGPT’s responses were rated higher in both quality and empathy by a team of licensed healthcare professionals. This research highlights ChatGPT’s ability to simplify complex medical information, demonstrating effective patient communication and personalized health education. Building on this, another pilot study investigated whether ChatGPT could improve the readability of online aortic stenosis patient education materials [39]. The study gathered 21 patient education materials from professional cardiothoracic surgical societies and academic institutions in the USA. These materials, originally written at a 10th-12th grade reading level, were inputted into ChatGPT with the prompt “translate to 5th-grade reading level.” The results showed that ChatGPT-3.5 successfully improved readability to the 6th-7th grade level across all measures, making complex medical information more accessible and supporting better patient understanding and engagement in managing cardiovascular health. Moreover, ChatGPT has demonstrated its ability to offer comprehensive guidance on hypertension management [40]. Further emphasizing ChatGPT’s potential, a study evaluated ChatGPT’s effectiveness in addressing queries related to lifestyle-related diseases and disorders, enhancing patient engagement and health literacy [41]. Twenty cases, each with four questions, were presented to ChatGPT, and two physicians rated the responses for accuracy and guidance. The study concluded that ChatGPT provides reasonably accurate and useful information, highlighting its potential as a virtual telehealth agent for lifestyle-related diseases. A recent study also confirmed that GPT-4 has shown promising potential in automated medical consultation for cardiovascular diseases, with medical accuracy comparable to human experts [42]. Additionally, ChatGPT’s capability to answer laypeople’s questions about cardiac arrest and CPR was assessed in a comprehensive study [43]. The answers provided by ChatGPT were evaluated by professionals and laypeople for accuracy, relevance, clarity, comprehensibility, and overall value. The study found that ChatGPT’s answers received high ratings for clarity, relevance, accuracy, and comprehensiveness, indicating its potential to enhance public medical education about cardiac arrest.
In summary, ChatGPT has shown significant promise in both professional and patient education by providing accurate, accessible, and comprehensive health information. Its potential to enhance medical education and training, as well as to support patient engagement and health literacy, underscores its importance in the future of healthcare.

5. Advancements in Cardiovascular Medical Research and Scholarly Communication with ChatGPT

Generative AI tools like ChatGPT have emerged as pivotal assets in cardiovascular research and academic writing. These advanced language models are transforming how researchers collect, analyze, and interpret data, as well as how they communicate their findings. By enhancing predictive modeling, facilitating data analysis, automating literature reviews, and streamlining the writing process, ChatGPT is significantly advancing the efficiency and accuracy of cardiovascular research and scholarly communication.
ChatGPT-4 has the potential to streamline automated machine learning for clinical studies, as its Advanced Data Analysis feature can make machine learning more accessible in medicine by simplifying complex data analyses [44]. Using data from 2,142 patients and controls to identify those at risk for cardiac amyloidosis, ChatGPT-4 autonomously selected and optimized machine learning models, employing techniques such as Gradient Boosting Machine and Random Forest classifiers. It utilized data preprocessing strategies like median imputation and standard scaling, followed by 5-fold cross-validation for model validation. The model achieved an Area Under the Receiver Operating Characteristic curve of 0.954, indicating high diagnostic accuracy [44]. This highlights ChatGPT-4’s significant value in advancing medical research and clinical practice through efficient and precise data analysis. Mohammed Ali et al. conducted a study aimed to explore the influence of the Social Vulnerability Index (SVI) and Digital Literacy (DL) on CVD mortality across US counties using ChatGPT-4 [45]. The researchers integrated data from the Centers for Disease Control and Prevention Wide-ranging Online Data for Epidemiologic Research and the US Census Bureau, identifying significant correlations between the SVI, DL, and CVD mortality through regression modeling. The findings highlighted that incorporating DL and SVI improves the accuracy of mortality predictions, underscoring the importance of these variables in public health research. ChatGPT-4 played a crucial role by identifying relevant variables, generating Python code for data analysis, formulating hypotheses, creating predictive models, and interpreting results, thereby demonstrating its efficiency in handling complex datasets and providing valuable insights. Moreover, ChatGPT has shown significant potential in cardiovascular imaging research [46]. Utilizing its data analyst functionality, ChatGPT enables researchers to handle descriptive statistics, intergroup analysis, and correlation analysis with high consistency and superior analytical efficiency compared to traditional biostatistical software like SAS, SPSS, and R. By generating Python code for data integration, hypothesis formulation, and statistical tests, ChatGPT allows researchers with limited programming skills to perform complex data analyses. Its capabilities in automatic data cleaning and visualization streamline the research process, enhancing focus on data interpretation. This functionality demonstrates ChatGPT’s ability to lower operational barriers, making sophisticated data analysis accessible to a broader range of researchers, thereby advancing the field of cardiovascular research. Additionally, ChatGPT significantly enhances research efficiency by automating data extraction processes. For instance, ChatGPT was utilized to extract procedural details from free-text reports on mechanical thrombectomy in patients with ischemic stroke [47]. The research compared the performance of GPT-4 and GPT-3.5 in accurately extracting data from these reports. ChatGPT-4 demonstrated a higher accuracy rate, successfully extracting 94.0% of data points from internal reports and 90.5% from external reports without the need for further postprocessing. Thus, ChatGPT can significantly reduce manual labor and minimize errors in data collection for prospective studies. Furthermore, ChatGPT has proven effective in automating the screening process for systematic reviews and meta-analyses in medical research. A study evaluated ChatGPT’s efficacy in screening 1,198 abstracts from three radiology subfields [48]. Researchers compared ChatGPT’s performance to that of general physicians (GPs) using metrics such as sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and workload savings. ChatGPT completed the screening within an hour, achieving a sensitivity of 95% and an NPV of 99%, slightly outperforming the GPs’ consensus. Additionally, ChatGPT demonstrated low false-negative counts and high workload savings (40-83%). Although ChatGPT had lower specificity and PPV compared to human raters, its high sensitivity and efficiency suggest that it could be a valuable first-line screening tool. Another study confirmed these findings, comparing ChatGPT to human researchers in identifying relevant studies for systematic reviews on mHealth interventions for medication adherence in ischemic stroke patients [49]. This study found that ChatGPT significantly reduced the time required for study identification, demonstrating substantial efficiency. Although human researchers showed higher precision (0.86 vs. 0.77) and relevance percentage (9.8% vs. 3%), ChatGPT identified studies much faster and retrieved a higher number of relevant studies overall.
Generative AI tools, including ChatGPT, have become widely used in the field of academic writing. A recent study reviewed the guidelines of the top 25 cardiology and cardiovascular medicine journals regarding the use of generative AI in scientific writing [50]. The study found that all journals permitted the use of AI tools with specific limitations, such as prohibiting AI from being listed as authors or used for image generation and requiring all authors to take full responsibility for their work. The use of AI in the peer review process is strictly prohibited. Researchers can use ChatGPT for writing as long as they take intellectual ownership and ethical responsibility for its content [51]. AI significantly aids the writing process [52]. For instance, ChatGPT played a pivotal role in a manuscript investigating the long-term administration of proton pump inhibitors (PPIs) and their potential link to adverse cardiovascular outcomes [53]. This research, which reviewed existing systematic reviews and meta-analyses, aimed to determine if prolonged PPI use is associated with major adverse cardiovascular events. ChatGPT was instrumental in generating search strings, screening titles and abstracts, extracting data, summarizing studies, and performing qualitative synthesis and risk of bias assessments. By automating and streamlining these stages, ChatGPT enhanced the efficiency and consistency of the writing process. The study demonstrated that ChatGPT holds significant promise as a tool for evidence synthesis and improving the quality and speed of academic manuscript creation. Furthermore, a case report on the diagnosis of ventricular septal rupture using point-of-care ultrasound in a patient with a history of acute myocardial infarction applied ChatGPT for data collection, literature review, and grammar correction [54]. The authors demonstrated that ChatGPT effectively assisted in gathering precise data, conducting comprehensive literature reviews, and ensuring grammatical accuracy. This highlights ChatGPT’s utility in medical case report writing by enhancing the efficiency and quality of the documentation process.
In summary, ChatGPT has proven to be a transformative tool in cardiovascular research and academic writing. The incorporation of ChatGPT in these areas not only improves efficiency and accuracy but also ensures high-quality and reliable scientific outputs, ultimately contributing to the advancement of cardiovascular medicine.

6. Challenges and Future Directions in Integrating ChatGPT into Cardiovascular Medicine

As discussed previously, ChatGPT has revolutionized cardiovascular medicine but faces challenges, including common digital-era issues like privacy and data security, as well as unique problems specific to its application.
Despite ChatGPT being one of the most advanced large language models in medical applications, a significant concern remains its potential for inaccuracies and hallucinations—instances where the model generates information that appears plausible but is entirely fabricated or incorrect. As highlighted in previous sections and supported by various studies, the error rates in different medical scenarios can range from 5% to 50%, presenting a substantial risk. While human errors are also prevalent, the mistakes made by ChatGPT are often harder to detect and correct, even fooling scientists [55]. Furthermore, ChatGPT’s errors can compound over time, as subsequent answers may build on its previous inaccuracies, leading to increasingly erroneous and potentially harmful outputs. This cascading effect poses a serious threat to patient safety and the overall integrity of medical practice. A notable example is the use of AI-generated imagery from DALL·E, another OpenAI tool, in CVD education. A study evaluating DALL·E images for congenital heart diseases found significant inaccuracies: 80.8% of images were anatomically incorrect, 85.2% had incorrect text labels, and 78.1% were deemed unusable for educational purposes [56]. This example underscores the broader challenges faced by AI applications in producing accurate and reliable content, further validating concerns about the potential risks associated with AI-generated outputs in medical contexts.
In addition to accuracy concerns, another significant issue is the fairness and potential biases in AI-generated content. A study examining gender and racial biases in content produced by seven large language models, including ChatGPT, revealed substantial biases [57]. Although ChatGPT demonstrated the lowest level of bias among the models and was the only one capable of declining to generate content when provided with biased prompts, this finding underscores the need for ongoing improvements in mitigating bias.
Furthermore, the application of ChatGPT in the medical field raises significant ethical risks and challenges [58]. A study demonstrated that ChatGPT lacks a consistent moral stance, often providing contradictory advice on the same ethical issue [59]. This inconsistency can influence the moral judgments of users, who may not fully recognize the extent of this impact. Such influence is particularly concerning in medicine, where ethical reasoning is foundational for healthcare professionals and students. Additionally, ChatGPT’s increasing influence in medical research and academic writing has heightened concerns regarding academic integrity and ethics. As reliance on such models for drafting papers grows, issues such as authorship, plagiarism, and the generation of incorrect scientific information come to the forefront. Addressing these ethical challenges is essential to ensure that ChatGPT’s use does not compromise the moral standards critical to both medical practice and academic integrity.
Lastly, the use of ChatGPT in medical education raises concerns about its impact on critical thinking [5]. While LLMs can aid learning, over-reliance on them may hinder students’ or professionals’ ability to evaluate information critically, a skill vital for distinguishing valuable input from irrelevant or incorrect data. This reliance can extend to medical reasoning, risking a decline in analytical and decision-making skills. Transparent regulation of LLM use in educational institutions is essential to preserve and foster critical thinking abilities.
By addressing these challenges, we can better integrate ChatGPT into medicine. Previous discussions have highlighted potential solutions [60,61]. This paper advances these ideas by proposing specific directions for future research and development to enhance the utility, safety, and ethical application of ChatGPT in cardiovascular medicine.
1. Enhancing Training Data Quality and Real-Time Knowledge Integration:
Improving ChatGPT’s performance in cardiovascular tasks requires incorporating high-quality, domain-specific texts during training, such as clinical notes, patient medical records, surgical reports, diagnostic results, treatment guidelines, and peer-reviewed literature. Expanding the training datasets to include more specialized medical knowledge and ensuring coverage of diverse countries, languages, ethnicities, and religions can reduce biases and enable ChatGPT to provide more accurate and relevant recommendations. Additionally, enabling the model to source real-time information from the internet and continuously update its knowledge base will ensure it remains current with the latest advancements in cardiovascular care.
2. Exploring Customized and Function-Specific Models:
Developing customized ChatGPT models tailored to specific subfields within cardiology is a promising research direction. This can be achieved by creating disease-specific models for conditions like atrial fibrillation and heart failure, as well as function-specific models for tasks such as medication guidance and secondary prevention advice. Customization allows the AI to focus on narrower data sets and contexts, reducing complexity and potential errors. Advanced research into prompt engineering, which involves refining the way questions and requests are posed to the AI, can also significantly improve the model’s accuracy and effectiveness [62]. By tailoring ChatGPT to specific diseases and functions and employing effective prompt engineering strategies, its application can become more targeted, reliable, and efficient in cardiovascular care.
3. Developing Uncertainty Indicators and Enhancing Human Oversight:
In cardiovascular care, it is crucial to have tools that indicate the confidence level of AI recommendations, helping users discern when to trust the AI’s advice and when to exercise caution. Clear signals about the reliability of AI-generated content ensure that ChatGPT is used as a helpful assistant rather than the sole decision-maker. To prevent over-reliance on AI and maintain high-quality clinical decisions, rigorous human oversight is essential. Experts can review initial AI recommendations, use these reviews to train the AI and balance efficiency with accuracy. Moreover, future development should establish strict regulations and legal frameworks to ensure human oversight and accountability. Verification by cardiology experts is indispensable in mitigating errors and validating accuracy, with clear accountability for AI-generated outputs resting with human professionals to ensure reliability and safety in clinical settings.
4. Conducting Rigorous and Expert-Supervised Clinical Trials:
Conducting high-quality, real-world clinical trials to validate ChatGPT’s use in cardiovascular care is essential. These trials should be meticulously designed and conducted under strict human and expert supervision to ensure the reliability and safety of ChatGPT-assisted interventions. Evaluating the effectiveness, acceptance, and practicality of these interventions in clinical practice is crucial. Key endpoints may include reductions in mortality and morbidity, improvements in treatment adherence, and enhanced patient satisfaction. Rigorous, well-supervised clinical research will provide the necessary evidence to assess ChatGPT’s true impact on cardiovascular disease management.
5. Establishing Ethical Guidelines and Responsibility:
Clear ethical guidelines for the use of ChatGPT are necessary, addressing issues of authorship, plagiarism, and accountability for AI-generated content. Transparency in the use of AI in research, clinical practice, and publications is crucial to prevent misuse and ensure ethical compliance. Scholars have suggested key recommendations, including prohibiting ChatGPT from being cited as an author, ensuring that users possess a basic understanding of the models, and restricting these models from generating the entirety of the manuscript text [63]. Proper acknowledgment of AI use is essential to maintain ethical standards and integrity in academic writing. These guidelines should be discussed and unified to form a broad consensus for effective management, thereby upholding the credibility and ethical standards of work involving ChatGPT.
By focusing on these directions, the research and development of ChatGPT can be aligned with the specific needs and challenges of the cardiovascular domain, ensuring its safe, ethical, and effective application in improving patient care, advancing research, and supporting education in cardiology.

7. Conclusion

ChatGPT has become a transformative tool in cardiovascular medicine, significantly enhancing clinical decision-making, education, and research. Despite facing challenges such as accuracy and ethical considerations, there are clear pathways to address these issues. With ongoing advancements and targeted solutions, the future of ChatGPT in cardiovascular medicine is promising. This technology has already benefited many patients, healthcare professionals, and medical students, and as it continues to evolve, it will further enhance their capabilities and drive greater advancements in the field.

Author Contributions

Conceptualization, M.L., and Y.S.; writing—original draft preparation, M.L., and C.R.; writing—review and editing, M.L., C.R., G.P., D.B., M.S., and M.Q.F.and Y.S.; supervision, Y.S.; funding acquisition, M.L., Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Cardiovascular Health Innovation Fund (2024) from Stanford University School of Medicine.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

In this section, you can acknowledge any support given which is not covered by the author contribution or funding sections. This may include administrative and technical support, or donations in kind (e.g., materials used for experiments).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Varghese, C., et al., Artificial intelligence in surgery. Nat Med, 2024. 30(5): p. 1257-1268.
  2. Cerrato, P.L. and J.D. Halamka, How AI drives innovation in cardiovascular medicine. Front Cardiovasc Med, 2024. 11: p. 1397921. [CrossRef]
  3. Will ChatGPT transform healthcare? Nat Med, 2023. 29(3): p. 505-506.
  4. Meng, X., et al., The application of large language models in medicine: A scoping review. iScience, 2024. 27(5): p. 109713. [CrossRef]
  5. Clusmann, J., et al., The future landscape of large language models in medicine. Commun Med (Lond), 2023. 3(1): p. 141. [CrossRef]
  6. Tsao, C.W., et al., Heart Disease and Stroke Statistics-2023 Update: A Report From the American Heart Association. Circulation, 2023. 147(8): p. e93-e621. [CrossRef]
  7. Singhal, K., et al., Large language models encode clinical knowledge. Nature, 2023. 620(7972): p. 172-180. [CrossRef]
  8. Wong, R.S., L.C. Ming, and R.A. Raja Ali, The Intersection of ChatGPT, Clinical Medicine, and Medical Education. JMIR Med Educ, 2023. 9: p. e47274. [CrossRef]
  9. Kris Jangjarat, T.K., Pongsakorn Limna, Rattaphong Sonsuphap Public perceptions towards ChatGPT a s the Robo -Assistant. Online Journal of Communication and Media Technologies, 2023. 13(3): p. e202338.
  10. Thirunavukarasu, A.J., et al., Large language models in medicine. Nat Med, 2023. 29(8): p. 1930-1940. [CrossRef]
  11. Shieh, A., et al., Assessing ChatGPT 4.0’s test performance and clinical diagnostic accuracy on USMLE STEP 2 CK and clinical case reports. Sci Rep, 2024. 14(1): p. 9330.
  12. Skalidis, I., et al., ChatGPT takes on the European Exam in Core Cardiology: an artificial intelligence success story? Eur Heart J Digit Health, 2023. 4(3): p. 279-281.
  13. Kung, T.H., et al., Performance of ChatGPT on USMLE: Potential for AI-assisted medical education using large language models. PLOS Digit Health, 2023. 2(2): p. e0000198. [CrossRef]
  14. Sandmann, S., et al., Systematic analysis of ChatGPT, Google search and Llama 2 for clinical decision support tasks. Nat Commun, 2024. 15(1): p. 2050. [CrossRef]
  15. Niko, M.M., et al., Comparing ChatGPT and Bing, in response to the Home Blood Pressure Monitoring (HBPM) knowledge checklist. Hypertens Res, 2024. 47(5): p. 1401-1409. [CrossRef]
  16. Guangyu Wang, G.Y., Zongxin Du, Longjun Fan, Xiaohu Li, ClinicalGPT: Large Language Models Finetuned with Diverse Medical Data and Comprehensive Evaluation. arXiv preprint 2023. arXiv:2306.09968.
  17. Van Bulck, L. and P. Moons, What if your patient switches from Dr. Google to Dr. ChatGPT? A vignette-based survey of the trustworthiness, value, and danger of ChatGPT-generated responses to health questions. Eur J Cardiovasc Nurs, 2024. 23(1): p. 95-98.
  18. Dave, T., S.A. Athaluri, and S. Singh, ChatGPT in medicine: an overview of its applications, advantages, limitations, future prospects, and ethical considerations. Front Artif Intell, 2023. 6: p. 1169595. [CrossRef]
  19. Lee, P., S. Bubeck, and J. Petro, Benefits, Limits, and Risks of GPT-4 as an AI Chatbot for Medicine. N Engl J Med, 2023. 388(13): p. 1233-1239. [CrossRef]
  20. Chlorogiannis, D.D., et al., The Role of ChatGPT in the Advancement of Diagnosis, Management, and Prognosis of Cardiovascular and Cerebrovascular Disease. Healthcare (Basel), 2023. 11(21). [CrossRef]
  21. Harskamp, R.E. and L. De Clercq, Performance of ChatGPT as an AI-assisted decision support tool in medicine: a proof-of-concept study for interpreting symptoms and management of common cardiac conditions (AMSTELHEART-2). Acta Cardiol, 2024. 79(3): p. 358-366.
  22. Salihu, A., et al., A study of ChatGPT in facilitating Heart Team decisions on severe aortic stenosis. EuroIntervention, 2024. 20(8): p. e496-e503.
  23. Rizwan, A. and T. Sadiq, The Use of AI in Diagnosing Diseases and Providing Management Plans: A Consultation on Cardiovascular Disorders With ChatGPT. Cureus, 2023. 15(8): p. e43106. [CrossRef]
  24. Hoppe, J.M., et al., ChatGPT With GPT-4 Outperforms Emergency Department Physicians in Diagnostic Accuracy: Retrospective Analysis. J Med Internet Res, 2024. 26: p. e56110.
  25. Kozaily, E., et al., Accuracy and consistency of online large language model-based artificial intelligence chat platforms in answering patients’ questions about heart failure. Int J Cardiol, 2024. 408: p. 132115. [CrossRef]
  26. Huang, X., et al., Evaluating the performance of ChatGPT in clinical pharmacy: A comparative study of ChatGPT and clinical pharmacists. Br J Clin Pharmacol, 2024. 90(1): p. 232-238. [CrossRef]
  27. Rao, A., et al., Proactive Polypharmacy Management Using Large Language Models: Opportunities to Enhance Geriatric Care. J Med Syst, 2024. 48(1): p. 41. [CrossRef]
  28. Kusunose, K., Revolution of echocardiographic reporting: the new era of artificial intelligence and natural language processing. J Echocardiogr, 2023. 21(3): p. 99-104. [CrossRef]
  29. Gunay, S., et al., Comparison of emergency medicine specialist, cardiologist, and chat-GPT in electrocardiography assessment. Am J Emerg Med, 2024. 80: p. 51-60. [CrossRef]
  30. Salam, B., et al., Generative Pre-trained Transformer 4 makes cardiovascular magnetic resonance reports easy to understand. J Cardiovasc Magn Reson, 2024. 26(1): p. 101035. [CrossRef]
  31. Liu, X., et al., Uncovering Language Disparity of ChatGPT on Retinal Vascular Disease Classification: Cross-Sectional Study. J Med Internet Res, 2024. 26: p. e51926.
  32. Saenger, J.A., et al., Delayed diagnosis of a transient ischemic attack caused by ChatGPT. Wien Klin Wochenschr, 2024. 136(7-8): p. 236-238.
  33. Pham, C., et al., ChatGPT’s Performance in Cardiac Arrest and Bradycardia Simulations Using the American Heart Association’s Advanced Cardiovascular Life Support Guidelines: Exploratory Study. J Med Internet Res, 2024. 26: p. e55037. [CrossRef]
  34. Hosseini, M., et al., An exploratory survey about using ChatGPT in education, healthcare, and research. PLoS One, 2023. 18(10): p. e0292216.
  35. Gurbuz, D.C. and E. Varis, Is ChatGPT knowledgeable of acute coronary syndromes and pertinent European Society of Cardiology Guidelines? Minerva Cardiol Angiol, 2024. 72(3): p. 299-303.
  36. Goodman, R.S., et al., Accuracy and Reliability of Chatbot Responses to Physician Questions. JAMA Netw Open, 2023. 6(10): p. e2336483.
  37. Ahn, C., Exploring ChatGPT for information of cardiopulmonary resuscitation. Resuscitation, 2023. 185: p. 109729.
  38. Ayers, J.W., et al., Comparing Physician and Artificial Intelligence Chatbot Responses to Patient Questions Posted to a Public Social Media Forum. JAMA Intern Med, 2023. 183(6): p. 589-596. [CrossRef]
  39. Rouhi, A.D., et al., Can Artificial Intelligence Improve the Readability of Patient Education Materials on Aortic Stenosis? A Pilot Study. Cardiol Ther, 2024. 13(1): p. 137-147.
  40. Layton, A.T., AI, Machine Learning, and ChatGPT in Hypertension. Hypertension, 2024. 81(4): p. 709-716.
  41. Mondal, H., et al., ChatGPT in Answering Queries Related to Lifestyle-Related Diseases and Disorders. Cureus, 2023. 15(11): p. e48296. [CrossRef]
  42. Jo, E., et al., Assessing GPT-4’s Performance in Delivering Medical Advice: Comparative Analysis With Human Experts. JMIR Med Educ, 2024. 10: p. e51282. [CrossRef]
  43. Scquizzato, T., et al., Testing ChatGPT ability to answer laypeople questions about cardiac arrest and cardiopulmonary resuscitation. Resuscitation, 2024. 194: p. 110077. [CrossRef]
  44. Tayebi Arasteh, S., et al., Large language models streamline automated machine learning for clinical studies. Nat Commun, 2024. 15(1): p. 1603.
  45. Ali, M.M., et al., Mapping the Heartbeat of America with ChatGPT-4: Unpacking the Interplay of Social Vulnerability, Digital Literacy, and Cardiovascular Mortality in County Residency Choices. J Pers Med, 2023. 13(12). [CrossRef]
  46. Daibes M, Lima BB. Cracking the heart code: using ChatGPT’s Data Analyst feature for cardiovascular imaging research. Int J Cardiovasc Imaging. 2024 Apr 23. [CrossRef]
  47. Lehnen, N.C., et al., Data Extraction from Free-Text Reports on Mechanical Thrombectomy in Acute Ischemic Stroke Using ChatGPT: A Retrospective Analysis. Radiology, 2024. 311(1): p. e232741.
  48. Issaiy, M., et al., Methodological insights into ChatGPT’s screening performance in systematic reviews. BMC Med Res Methodol, 2024. 24(1): p. 78.
  49. Ruksakulpiwat, S., et al., Assessing the Efficacy of ChatGPT Versus Human Researchers in Identifying Relevant Studies on mHealth Interventions for Improving Medication Adherence in Patients With Ischemic Stroke When Conducting Systematic Reviews: Comparative Analysis. JMIR Mhealth Uhealth, 2024. 12: p. e51526. [CrossRef]
  50. Inam, M., et al., A review of top cardiology and cardiovascular medicine journal guidelines regarding the use of generative artificial intelligence tools in scientific writing. Curr Probl Cardiol, 2024. 49(3): p. 102387. [CrossRef]
  51. Basgier, C. and S. Sharma, Should scientists delegate their writing to ChatGPT? Nature, 2023. 624(7992): p. 523.
  52. Gruda, D., Three ways ChatGPT helps me in my academic writing. Nature, 2024.
  53. Teperikidis, E., et al., Does the long-term administration of proton pump inhibitors increase the risk of adverse cardiovascular outcomes? A ChatGPT powered umbrella review. Acta Cardiol, 2023. 78(9): p. 980-988.
  54. Lal Vallath, A., et al., Ventricular Septal Rupture and Artificial Intelligence (AI)-Assisted Healthcare. Cureus, 2023. 15(3): p. e36581.
  55. Else, H., Abstracts written by ChatGPT fool scientists. Nature, 2023. 613(7944): p. 423.
  56. Temsah, M.H., et al., Art or Artifact: Evaluating the Accuracy, Appeal, and Educational Value of AI-Generated Imagery in DALL.E 3 for Illustrating Congenital Heart Diseases. J Med Syst, 2024. 48(1): p. 54.
  57. Fang, X., et al., Bias of AI-generated content: an examination of news produced by large language models. Sci Rep, 2024. 14(1): p. 5224.
  58. Ong, J.C.L., et al., Ethical and regulatory challenges of large language models in medicine. Lancet Digit Health, 2024. 6(6): p. e428-e432. [CrossRef]
  59. Krügel S, Ostermaier A, Uhl M. ChatGPT’s inconsistent moral advice influences users’ judgment. Sci Rep. 2023 Apr 6;13(1):4569.
  60. van Dis, E.A.M., et al., ChatGPT: five priorities for research. Nature, 2023. 614(7947): p. 224-226.
  61. Ullah, E., et al., Challenges and barriers of using large language models (LLM) such as ChatGPT for diagnostic medicine with a focus on digital pathology - a recent scoping review. Diagn Pathol, 2024. 19(1): p. 43. [CrossRef]
  62. Mishra, V., et al., Evaluation of Prompts to Simplify Cardiovascular Disease Information Generated Using a Large Language Model: Cross-Sectional Study. J Med Internet Res, 2024. 26: p. e55388.
  63. Kim, J.K., et al., ChatGPT and large language model (LLM) chatbots: The current state of acceptability and a proposal for guidelines on utilization in academic medicine. J Pediatr Urol, 2023. 19(5): p. 598-604. [CrossRef]
Figure 1. ChatGPT in Cardiovascular Medicine. This graphical abstract summarizes the article’s overview of ChatGPT’s innovative applications in cardiovascular medicine, its persistent challenges, and future directions for optimizing its role and effectiveness.
Figure 1. ChatGPT in Cardiovascular Medicine. This graphical abstract summarizes the article’s overview of ChatGPT’s innovative applications in cardiovascular medicine, its persistent challenges, and future directions for optimizing its role and effectiveness.
Preprints 116489 g001
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

© 2024 MDPI (Basel, Switzerland) unless otherwise stated