REVIEW ARTICLE |
https://doi.org/10.5005/jp-journals-11002-0095 |
Artificial Intelligence in Newborn Medicine
1Department of Pediatrics, Louisiana State University, Shreveport, Louisiana, United States of America
2Global Newborn Society, Clarksville, Maryland, United States of America
3Banaras Hindu University Institute of Excellence, Varanasi, Uttar Pradesh, India
4Department of Radiology, Baylor College of Medicine, Houston, Texas, United States of America
Corresponding Author: Akhil Maheshwari, Department of Pediatrics, Louisiana State University, Shreveport, Louisiana, United States of America, Phone: +7089108729, e-mail: Akhil@globalnewbornsociety.org; Thierry AGM Huisman, Department of Radiology, Baylor College of Medicine, Houston, Texas, United States of America, Phone: +8328247237, e-mail: huisman@texaschildrens.org
How to cite this article: Maheshwari A, Huisman TAGM. Artificial Intelligence in Newborn Medicine. Newborn 2024;3(2):96–110.
Source of support: Nil
Conflict of interest: Dr Akhil Maheshwari and Dr Thierry AGM Huisman are associated as the Editorial Board Member of this journal and this manuscript was subjected to this journal’s standard review procedures, with this peer review handled independently of these Editorial Board Members and their research group.
Received on: 14 May 2024; Accepted on: 20 June 2024; Published on: 21 June 2024
ABSTRACT
The development of artificial intelligence (AI) algorithms has evoked a mixed-feeling reaction, a combination of excitement but also some trepidation, with reminders of caution coming up each time a novel AI-related academic/medical software program is proposed. There is awareness, with some hesitancy, that these algorithms could turn out to be a continuous, transformational source of clinical and educational information. Several AI algorithms with varying strengths and weaknesses are known, but the deep-learning pathways known as the Generative Pre-trained Transformers (GPT) have evoked the most interest as clinical decision-support systems. Again, these tools still need validation and all steps should undergo multiple checks and cross-checks prior to any implementation in human medicine. If, however, testing eventually confirms the utility of these pathways, there is a possibility of a non-incremental advancement of immense value. Artificial intelligence can be helpful by facilitating appropriate analysis of the large bodies of data that are available but are not being uniformly and comprehensively analyzed at all centers. It could promote appropriate, timely diagnoses, testing for efficacy with less bias, fewer diagnostic and medication errors, and good follow-up. Predictive modeling can help in appropriate allocation of resources by identifying at-risk newborns right at the outset. Artificial intelligence can also help develop information packets to engage and educate families. In academics, it can help in an unbiased, all-inclusive analysis of medical literature on a continuous basis for education and research. We know that there will be challenges in protection of privacy in handling data, bias in algorithms, and in regulatory compliance. Continued efforts will be needed to understand and streamline AI. However, if the medical community hesitates today in overseeing this juggernaut, the inclusion (or not) of AI in medicine might not stop—it might just gradually get extrapolated into patient care from other organizations/industry for cost reasons, not justification based on actual clinical data. If we do not get involved in this process to oversee the development/incorporation of AI in newborn medicine, the questions in making decisions will just change from who, to which, when, and how. Maybe this will not be the most appropriate scenario. To conclude, AI has definite benefits; we should embrace AI developments as valuable tools that can assist physicians in analyzing large and complex datasets, which will facilitate the identification of key facts/findings that might be missed if studied by humans. On the other hand, a well-designed and critical expert review board is mandatory to prevent AI-generated systematic errors.
Keywords: Critical, Generative pre-trained transformers, Neonate, Patient triage, Predictive modeling techniques, Premature, Resource allocation, Telemedicine consultations, Timely detection.
KEY POINTS
The development of artificial intelligence (AI) algorithms is evoking a mixed-feeling reaction, some excitement and some trepidation, as new medical possibilities emerge every day.
AI could strengthen clinical support systems to assist healthcare providers in timely detection of neonatal diseases, in making evidence-based decisions, reducing diagnostic errors, and in improving patient outcomes.
Predictive modeling techniques can enable the identification of at-risk newborns and the early intervention of complications, such as sepsis and neurological disorders.
In newborn health care, AI could help optimize resource allocation, patient triage, and telemedicine consultations, thereby enhancing access to medical expertise, particularly in underserved regions.
To reiterate, there is a need for extreme caution in evaluation of these programs. This could be a paradigm shift, and like those in the past, we need cautious multi/inter-disciplinary collaboration to test and if viable, develop it, not outrightly reject it.
Introduction
Artificial intelligence (AI) encompasses various techniques and approaches that can be applied to medicine to improve healthcare delivery, precision diagnosis, treatment, and patient outcomes.1 These tools show immense promise in transforming various aspects of newborn medicine, ranging from early detection and diagnosis of neonatal conditions to personalized precision treatment and long-term care.2 This article explores the potential uses of AI in newborn medicine, highlighting its value in the timely detection of neonatal diseases and developmental abnormalities, analysis of vital signs, medical images, and genetic information.3,4 It can also process information, ranging from medical literature pertaining to research articles, clinical trials, and/or case studies.5 However, despite all strengths and the significant potential of AI in newborn medicine, challenges such as data privacy, algorithm bias, and regulatory compliance need to be addressed to ensure the responsible and ethical deployment of AI technologies.6 By leveraging the capabilities of AI, newborn medicine stands to benefit from improved diagnostic accuracy, personalized treatment strategies, and enhanced healthcare delivery, ultimately leading to better health outcomes for neonates.7
One particular set of algorithms, the Generative Pre-trained Transformers (GPTs), has emerged as an important tool in natural language processing (NLP).8,9 These can be developed with relatively limited training data, analyze large volumes of neonatal data, streamline clinical documentation, and provide clinical decision support systems.4 By automating these tasks, GPT can reduce administrative burden, improve accuracy of documentation, and enhance the efficiency of clinical workflows while enhancing the physicians wellness and limiting the risk of physician burn-out.10,11 Indirectly, this can promote evidence-based decisions, reduce diagnostic errors, and improve patient outcomes.12 GPT-powered chatbots and virtual assistants can also engage families and provide personalized medical information, answer medical queries, and facilitate remote consultations.13 These virtual assistants can enhance patient satisfaction and empower families to make informed decisions.14 GPT can also help generate patient-friendly educational materials, such as articles and videos, promoting health literacy and adherence to treatment regimens.15 Predictive modeling techniques can enable the identification of at-risk infants and promote early intervention in conditions related to the severity-of-illness such as sepsis.16 This information can help optimize healthcare delivery with resource allocation, patient triage, and telemedicine consultations.17 This can enhance access to newborn care services, particularly in underserved regions.18
Please note GPT models can promote continuing medical education of care-providers.19 The algorithms can help identify questions relevant for clinical/translational research, identify important studies conducted in the past, and summarize complex findings for education.20 It could identify the most important clinical needs where new drugs are needed, predict drug interactions, and optimize drug design.21 To summarize, GPTs have emerged as valuable tools in medicine, in healthcare delivery, medical research, and patient engagement. By leveraging the power of NLP, GPT enhances clinical documentation, facilitates medical literature analysis, improves patient communication, and accelerates drug discovery efforts.
In this article, we have focused on the strengths and weaknesses of GPT, particularly its version 3.5. It can help determine infrastructure needs for processing and information systems, costs, biases, and the evaluation metrics.5 For application in clinical medicine, there is a need to ensure compliance with the Health Insurance Portability and Accountability Act, collaborate with healthcare providers, and ensure access and technical understanding of these software tools.22 These software models are being continuously updated to improve the human-like logical and intellectual responses to prompts.23 However, there is a need for caution as questions remain about safety and accuracy before its full-scale operationalization and its use in clinical practice and research. Here, we introduce AI and GPT for its capabilities, considerations for its implementation and operationalization in clinical practice, and the need for caution and well-designed control mechanisms.
History of Development of AI
These developments have involved a series of events in scientific research, technological advancements, and interdisciplinary collaboration over several decades:
Early Foundations
In the 1950s, the field of AI emerged with the seminal work of researchers such as Alan Turing, John McCarthy, Marvin Minsky, Allen Newel, and Herbert Simon.24
Alan M Turing, an English mathematician, computer scientist, cryptanalyst, philosopher, and theoretical biologist is considered the father of theoretical computer science and artificial intelligence.25,26 During the Second World War, he was involved with the British code-breaking center that produced ultra-intelligence. His work assisted breaking the German naval ciphers, most importantly the enigma machine.
John McCarthy coined the term “Artificial Intelligence” as early as 1955.25 He organized an 8 week summer research conference in Dartmouth, New Hampshire in 1956 where 11 mathematicians and scientists laid the foundation for AI as a new scientific field. The topics of the meeting focused on computers, NLP, neural networks, theory of computation, abstraction, and creativity. The expertise of the participants was very diverse, including economists, political scientists, cognitive psychologists, computer scientists, and electrical engineers to mention a few.
Early AI research focused on symbolic AI, which used symbolic representations and logic-based reasoning to simulate human intelligence.27
Key developments during this period included the development of expert systems, logical reasoning systems, and early forms of machine learning (ML) algorithms.28
Period of Low Progress
During the 1970s–1980s, AI research faced significant challenges and setbacks, leading to a period known as the “AI winter.”29
Financial support for AI research declined, and there was skepticism about the feasibility of achieving human-level intelligence with existing approaches.
Despite these challenges, research continued in areas such as expert systems, NLP, and robotics.24
Rise of Machine Learning (ML)
In the 1980s and 1990s, ML emerged as a dominant paradigm in AI research.30
Researchers explored various ML techniques, including neural networks, genetic algorithms, and statistical methods, to develop AI systems that could learn from data.31
Key developments during this period included the back propagation algorithm for training neural networks, the development of support vector machines (SVMs), and the rise of Bayesian methods in ML.32
Computational Advances
The 2000s witnessed significant advances in computational power, data availability, and algorithmic innovation, fueling rapid progress in AI research.33
Deep learning (DL), a subfield of ML, focused on training deep neural networks, gained prominence due to its ability to learn hierarchical representations from data.34
Breakthroughs in DL, such as the development of convolutional neural networks (CNNs) for image recognition and recurrent neural networks (RNNs) for sequence modeling, revolutionized AI applications in computer vision, NLP, and robotics.35
The IBM Watson Health was launched in 2017 after IBM-Watson defeated two human “Jeopardy” champions in 2011. The initial experience showed that AI, DL, and ML require a careful review of the quality and size of the input data, the applied algorithm, and the validity of the output results before it can be rolled out in clinical healthcare.
Interdisciplinary Collaboration and Applications (Present)
AI research today involves interdisciplinary collaboration between computer scientists, mathematicians, engineers, neuroscientists, and domain experts from various fields.36
These technologies have been applied to a wide range of domains, including healthcare, finance, transportation, agriculture, and entertainment, and has transformed industry.37
Ongoing research in AI focuses on addressing challenges such as data privacy and governance, fairness, interpretability, and ethical considerations, as well as advancing the capabilities of AI systems to achieve or surpass human-level intelligence in complex tasks.38
Overall, the development of AI has been a gradual process, driven by scientific curiosity, technological innovation, and real-world applications. It has involved contributions from researchers and practitioners across multiple disciplines. In addition, the development of AI algorithms also has significant economic implications and potential. Many for profit organizations are progressively offering commercial AI solutions.
Potential Importance of AI in Healthcare
AI has been viewed as a promising advancement in medicine. Specific medical tasks, available data, computational resources, and regulatory considerations seem to be important determinants.7,39–42 Here are some types of AI commonly used in the broader picture of healthcare and medicine:
Machine Learning (ML)
ML is focused on the development and study of statistical algorithms that can learn from data, generalize the paradigms to unseen data, and then perform tasks without explicit instructions.43 ML models can be trained on large datasets of medical images, including objective imaging-based scalars such as apparent diffusion coefficients or concentration of metabolic markers; patient records; genomic data; and other healthcare data to assist in diagnosis, treatment planning, and personalized medicine.44–50 These techniques seem to be particularly effective for tasks such as medical image analysis, disease classification, predictive modeling, and clinical decision support.51 There are three broad patterns:
Supervised learning: These algorithms are trained on labeled datasets, where input–output pairs are used to learn the mapping between input features and target labels. Supervised learning techniques are used in medical image analysis, disease classification, and predictive modeling.52
Unsupervised learning: These involve training algorithms on unlabeled data to discover hidden patterns or structures within the data. Unsupervised learning techniques are used in clustering, anomaly detection, and dimensionality reduction tasks in medicine.53
Semi-supervised learning: These combine elements of supervised and unsupervised learning by leveraging a small amount of labeled data and a larger amount of unlabeled data to improve model performance. Semi-supervised learning techniques are used when labeled data are scarce or expensive to obtain in medical applications.54
Deep Learning (DL)
DL is a subfield of ML. It has shown remarkable success in various medical imaging tasks, including radiology, pathology, and dermatology. DL models, such as CNNs and RNNs can learn hierarchical representations from raw data, enabling automatic identification of patterns from medical images and signals.55,56 DL techniques are also used in medical NLP tasks, such as clinical documentation, medical transcription, and patient data extraction from electronic health records (EHRs). There are three broad patterns:
Convolutional neural networks (CNNs): These networks are DL models designed to process structured grid-like data, such as images. CNNs are widely used in medical imaging tasks, including image classification, segmentation, object detection, and pattern recognition.57
Recurrent neural networks (RNNs): These are DL models with recurrent connections that enable them to capture sequential dependencies in data. RNNs are used in tasks involving sequential data, such as time-series analysis, NLP, and EHR analysis.58
Transformer models: These include GPT (Bidirectional Encoder Representations from Transformers) and Bidirectional Encoder Representations from Transformers (BERT), which are DL architectures designed to process data with self-attention mechanisms.59 Transformer models are used in medical NLP tasks, including language translation, text summarization, and clinical documentation. Later in this review, we have focused on GPT and its importance in medicine and its branches such as neonatology.60
Expert Systems
These are AI systems designed to mimic the decision-making processes of human experts in specific domains. Expert systems use rules-based approaches and knowledge representation techniques to reason and make decisions. Expert systems are used in medical diagnosis, treatment planning, and clinical decision support systems.61 These systems are particularly useful in domains where expert knowledge is well-defined and can be codified into rules or algorithms, such as pathology, radiology, and dermatology.
Natural Language Processing (NLP)
NLP involves techniques for processing and understanding human language. These algorithms can help in:
Processing unstructured text data in medicine, including in clinical notes, research articles, and communication between healthcare providers and patients.62
Retrieval of information from medical texts. Natural language processing can then help with sentiment analysis and to summarize documents- and answer-related questions.63
Medical literature mining to develop clinical decision support systems and patient communication platforms.64
Computer Vision
These involve techniques for analyzing and interpreting visual data, such as medical images and videos. Computer vision techniques are used in medical imaging modalities, including radiology, pathology, and dermatology, to assist in diagnosis, image interpretation, and treatment planning.65 Pattern recognition of imaging findings has been useful in neonatal and pediatric neuroradiology.66,67 In these programs, feeding new magnetic resonance imaging data evokes an output with a list of differential diagnoses, with probabilities and 95% confidence intervals for each entity; specific data on the MRI findings of few cases could be added to the database to improve the experience and accuracy of the program.
Reinforcement Learning (RL)
These modules involve training agents to interact with an environment and learn optimal actions through a trial-and-error process. Reinforcement learning techniques are used to plan optimized, personalized treatments adaptive therapies and medical robotics.68,69
These are just a few examples of the types of AI that can be used in medicine. The choice of AI techniques depends on the specific task or application, as well as factors such as available data, computational resources, and regulatory considerations. Reinforcement learning techniques are still relatively new in medicine but hold promise for addressing complex and dynamic healthcare challenges. Integrating multiple AI techniques and approaches can lead to more robust and effective solutions.
Generative Pre-trained Transformers (GPTs), An ML Algorithm Used in Medicine
GPT is one of many AI models that have been used in medicine. The preference for GPT over other AI algorithms depends on the specific tasks or applications in the medical domain.70
Possible Advantages of Using GPT and Similar Language Models in Medical Tasks
Natural language understanding (NLU): GPT excels in understanding and generating human-like text, making it valuable for tasks such as medical documentation, literature analysis, and patient communication. Its ability to process and generate text in natural language allows it to assist healthcare professionals in writing clinical notes, summarizing medical literature, and engaging with patients effectively.71 This may also positively impact the health and resilience of healthcare professionals by reducing administrative tasks. The GPT has also been found to be effective with minor differences in syntax and accents.
Flexible and versatile: GPT is a highly flexible and versatile model that can be fine-tuned for a wide range of medical tasks and applications. It can be adapted to different medical specialties, languages, and healthcare settings by training it on domain-specific data and fine-tuning its parameters for specific tasks, such as medical question answering, clinical decision support, and medical image analysis.28
Large-scale pre-training: GPT is typically pre-trained on large amounts of text data from diverse sources, enabling it to learn rich representations of language and knowledge from a broad range of domains. This large-scale pre-training helps GPT capture complex linguistic patterns, domain-specific terminology, and contextual information relevant to medical tasks, making it effective in understanding and generating medical text.12
Contextual understanding: GPT models, particularly those with large numbers of parameters like GPT-3, excel in capturing contextual information and understanding the nuances of language. This contextual understanding allows GPT to generate coherent and contextually relevant responses to queries, making it useful for tasks such as medical question answering, clinical documentation, and patient education.72
Continuous learning: GPT can be continuously updated and improved over time by fine-tuning it on new data and tasks. It can adapt to evolving medical knowledge, guidelines, and practices, ensuring that it remains up-to-date and relevant for medical applications.73
GPT offers several advantages for certain medical tasks, but it is important to recognize that no single AI model is universally “better” than others in all contexts. Different AI models, including neural networks, ML algorithms, and statistical models, have their own strengths and limitations, and the choice of model depends on factors such as the task requirements, available data, computational resources, and ethical considerations.
Types of GPT Codes Used in Medicine
In medical AI research and applications, many versions of GPT have been tested. Three versions of GPT are best-known:
GPT-1 was the first version of GPT released by OpenAI. It had 117 million parameters and was trained on a diverse range of internet text. GPT-1 demonstrated strong performance on various NLP tasks but was later surpassed by more sophisticated models integrating larger data sets and parameter.
GPT-2 was a significant advancement over GPT-1, featuring 1.5 billion parameters. It was trained on a massive dataset scraped from the internet, allowing it to generate highly coherent and contextually relevant text. GPT-2 attracted attention due to concerns about its potential misuse for generating fake news or deceptive content, leading OpenAI to initially release it in a controlled manner. However, the full version was later made publicly available.
GPT-3 is a powerful version of the model, boasting a staggering 175 billion parameters. It is the largest publicly known language model to date. GPT-3 demonstrated remarkable capabilities in generating human-like text, understanding context, and performing a wide range of NLP tasks, including translation, question answering, text completion, and more. GPT-3 has been widely used and studied by researchers and developers across various domains.74
GPT-4 is a multimodal large language model that was released in early 2023. It is a transformer-based model where pre-training uses both public data and “data licensed from third-party providers.” The model has been fine-tuned with reinforcement learning feedback from humans and AI for human alignment and policy compliance.
Each version of GPT has built upon the advancements of its predecessor, with improvements in model architecture, training data, and scale. The increasing size and complexity of these models have contributed to significant advancements in natural language understanding and generation, enabling a wide range of applications across industries.75 Below is a simplified flow diagram outlining the general process of developing a GPT model for a specific application, such as medicine:
sql
Copy code
Start
|
|– Define Task and Objectives
|
|– Collect Data
| |– Gather Raw Text Data
| |– Pre-process Data (Tokenization, Cleaning, etc.)
| |– Create Training, Validation, and Test Sets
|
|– Pre-train GPT Model
| |– Initialize Model Architecture
| |– Train on Large Text Corpus (Unsupervised Learning)
| |– Fine-tune Hyperparameters
| |– Monitor Training Progress
|
|– Fine-tune GPT Model
| |– Initialize Pre-trained Model
| |– Fine-tune on Domain-specific Data (Supervised Learning)
| |– Fine-tune Task-specific Head (if applicable)
| |– Validate and Tune Hyperparameters
| |– Evaluate Performance on Validation Set
|
|– Evaluate and Test
| |– Evaluate Model Performance on Test Set
| |– Assess Metrics (Accuracy, Precision, Recall, etc.)
| |– Iterate and Refine Model (if necessary)
|
|– Deployment
| |– Integrate Model into Application or Workflow
| |– Monitor Model Performance in Real-world Settings
| |– Provide Updates and Maintenance as Needed
|
End
This flow diagram outlines the main steps involved in developing and deploying a GPT model for a specific task, such as medical text generation or medical question answering. Each step involves a series of subtasks, such as data collection, pre-processing, model training, fine-tuning, evaluation, and deployment. Additionally, there may be iterative loops where the model is refined based on performance feedback or changes in requirements.
Customized Versions of GPT
Customized versions of GPT can be developed for specific needs with appropriate expertise in ML and NLP.76,77 Here are some general steps (Fig. 1):
Fig. 1: An algorithm to develop a problem-solving model based on machine learning/natural language processing
A clearly defined specific NLP task, such as text generation, text classification, sentiment analysis, language translation, or summarization.
Collection of data: A large dataset of good quality and diversity is desirable.
Selection of the architecture of the NLP model: A transformer-based architecture resembling GPT or other models can be used per requirements/constraints of the project.
Pre-training of the NLP model on the dataset using self-supervised/supervised learning; this involves training the model to predict the next word(s) in a sentence given the context.
Fine-tuning of the pre-trained model on specific examples or relevant tasks. These polishing steps improve its performance on the target task.
Evaluation of the performance of the model on a separate validation or test dataset to assess its effectiveness and identify areas for improvement.
Iterative refinement of the design, training, and evaluation process to improve the model. This may need change in hyperparameters or architectures, or testing with additional data.
Once the customized language model looks satisfactory, the model can be deployed in the application or integrated into the workflow to start making predictions or generating text.
This flow diagram provides a high-level overview, and the specific details and implementation may vary depending on the task, dataset, model architecture, and research goals. Additionally, newer developments and research in the field may introduce variations or improvements to the overall process. Customized versions of GPT or a similar language model can be developed for specific needs, although it might require significant expertise in ML and NLP. Here are some general steps to consider:
Designing and training a customized language model can be a complex and resource-intensive process, requiring expertise in ML, NLP, and data engineering. Additionally, there might be a need for powerful computing resources, such as graphics processing units or Tensor Processing Units to train large-scale models effectively.
Parameters Used in the Development of GPT for Medicine
This is an evolving process.78 Here are some general information about the parameters typically included in training GPT models for medical applications:
Model architecture: The GPT models are based on a transformer architecture, which consists of multiple layers of self-attention mechanisms and feedforward neural networks.79 The number of layers, hidden units, attention heads, and other architectural parameters can vary depending on the size and complexity of the model.
Pre-training data: Models pre-trained on large amounts of text data to learn language representations can be useful.78,80 The pre-training data may include various sources of biomedical literature, clinical notes, EHRs, drug labels, and other healthcare-related text. These datasets can range from millions to billions of tokens.
Fine-tuning data: After pre-training, GPT models are fine-tuned on domain-specific data to adapt them to specific tasks or applications.81,82 This fine-tuning may focus on clinical notes, medical questions/answers, EHR, and other healthcare-related documents. The size of the fine-tuning dataset can vary depending on the specific task.
Training hyperparameters: Various hyperparameters involved in training GPT models, such as learning rate, batch size, optimization algorithm, and dropout probability are typically tuned empirically to optimize model performance on the target task or dataset.83,84
Task-specific head: In some implementations, GPT models may include a task-specific head or output layer tailored to the specific medical task at hand.85 For example, in medical question answering, the output layer may consist of a softmax classifier trained to predict the most likely answer to a given medical question.86
Evaluation metrics: During training and evaluation, many metrics may help evaluate GPT models on medical tasks.87 Evaluation metrics include accuracy, precision, recall, F1 score, and perplexity, depending on the specific task and evaluation criteria.
Notably, the exact parameters used in GPT development for medicine may vary with specific implementation, datasets, task requirements, and research goals. Newer developments and research in the field may introduce variations or improvements to the training and fine-tuning processes.
GPT Models Used in Medicine
GPT models have been utilized for various applications in medicine.88,89 Although we still do not have specific GPT models for these tasks, the GPT architecture, particularly in GPT-3, has been adapted for medical use. Here are some examples:
BioGPT: A version fine-tuned specifically for biomedical applications.90 It has been trained on biomedical literature and clinical text to understand medical terminology, concepts, and contexts. BioGPT has been used for tasks such as medical text generation, clinical documentation, and medical question answering.
Clinical GPT: Clinical GPT is another variant fine-tuned on EHRs, medical literature, and healthcare-related text to assist healthcare providers in clinical documentation, patient management, and decision support.91 It aims to improve clinical workflows while ensuring compliance with healthcare regulations and standards.
GPT-M: A modified version of GPT-3 optimized for medical text generation and understanding. It has been trained on a curated dataset of medical documents, including clinical notes, research articles, and drug labels. The GPT-M can generate medical reports, summarize patient information, and answer queries accurately.
MedGPT: A variant designed specifically for answering medical questions and knowledge retrieval.92 It aims to assist healthcare professionals in accessing medical knowledge efficiently to support clinical decision-making.
To summarize, GPT models have been adapted and customized for medical applications on domain-specific data and tasks. By leveraging the power of NLP, these specialized GPT models contribute to improving healthcare delivery, clinical decision support, and medical research. Many GPT and other large language models have several applications in the field of medicine, including:93
Clinical documentation and note generation: GPT can be used to assist healthcare professionals in generating clinical notes, summaries, and reports.94 By providing relevant patient information as input, the model can generate structured, accurate documentation to improve the efficiency of medical record-keeping.
Medical literature summarization: GPT can help summarize large volumes of medical literature, including research articles, clinical trials, and case studies.95 This can assist healthcare professionals in staying up-to-date with the latest advancements in their field and making informed decisions about patient care.
Medical chatbots and virtual assistants: These can provide patients with personalized medical information, answer their questions about symptoms, treatments, medications, and even schedule appointments.96 These virtual assistants can improve patient engagement, provide continuous support all over the day, and alleviate the burden on healthcare providers.
Clinical decision-support systems: GPT can be integrated into clinical decision support systems to assist healthcare providers in making diagnostic and treatment decisions.97 By analyzing patient data, medical history, and symptoms, GPT can provide recommendations for appropriate diagnostic tests, treatment options, and medication dosages.
NLP for electronic health records (EHR): GPT can help extract information from unstructured EHR. This includes identifying key medical concepts, relevant information for research or analysis, and improving the accuracy of coding and billing.98
Patient education and health communication: GPT can generate patient-friendly educational materials, such as articles, pamphlets, and videos, to help patients better understand their medical conditions, treatment options, and preventive care measures.99 This can improve health literacy, patient engagement, and adherence to treatment plans.
Drug discovery and development: GPT can help analyze biomedical literature, clinical trial data, and molecular structures to accelerate drug discovery and development processes.100,101 By identifying drug candidates, predicting drug interactions, and optimizing drug design, GPT can contribute to advancing medical research and improving patient outcomes.100
These are just a few examples of how GPT and other large language models are being utilized in the field of medicine. With continuing progress, we anticipate to see more innovative applications that leverage NLP to improve healthcare delivery and patient outcomes.102–104
Accuracy of GPT for Medicine
The accuracy of GPT for medicine depends on several factors, including the quality of the training data, the specific medical task, and the fine-tuning process.105 While GPT and similar language models have shown impressive capabilities in natural language understanding and generation, their performance in medical applications varies depending on the complexity and domain-specific nature of the tasks:
Training data quality: The quality and quantity of the training data used to fine-tune the model are important determinants of the accuracy of GPT models.106 Training data that are representative of the medical domain and cover a wide range of medical topics, specialties, and terminology are essential for achieving high accuracy.52
Task complexity: The accuracy of GPT varies depending on the complexity of the medical task.107 GPT may perform well on relatively simple tasks, such as medical text generation or summarization, but may not be as effective in tasks that require deep domain knowledge or specialized expertise such as clinical management.108
Domain specificity: GPT’s accuracy for medicine is affected by its ability to understand and generate medical terminology, concepts, and contexts. GPT models fine-tuned on large volumes of medical text demonstrate better accuracy for medical tasks compared with models trained on generic text data.109
Evaluation metrics: GPT models developed for medicine are typically evaluated using standard metrics such as precision, recall, F1 score, and perplexity.110 However, should also be evaluated clinical relevance, interpretability, and generalizability when assessing the performance of GPT models in medical applications.
Ethical and regulatory considerations: Besides accuracy, ethical considerations, such as patient privacy, algorithm bias, and regulatory compliance, also play a crucial role in determining the suitability of GPT for medical applications.111,112
Overall, while GPT and similar language models have demonstrated promising capabilities in medicine, the accuracy is not guaranteed and varies depending on the specific task and application. It’s essential to carefully evaluate the performance of GPT models in medical settings before deploying these in clinical practice. Interdisciplinary collaboration between AI researchers, healthcare professionals, and domain experts is critical for leveraging GPT effectively in medicine while ensuring patient safety and quality of care.
Alternatives to GPT
There are several alternatives to GPT for NLP tasks, each with its own strengths and weaknesses.79 Here are some notable alternatives:
Bidirectional Encoder Representions from Transformers: BERT is another widely used transformer-based model developed by Google. Unlike GPT, which is trained in a left-to-right autoregressive manner, BERT is trained bidirectionally, allowing it to capture context from both directions. BERT is known for its effectiveness in various NLP tasks, including question answering, sentiment analysis, and named entity recognition.113
XLNet: XLNet is a transformer-based model that extends BERT’s pre-training approach by leveraging permutation-based language modeling. XLNet can achieve state-of-the-art performance on several benchmark NLP tasks by considering all possible permutations of the input sequence during training, allowing it to capture bidirectional context more effectively.
Transformer-XL: Transformer-XL is a variant of the above transformer model. It introduces a novel architecture that addresses the limitation of the fixed-length context window in traditional transformers by allowing for longer-term dependency modeling. Transformer-XL is particularly useful for tasks requiring long-range context understanding, such as language modeling and document summarization.
Robustly-optimized BERT approach (RoBERTa): Robustly-optimized BERT approach is a variant of BERT developed by Facebook AI. It improves upon BERT’s pre-training objectives and training strategies to achieve better performance on downstream NLP tasks. Robustly-optimized BERT approach adopts larger batch sizes, longer training sequences, and dynamically changing masking patterns during pre-training, leading to improved robustness and generalization.
A little BERT (ALBERT): A little BERT is a lightweight variant of BERT developed by Google Research that achieves performance that is comparable to BERT with fewer parameters. It introduces parameter-sharing techniques and factorized embedding parameterization to reduce model size and computational cost while maintaining high performance on various NLP tasks.
T5 (Text-to-text Transfer Transformer): T5 is a transformer-based model developed by Google that frames all NLP tasks as text-to-text tasks, allowing for unified training and evaluation procedures. T5 achieves state-of-the-art performance on a wide range of NLP benchmarks by leveraging large-scale pre-training and fine-tuning on task-specific data.
These alternatives to GPT offer different approaches to modeling and training transformer-based architectures for NLP tasks. Depending on the specific requirements of a task or application, researchers and practitioners may choose one of these models based on factors such as performance, efficiency, model size, and computational resources available.
GPT in Medical Records
The program can be used to assist in correcting medical notes, particularly in tasks such as proofreading, grammar correction, and ensuring adherence to medical terminology and conventions.114 Here’s how GPT can be applied in correcting medical notes:
Grammar correction: GPT can help identify grammatical errors, punctuation mistakes, and spelling errors in medical notes. It can suggest corrections and improvements to ensure that the notes adhere to proper grammar and writing conventions.115 Many of the newer GPT models also include automated summary tools of frequently seen findings, the so-called “smart phrases,” to make the reports more comprehensive with higher levels of detail.116
Language consistency: GPT can assist in maintaining consistency in language and terminology throughout medical notes. It can help ensure that medical terms, abbreviations, and acronyms are used consistently and correctly across different sections of the notes.19
Clarity and readability: GPT can help improve the clarity and readability of medical notes by suggesting revisions to modify/replace linguistically unusual phrasings, convoluted sentences, or ambiguous language.117 It can help streamline the writing style to ensure that the notes are easier to understand for other healthcare professionals.
Medical terminology: GPT can assist in accurate use of medical terminology and terminology specific to different medical specialties.19 It can help identify incorrect or outdated terminology and suggest appropriate replacements based on current medical standards.
Formatting and structure: GPT can help ensure that medical notes follow the appropriate formatting and structure guidelines.118 It can assist in organizing information, formatting headings and subheadings, and ensuring a logical, coherent structure of the notes.
Quality assurance: GPT can aid in quality assurance by flagging potential inconsistencies, errors, or omissions in medical notes.119,120 It can help reviewers identify areas that may require further clarification or revision to ensure the accuracy and completeness of the notes.
Overall, GPT can be valuable for correcting medical notes, but it should be used as a supplement to human expertise and judgment. Healthcare professionals should critically evaluate the suggestions and feedback generated by GPT and ensure that the corrected notes adhere to relevant medical guidelines, standards, and best practices.
GPT in Education and Research
Use of GPT to Query Databases Such as PubMed
GPT can potentially be used to assist in querying large databases but there are some considerations to keep in mind:121
Natural language interface (NLI): GPT could be employed as part of a NLI for querying databases.122 Users could input their queries in natural language, and GPT could help parse and interpret those queries to generate more structured search queries that are compatible with the database’s search interface.
Query expansion and refinement: GPT could assist in refining and expanding search queries based on contextual information provided by the user.123 For example, if a user provides a vague or ambiguous query, GPT could help clarify the user’s intent and suggest additional terms or concepts to include in the search.124
Summarization of results: After retrieving search results from PubMed or other databases, GPT could assist in summarizing and presenting the key findings or insights in a more digestible format.125 This could include generating concise summaries of research articles, identifying relevant studies, or highlighting important keywords or concepts.
Contextual understanding: GPT’s ability to understand context could be leveraged to improve search relevance and accuracy.126 For example, it could utilize previous search history, user preferences, or the specific domain of interest to tailor search results more effectively.
To summarize, GPT can potentially assist in various aspects of database querying, but it is important to note that the actual querying and retrieval of data from databases like PubMed typically utilizes more specialized tools and techniques. PubMed provides its own search interface and application programming interface (API) keys for accessing its database, and many researchers often use tools like Python with libraries such as Biopython or other specialized search engines.127–129 Overall, even though GPT can augment the querying process by providing a NLI and assisting in query refinement and result summarization, it would typically be used in conjunction with other tools and techniques for more efficient and effective database querying.
GPT to Develop Medical Manuscripts
The GPT can provide meaningful summaries with various types of images and appendices. However, the development of full manuscripts still needs work.130
GPT to Review Medical Manuscripts
GPT can be used to assist in reviewing manuscripts, particularly in tasks such as summarization, language correction, and providing feedback on the clarity and coherence of writing.131 Here’s how GPT can be applied in manuscript review:
Summarization: GPT can generate concise summaries of manuscripts, highlighting key findings and contributions. Reviewers can use these summaries for understanding the key points of the manuscript and provide feedback on its overall coherence and structure.132
Language correction: GPT can help identify grammatical errors, typos, and awkward phrasings in manuscripts. Reviewers can use GPT to suggest corrections and improvements to the writing style, ensuring clarity and readability.132
Feedback generation: GPT can generate feedback on various aspects of the manuscript, such as the strength of the arguments, the relevance of the literature cited, and the validity of the methodology. Reviewers can use GPT-generated feedback as a starting point for providing detailed critiques and suggestions for improvement.126
Plagiarism detection: While not a primary function of GPT, it can assist in identifying potential instances of plagiarism by comparing text passages in the manuscript with existing literature and databases. Reviewers can use GPT to flag suspicious similarities and recommend further investigation.133
Reviewer assistance: GPT can aid reviewers by providing additional context or background information for the manuscript. Reviewers can use GPT to look up relevant references, definitions, or explanations to enhance their understanding of the subject matter.126
Thus, GPT can be a valuable tool in manuscript review, but it should be viewed as a supplement to human expertise and judgment rather than a replacement. Reviewers should critically evaluate the suggestions and feedback generated by GPT based on their own expertise in the field. They should also be aware of the limitations and biases of these models.
Customized versions of GPT or similar language models can be developed for specific needs, using ML, NLP, and data engineering. Additional computing resources such as GPUs or TPUs might be needed to effectively train large-scale models.
GPT to Make Images for Medical Manuscripts
GPTs are designed primarily for NLP tasks, such as text generation and understanding. However, newer generative models with embedded DL can be tailored for image generation. An example is shown in Figure 2; there are several possible versions of a photographic image (one of the two authors).134,135
Figs 2A to H: Newer generative models with embedded deep-learning can be tailored for image generation. The images above show. (A) A photograph of one of the authors (Dr. Maheshwari) that has been modified to make (B–H) A series of simplified cartoon (“toon”) “avatars” using an AI algorithm
Generative adversarial networks (GANs): These DL models are comprised of two neural networks, a generator and a discriminator, which are trained simultaneously.136 The generator can learn to produce realistic images from random noise, while the discriminator learns to distinguish between real and fake images. Generative adversarial networks can generate high-quality images across various domains.137 Artificial intelligence has been used to generate chatbots, which can use algorithms with AI, ML, NLU, and NLP to simulate a human conversation with text messages in a chat window. These windows can produce text, images, sounds, software, and other digital media.
Variational autoencoders (VAEs) are another type of generative models that can learn and generate new data points by capturing the distribution of the input data. Variational autoencoders can be trained to encode input data into a lower-dimensional latent space and then decode it back into the original space. These can be used for tasks like image generation and reconstruction.138
Autoregressive models: GPT itself is an autoregressive model for text generation, but similar architectures can be applied to image generation tasks.8 Models such as PixelRNN and PixelCNN can generate images that show pixel-by-pixel congruity with previous versions.139,140
As shown in Figures 3 to 5, current image generators can produce interesting images but there is still room for improvement. Figure 3 shows output images of the two principal authors of this article. Panel A shows an image that resembles Dr. Maheshwari but the infant seems to show congenital anomalies. In panel B, most observers did not identify the image of the physician with Dr. Huisman. Figure 4 shows the typical clinical setting of two physicians attending to a newborn baby. However, there are multiple incorrect background details and what appear to be structural abnormalities in the baby. In Figure 5, the output function failed to illustrate the requested germinal matrix hemorrhage correctly; the figure showed a whirled red color that faintly resembled an acute hemorrhage.
Figs 3A and B: Current image generators can produce interesting images but there is still room for improvement. This figure shows output images of the two principal authors of this article, produced by ChatOn powered by the Chat GPT image generator (AIBY, Florida, USA; https://aiby.com). (A) Image created using the input function: “Create an image of Dr. Akhil Maheshwari as a young neonatologist.” The image shows features identifiable with those in his photograph. However, the infant in the image shows “malformed/bent” lower extremities; (B) Another image created using the input function: “Create an image of Dr. Thierry A.G.M. Huisman as a young pediatric neuroradiologist” shows an expert with a stethoscope around his neck. The screen shows soft organs of the chest and upper abdomen with an unusual anatomy. A surgical lamp can be seen above the imaging equipment. Many observers did not identify the expert in the image as resembling Dr. Huisman
Fig. 4: Output images of a neonatologist and radiologist produced by ChatOn powered by the Chat GPT image generator (AIBY, Florida, USA; https://aiby.com). The input function was “Create an image of a neonatologist and pediatric radiologist.” The figure correctly shows the typical clinical setting of two physicians attending to a newborn baby. However, multiple incorrect background details including a “malformed” chest imaging and a “holographic” bony upper chest and skull are seen. Looking at the baby, it almost seems that there might be a hemangioma on the dorsum of the nose. The eyeball appears red, and the ears look ‘low-set’. The face seems to show swelling below the zygomatic arch. The abdomen looks a bit distended. In addition, the upper and lower extremities of the newborn seem malformed with subcutaneous edema and syndactyly. The left foot seems to show a hemangioma or a subcutaneous hematoma. So many artifacts!
Fig. 5: ChatOn powered by Chat GPT image generator (AIBY, Florida, USA, https://aiby.com) output images of a germinal matrix hemorrhage. We used an input function “Create an image of a germinal matrix hemorrhage in a newborn.” The output function failed to illustrate a germinal matrix hemorrhage correctly; the figure shows a whirled red color faintly suggesting an acute hemorrhage
GPT to Tabulate Data for Medical Manuscripts
GPT tools are not specifically designed to create visualizations like bar diagrams or pie charts as these typically require manipulation of numerical data. These graphics are easier prepared using specialized software tools such as:
Python-based libraries: Python is a general-purpose programming language; the design philosophy emphasizes code readability. It is a dynamically typed, garbage-collected program that can support multiple programming paradigms, including structured, object-oriented and functional programming.141–143
–Matplotlib is a widely used library that can be used for creating static, interactive, and animated visualizations. It can generate bar plots, pie charts, line plots, and scatter plots.144
–Seaborn is an advancement over Matplotlib; it provides a high-level interface for creating attractive and informative statistical graphics. These advanced visualizations can support options for further customization.144
–Plotly provides interactive plots and dashboards. It supports many chart types, including bar charts, pie charts, line charts, and scatter plots. Plotly can also be used in conjunction with Dash for building interactive web applications.145,146
ggplot2 is a plotting system for the R programming language inspired by the Grammar of Graphics.147 It is a powerful framework for creating statistical graphics such as bar plots, pie charts, and histograms.148
Business intelligence (BI) tools like Tableau and Power BI offer interfaces for creating visualizations such as bar charts, pie charts, heatmaps, and dashboards. These tools can be customized and help connecting to data sources directly.149
GPT can be useful in creating visualizations such as bar diagrams or pie charts, and generating textual descriptions and/or explanations of data visualizations.
GPT for Language Translation
GPT-based models can be used for text generation and summarization, answering of questions, and language translation.150 Some transformer-based models have been specifically tailored for language translation. The GPT transformer models have been used in Google’s transformer-based translation models.151
Models like BERT and its variants can be fine-tuned for translation tasks. Models like T5 can be trained in a “translation mode” to perform translation tasks. Generally, GPT can generate text and understand context well, for translation tasks; specialized models can be even more effective.
GPT to Tabulate Data for Medical Manuscripts
GPT models can be tailored to assist in certain aspects related to tabulating data, even though these are not specifically designed for tabulating data or manipulating structured data formats like tables.152 These can be useful for:
Data summarization: These can summarize textual data such as in descriptions or explanations of tabulated data to present key points or trends.
Data interpretation: These models can help interpret textual descriptions of data as in identifying important features or insights and conveying those in natural language.
Natural language interface: The GPT models can be used in NLI for interacting with tabular data. Users might be able to ask questions in natural language about the data, and then GPT could help parse and interpret those queries to retrieve relevant information from the dataset.
However, if the primary goal is to create or manipulate tables from raw data, other tools and techniques might be more suitable:
Spreadsheet software: Tools like Microsoft Excel, Google Sheets, or specialized data analysis software provide powerful features for tabulating and manipulating structured data.
Data processing libraries: Programming languages like Python offer libraries such as Pandas153 which can help generate functions for creating, modifying, and summarizing tabular data.
Business intelligence software like Tableau, Power BI, or Looker are designed specifically for data visualization and analysis, using interactive dashboards and tabular data sources.154
Even though GPT might not be the most suitable tool for direct tabulation or statistical analysis of data to determine whether the observed differences or relationships are likely to be real or simply due to random chance.155 The actual statistical analysis would need statistical software or programming languages such as R, Python with libraries like NumPy, SciPy, or statsmodels, or dedicated statistical software packages like SPSS or SAS.156,157 Then, GPT applications can be used to query data, where an NLI could be used to parse the retrieved information, summarize/interpret the results of data analysis, identify patterns, and convey those in natural language.
CONCLUSIONS
Artificial intelligence, particularly GPT, is emerging as an important tool for developing the infrastructure for processing needs and information systems; operating costs; biases in models; and evaluation metrics. Further work is needed for the development of operational factors that drive the adoption of AI in the US healthcare system, such as ensuring compliance with the health insurance portability and accountability act, team-building/collaboration with healthcare providers, and ensuring continued development and training for the use of AI tools so that correct questions can be asked. We will need teams that include healthcare practitioners, AI developers, clinicians, and decision makers, which can develop a deep understanding of the use of the powerful AI tools integrated into hospital systems and healthcare. Recent developments in analysis of electrocardiograms, electroencephalograms, data from genetics, and prediction of chronic behavioral and other conditions holds promise for newborn medicine.
We all know that there will be challenges in protection of privacy in handling data, bias in algorithms, and in regulatory compliance. A well-designed and critical expert review board will be important for preventing AI-generated systematic errors. Continued efforts will be needed to understand and streamline AI. However, if the medical community hesitates today in overseeing this juggernaut, the inclusion (or not) of AI in medicine might not stop—it might just gradually get extrapolated into patient care from other organizations/industry for cost reasons, not because of justification based on actual clinical data. If we do not get involved in this process to oversee the development/incorporation of AI in newborn medicine, the questions in making decisions will just change from who, to which, when, and how. Maybe this will not be the most appropriate scenario. Hence, we should embrace and participate in the development and regulation of AI. These are valuable tools that can be developed for accurate analysis of large and complex datasets, likely in a more accurate fashion than human observers alone.
ORCID
Akhil Maheshwari https://orcid.org/0000-0003-3613-4054
REFERENCES
1. Davenport T, Kalakota R. The potential for artificial intelligence in healthcare. Future Healthc J 2019;6(2):94–98. DOI: 10.7861/futurehosp.6-2-94.
2. Sullivan BA, Beam K, Vesoulis ZA, et al. Transforming neonatal care with artificial intelligence: Challenges, ethical consideration, and opportunities. J Perinatol 2024;44(1):1–11. DOI: 10.1038/s41372-023-01848-5.
3. Khan M, Khurshid M, Vatsa M, et al. On AI approaches for promoting maternal and neonatal health in low resource settings: A review. Front Public Health 2022;10:880034. DOI: 10.3389/fpubh.2022.880034.
4. Beam K, Sharma P, Levy P, et al. Artificial intelligence in the neonatal intensive care unit: The time is now. J Perinatol 2024;44(1):131–135. DOI: 10.1038/s41372-023-01719-z.
5. Al Kuwaiti A, Nazer K, Al-Reedy A, et al. A review of the role of artificial intelligence in healthcare. J Pers Med 2023;13(6):951. DOI: 10.3390/jpm13060951.
6. Naik N, Hameed BMZ, Shetty DK, et al. Legal and ethical consideration in artificial intelligence in healthcare: Who takes responsibility? Front Surg 2022;9:862322. DOI: 10.3389/fsurg.2022.862322.
7. Johnson KB, Wei WQ, Weeraratne D, et al. Precision medicine, AI, and the future of personalized health care. Clin Transl Sci 2021;14(1):86–93. DOI: 10.1111/cts.12884.
8. Korngiebel DM, Mooney SD. Considering the possibilities and pitfalls of Generative Pre-trained Transformer 3 (GPT-3) in healthcare delivery. NPJ Digit Med 2021;4(1):93. DOI: 10.1038/s41746-021-00464-x.
9. Brown TB, Mann B, Ryder N, et al. Language models are few-shot learners. 2020. Available from: https://arxiv.org/abs/2005.14165.
10. Holderried F, Stegemann-Philipps C, Herschbach L, et al. A generative pretrained transformer (gpt)-powered chatbot as a simulated patient to practice history taking: Prospective, mixed methods study. JMIR Med Educ 2024;10:e53961. DOI: 10.2196/53961.
11. Sirrianni J, Sezgin E, Claman D, et al. Medical text prediction and suggestion using generative pretrained transformer models with dental medical notes. Methods Inf Med 2022;61(5–06):195–200. DOI: 10.1055/a-1900-7351.
12. Sezgin E, Sirrianni J, Linwood SL. Operationalizing and implementing pretrained, large artificial intelligence linguistic models in the US Health Care System: Outlook of generative pretrained transformer 3 (GPT-3) as a service model. JMIR Med Inform 2022;10(2):e32875. DOI: 10.2196/32875.
13. Amisha, Malik P, Pathania M, et al. Overview of artificial intelligence in medicine. J Family Med Prim Care 2019;8(7):2328–2331. DOI: 10.4103/jfmpc.jfmpc_440_19.
14. Jadczyk T, Wojakowski W, Tendera M, et al. Artificial intelligence can improve patient management at the time of a pandemic: The role of voice technology. J Med Internet Res 2021;23(5):e22959. DOI: 10.2196/22959.
15. Islam MR, Urmi TJ, Mosharrafa RA, et al. Role of ChatGPT in health science and research: A correspondence addressing potential application. Health Sci Rep 2023;6(10):e1625. DOI: 10.1002/hsr2.1625.
16. Ramgopal S, Sanchez-Pinto LN, Horvat CM, et al. Artificial intelligence-based clinical decision support in pediatrics. Pediatr Res 2023;93(2):334–341. DOI: 10.1038/s41390-022-02226-1.
17. Bekbolatova M, Mayer J, Ong CW, et al. Transformative potential of AI in healthcare: Definitions, applications, and navigating the ethical landscape and public perspectives. Healthcare (Basel) 2024;12(2):25. DOI: 10.3390/healthcare12020125.
18. Ganju A, Satyan S, Tanna V, et al. AI for improving children’s health: A community case study. Front Artif Intell 2020;3:544972. DOI: 10.3389/frai.2020.544972.
19. Abd-Alrazaq A, AlSaad R, Alhuwail D, et al. Large language models in medical education: Opportunities, challenges, and future directions. JMIR Med Educ 2023;9:e48291. DOI: 10.2196/48291.
20. Bernstam EV, Shireman PK, Meric-Bernstam F, et al. Artificial intelligence in clinical and translational science: Successes, challenges and opportunities. Clin Transl Sci 2022;15(2):309–321. DOI: 10.1111/cts.13175.
21. Paul D, Sanap G, Shenoy S, et al. Artificial intelligence in drug discovery and development. Drug Discov Today 2021;26(1):80–93. DOI: 10.1016/j.drudis.2020.10.010.
22. Rezaeikhonakdar D. AI chatbots and challenges of HIPAA compliance for AI developers and vendors. J Law Med Ethics 2023;51(4):988–995. DOI: 10.1017/jme.2024.15.
23. Dillion D, Tandon N, Gu Y, et al. Can AI language models replace human participants? Trends Cogn Sci 2023;27(7):597–600. DOI: 10.1016/j.tics.2023.04.008.
24. Kaul V, Enslin S, Gross SA. History of artificial intelligence in medicine. Gastrointest Endosc 2020;92(4):807–812. DOI: 10.1016/j.gie.2020.06.040.
25. Mintz Y, Brodie R. Introduction to artificial intelligence in medicine. Minim Invasive Ther Allied Technol 2019;28(2):73–81. DOI: 10.1080/13645706.2019.1575882.
26. Green JBA. Computational biology: Turing’s lessons in simplicity. Biophys J 2021;120(19):4139–4141. DOI: 10.1016/j.bpj.2021.08.041.
27. Sarker IH. AI-based modeling: Techniques, applications and research issues towards automation, intelligent and smart systems. SN Comput Sci 2022;3(2):158. DOI: 10.1007/s42979-022-01043-x.
28. Xu Y, Liu X, Cao X, et al. Artificial intelligence: A powerful paradigm for scientific research. Innovation (Camb) 2021;2(4):100179. DOI: 10.1016/j.xinn.2021.100179.
29. London AJ. Groundhog day for medical artificial intelligence. Hastings Cent Rep 2018;48(3): inside back cover. DOI: 10.1002/hast.842.
30. Ahmed Z, Mohamed K, Zeeshan S, et al. Artificial intelligence with multi-functional machine learning platform development for better healthcare and precision medicine. Database (Oxford) 2020;2020:baaa010. DOI: 10.1093/database/baaa010.
31. Swic A, Wolos D, Gola A, et al. The use of neural networks and genetic algorithms to control low rigidity shafts machining. Sensors (Basel) 2020;20(17):4683. DOI: 10.3390/s20174683.
32. Ben-Hur A, Weston J. A user’s guide to support vector machines. Methods Mol Biol 2010;609:223–239. DOI: 10.1007/978-1-60327-241-4_13.
33. Ahuja AS. The impact of artificial intelligence in medicine on the future role of the physician. PeerJ 2019;7:e7702. DOI: 10.7717/peerj.7702.
34. Kufel J, Bargiel-Laczek K, Kocot S, et al. What is machine learning, artificial neural networks and deep learning? – Examples of Practical Applications in Medicine. Diagnostics (Basel) 2023;13(15):2582. DOI: 10.3390/diagnostics13152582.
35. Vaz JM, Balaji S. Convolutional neural networks (CNNs): Concepts and applications in pharmacogenomics. Mol Divers 2021;25(3):1569–1584. DOI: 10.1007/s11030-021-10225-3.
36. Krause-Juttler G, Weitz J, Bork U. Interdisciplinary collaborations in digital health research: Mixed methods case study. JMIR Hum Factors 2022;9(2):e36579. DOI: 10.2196/36579.
37. Ding H, Tian J, Yu W, et al. The application of artificial intelligence and big data in the food industry. Foods 2023;12(24):4511. DOI: 10.3390/foods12244511.
38. Ueda D, Kakinuma T, Fujita S, et al. Fairness of artificial intelligence in healthcare: Review and recommendations. Jpn J Radiol 2024;42(1):3–15. DOI: 10.1007/s11604-023-01474-3.
39. Khan B, Fatima H, Qureshi A, et al. Drawbacks of artificial intelligence and their potential solutions in the healthcare sector. Biomed Mater Devices 2023;1–8. DOI: 10.1007/s44174-023-00063-2.
40. Alowais SA, Alghamdi SS, Alsuhebany N, et al. Revolutionizing healthcare: The role of artificial intelligence in clinical practice. BMC Med Educ 2023;23(1):689. DOI: 10.1186/s12909-023-04698-z.
41. Sahu M, Gupta R, Ambasta RK, et al. Artificial intelligence and machine learning in precision medicine: A paradigm shift in big data analysis. Prog Mol Biol Transl Sci 2022;190(1):57–100. DOI: 10.1016/bs.pmbts.2022.03.002.
42. Dias R, Torkamani A. Artificial intelligence in clinical and genomic diagnostics. Genome Med 2019;11(1):70. DOI: 10.1186/s13073-019-0689-8.
43. Rowe M. An introduction to machine learning for clinicians. Acad Med 2019;94(10):1433–1436. DOI: 10.1097/ACM.0000000000002792.
44. Rana M, Bhushan M. Machine learning and deep learning approach for medical image analysis: Diagnosis to detection. Multimed Tools Appl 2022;1–39. DOI: 10.1007/s11042-022-14305-w.
45. Ren T, Li H, Modest MF, et al. Machine learning applied to the retrieval of three-dimensional scalar fields of laminar flames from hyperspectral measurements. J Quant Spectrosc Radiat Transf 2022;279:108047. DOI: 10.1016/j.jqsrt.2021.108047.
46. Gaudiano C, Mottola M, Bianchi L, et al. An apparent diffusion coefficient-based machine learning model can improve prostate cancer detection in the grey area of the prostate imaging reporting and data system category 3: A single-centre experience. Cancers (Basel) 2023;15(13):3438. DOI: 10.3390/cancers15133438.
47. Martinelli DD. Machine learning for metabolomics research in drug discovery. Intell-Based Med 2023;8:100101. DOI: 10.1016/j.ibmed.2023.100101.
48. Nair M, Lundgren LE, Soliman A, et al. Machine learning model for readmission prediction of patients with heart failure based on electronic health records: Protocol for a quasi-experimental study for impact assessment. JMIR Res Protoc 2024;13:e52744. DOI: 10.2196/52744.
49. Lourenco VM, Ogutu JO, Rodrigues RAP, et al. Genomic prediction using machine learning: A comparison of the performance of regularized regression, ensemble, instance-based and deep learning methods on synthetic and empirical data. BMC Genom 2024;25(1):152. DOI: 10.1186/s12864-023-09933-x.
50. Peng J, Jury EC, Donnes P, et al. Machine learning techniques for personalised medicine approaches in immune-mediated chronic inflammatory diseases: Applications and challenges. Front Pharmacol 2021;12:720694. DOI: 10.3389/fphar.2021.720694.
51. Javaid M, Haleem A, Singh RP, et al. Significance of machine learning in healthcare: Features, pillars and applications. Int J Intell Netw 2022;3:58–73. DOI: 10.1016/j.ijin.2022.05.002.
52. Stafie CS, Sufaru IG, Ghiciuc CM, et al. Exploring the intersection of artificial intelligence and clinical healthcare: A multidisciplinary review. Diagnostics (Basel) 2023;13(12):1995. DOI: 10.3390/diagnostics13121995.
53. Eckhardt CM, Madjarova SJ, Williams RJ, et al. Unsupervised machine learning methods and emerging applications in healthcare. Knee Surg Sports Traumatol Arthrosc 2023;31(2):376–381. DOI: 10.1007/s00167-022-07233-7.
54. Jiao R, Zhang Y, Ding L, et al. Learning with limited annotations: A survey on deep semi-supervised learning for medical image segmentation. Comput Biol Med 2024;169:107840. DOI: 10.1016/j.compbiomed.2023.107840.
55. Sarker IH. Deep learning: A comprehensive overview on techniques, taxonomy, applications and research directions. SN Comput Sci 2021;2(6):420. DOI: 10.1007/s42979-021-00815-1.
56. Chang W-J, Chen L-B, Hsu C-H, et al. A deep learning-based intelligent medicine recognition system for chronic patients. IEEE Access 2019;7:44441–44458. DOI: 10.1109/ACCESS.2019.2908843.
57. Iglesias LL, Bellon PS, Del Barrio AP, et al. A primer on deep learning and convolutional neural networks for clinicians. Insights Imaging 2021;12(1):117. DOI: 10.1186/s13244-021-01052-z.
58. Das S, Tariq A, Santos T, et al. Recurrent neural networks (RNNs): Architectures, training tricks, and introduction to influential research. In: Colliot O, editor. Machine Learning for Brain Disorders. New York, NY: Humana; 2023; pp. 117–138. DOI: 10.1007/978-1-0716-3195-9_4.
59. Sibilano E, Buongiorno D, Lassi M, et al. Understanding the role of self-attention in a transformer model for the discrimination of SCD from MCI using resting-state EEG. IEEE J Biomed Health Inform 2024;18. DOI: 10.1109/JBHI.2024.3390606.
60. Denecke K, May R, Rivera-Romero O. Transformer models in healthcare: A survey and thematic analysis of potentials, shortcomings and risks. J Med Syst 2024;48(1):23. DOI: 10.1007/s10916-024-02043-5.
61. Holman JG, Cookson MJ. Expert systems for medical applications. J Med Eng Technol 1987;11(4):151–519. DOI: 10.3109/03091908709008986.
62. Nadkarni PM, Ohno-Machado L, Chapman WW. Natural language processing: An introduction. J Am Med Inform Assoc 2011;18(5):544–551. DOI: 10.1136/amiajnl-2011-000464.
63. Petersen CL, Halter R, Kotz D, et al. Using natural language processing and sentiment analysis to augment traditional user-centered design: Development and usability study. JMIR Mhealth Uhealth 2020;8(8):e16862. DOI: 10.2196/16862.
64. van de Burgt BWM, Wasylewicz ATM, Dullemond B, et al. Combining text mining with clinical decision support in clinical practice: A scoping review. J Am Med Inform Assoc 2023;30(3):588–603. DOI: 10.1093/jamia/ocac240.
65. Pun T, Gerig G, Ratib O. Image analysis and computer vision in medicine. Comput Med Imaging Graph 1994;18(2):85–96. DOI: 10.1016/0895-6111(94)90017-5.
66. Poretti A, Valk J, Huisman TA. A neuroimaging pattern-recognition approach in pediatric intracranial calcifications. Dev Med Child Neurol 2013;55(1):7–8. DOI: 10.1111/j.1469-8749.2012.04439.x.
67. Bosemani T, Anghelescu C, Boltshauser E, et al. Subthalamic nucleus involvement in children: A neuroimaging pattern-recognition approach. Eur J Paediatr Neurol 2014;18(3):249–256. DOI: 10.1016/j.ejpn.2013.09.010.
68. Botvinick M, Wang JX, Dabney W, et al. Deep reinforcement learning and its neuroscientific implications. Neuron 2020;107(4):603–616. DOI: 10.1016/j.neuron.2020.06.014.
69. Liu S, See KC, Ngiam KY, et al. Reinforcement learning for clinical decision support in critical care: Comprehensive review. J Med Internet Res 2020;22(7):e18477. DOI: 10.2196/18477.
70. Groza T, Caufield H, Gration D, et al. An evaluation of GPT models for phenotype concept recognition. BMC Med Inform Decis Mak 2024;24(1):30. DOI: 10.1186/s12911-024-02439-w.
71. Alqahtani T, Badreldin HA, Alrashed M, et al. The emergent role of artificial intelligence, natural learning processing, and large language models in higher education and research. Res Social Adm Pharm 2023;19(8):1236–1242. DOI: 10.1016/j.sapharm.2023.05.016.
72. Uppalapati VK, Nag DS. A comparative analysis of AI models in complex medical decision-making scenarios: Evaluating ChatGPT, Claude AI, Bard, and Perplexity. Cureus 2024;16(1):e52485. DOI: 10.7759/cureus.52485.
73. Feng J, Phillips RV, Malenica I, et al. Clinical artificial intelligence quality improvement: Towards continual monitoring and updating of AI algorithms in healthcare. NPJ Digit Med 2022;5(1):66. DOI: 10.1038/s41746-022-00611-y.
74. Haim GB, Braun A, Eden H, et al. AI in the ED: Assessing the efficacy of GPT models vs. physicians in medical score calculation. Am J Emerg Med 2024;79:161–166. DOI: 10.1016/j.ajem.2024.02.016.
75. Thirunavukarasu AJ, Ting DSJ, Elangovan K, et al. Large language models in medicine. Nat Med 2023;29(8):1930–1940. DOI: 10.1038/s41591-023-02448-8.
76. Sathe TS, Roshal J, Naaseh A, et al. How I GPT it: Development of custom artificial intelligence (AI) chatbots for surgical education. J Surg Educ 2024;81(6):772–775. DOI: 10.1016/j.jsurg.2024.03.004.
77. Masters K, Benjamin J, Agrawal A, et al. Twelve tips on creating and using custom GPTs to enhance health professions education. Med Teach 2024;1–5. DOI: 10.1080/0142159X.2024.2305365.
78. Clusmann J, Kolbinger FR, Muti HS, et al. The future landscape of large language models in medicine. Commun Med (Lond) 2023;3(1):141. DOI: 10.1038/s43856-023-00370-1.
79. Zhang H, Shafiq MO. Survey of transformers and towards ensemble learning using transformers for natural language processing. J Big Data 2024;11(1):25. DOI: 10.1186/s40537-023-00842-0.
80. Xie H, Ioannidis VN, Yang C, et al. Graph-aware language model pre-training on a large graph corpus can help multiple graph applications. KDD 2023;2023:5270–5281. DOI: 10.1145/3580305.3599833.
81. Xie Z, Evangelopoulos X, Omar OH, et al. Fine-tuning GPT-3 for machine learning electronic and functional properties of organic molecules. Chem Sci 2024;15(2):500–510. DOI: 10.1039/d3sc04610a.
82. Tinn R, Cheng H, Gu Y, et al. Fine-tuning large neural language models for biomedical natural language processing. Patterns (N Y) 2023;4(4):100729. DOI: 10.1016/j.patter.2023.100729.
83. Nematzadeh S, Kiani F, Torkamanian-Afshar M, et al. Tuning hyperparameters of machine learning algorithms and deep neural networks using metaheuristics: A bioinformatics study on biomedical and biological cases. Comput Biol Chem 2022;97:107619. DOI: 10.1016/j.compbiolchem.2021.107619.
84. Mansouri-Benssassi E, Rogers S, Reel S, et al. Disclosure control of machine learning models from trusted research environments (TRE): New challenges and opportunities. Heliyon 2023;9(4):e15143. DOI: 10.1016/j.heliyon.2023.e15143.
85. Guo LL, Steinberg E, Fleming SL, et al. EHR foundation models improve robustness in the presence of temporal distribution shift. Sci Rep 2023;13(1):3767. DOI: 10.1038/s41598-023-30820-8.
86. Mesfin WM, Cho S, Lee J, et al. Deep-learning-based segmentation of fresh or young concrete sections from images of construction sites. Materials (Basel) 2021;14(21):6311. DOI: 10.3390/ma14216311.
87. Abbasian M, Khatibi E, Azimi I, et al. Foundation metrics for evaluating effectiveness of healthcare conversations powered by generative AI. NPJ Digit Med 2024;7(1):82. DOI: 10.1038/s41746-024-01074-z.
88. Lecler A, Duron L, Soyer P. Revolutionizing radiology with GPT-based models: Current applications, future possibilities and limitations of ChatGPT. Diagn Interv Imaging 2023;104(6):269–274. DOI: 10.1016/j.diii.2023.02.003.
89. Hou W, Ji Z. GeneTuring tests GPT models in genomics. bioRxiv 2023. DOI: 10.1101/2023.03.11.532238.
90. Luo R, Sun L, Xia Y, et al. BioGPT: Generative pre-trained transformer for biomedical text generation and mining. Brief Bioinform 2022;23(8):1–7. DOI: 10.1093/bib/bbac409.
91. Wang G, Yang G, Du Z, et al. ClinicalGPT: Large language models finetuned with diverse medical data and comprehensive evaluation. arXiv 2023:2306.09968. DOI: 10.48550/arXiv.2306.09968.
92. Kraljevic Z, Shek A, Bean D, et al. MedGPT: Medical concept prediction from clinical narratives. arXiv 2021;2107.03134. DOI: 10.48550/arXiv.2107.03134.
93. Wojcik S, Rulkiewicz A, Pruszczyk P, et al. Beyond ChatGPT: What does GPT-4 add to healthcare? The dawn of a new era. Cardiol J 2023;30(6):1018–1025. DOI: 10.5603/cj.97515.
94. Nguyen J, Pepping CA. The application of ChatGPT in healthcare progress notes: A commentary from a clinical and research perspective. Clin Transl Med 2023;13(7):e1324. DOI: 10.1002/ctm2.1324.
95. Qazvinian V, Radev DR, Mohammad SM, et al. Generating extractive summaries of scientific paradigms. JAIR 2013;46(1):165–201. DOI: 10.1613/jair.3732.
96. Sun G, Zhou YH. AI in healthcare: Navigating opportunities and challenges in digital communication. Front Digit Health 2023;5:1291132. DOI: 10.3389/fdgth.2023.1291132.
97. Liu S, Wright AP, Patterson BL, et al. Assessing the value of ChatGPT for clinical decision support optimization. medRxiv 2023;2023.02.21.23286254. DOI: 10.1101/2023.02.21.23286254.
98. Hossain E, Rana R, Higgins N, et al. Natural language processing in electronic health records in relation to healthcare decision-making: A systematic review. Comput Biol Med 2023;155:106649. DOI: 10.1016/j.compbiomed.2023.106649.
99. Montazeri M, Galavi Z, Ahmadian L. What are the applications of ChatGPT in healthcare: Gain or loss? Health Sci Rep 2024;7(2):e1878. DOI: 10.1002/hsr2.1878.
100. Liu Z, Roberts RA, Lal-Nag M, et al. AI-based language models powering drug discovery and development. Drug Discov Today 2021;26(11):2593–2607. DOI: 10.1016/j.drudis.2021.06.009.
101. Wang R, Feng H, Wei GW. ChatGPT in drug discovery: A case study on anticocaine addiction drug development with Chatbots. J Chem Inf Model 2023;63(22):7189–7209. DOI: 10.1021/acs.jcim.3c01429.
102. Ambrosy AP, Parikh RV, Sung SH, et al. A natural language processing-based approach for identifying hospitalizations for worsening heart failure within an integrated health care delivery system. JAMA Netw Open 2021;4(11):e2135152. DOI: 10.1001/jamanetworkopen.2021.35152.
103. Yelne S, Chaudhary M, Dod K, et al. Harnessing the power of AI: A comprehensive review of its impact and challenges in nursing science and healthcare. Cureus 2023;15(11):e49252. DOI: 10.7759/cureus.49252.
104. Tamang S, Humbert-Droz M, Gianfrancesco M, et al. Practical considerations for developing clinical natural language processing systems for population health management and measurement. JMIR Med Inform 2023;11:e37805. DOI: 10.2196/37805.
105. Johnson D, Goodman R, Patrinely J, et al. Assessing the accuracy and reliability of AI-generated medical responses: An evaluation of the Chat-GPT Model. Res Sq 2023;rs.3.rs-2566942. DOI: 10.21203/rs.3.rs-2566942/v1.
106. Levine DM, Tuwani R, Kompa B, et al. The diagnostic and triage accuracy of the GPT-3 artificial intelligence model. medRxiv 2023. DOI: 10.1101/2023.01.30.23285067.
107. Abbas A, Rehman MS, Rehman SS. Comparing the performance of popular large language models on the national board of medical examiners sample questions. Cureus 2024;16(3):e55991. DOI: 10.7759/cureus.55991.
108. Chen Q, Sun H, Liu H, et al. An extensive benchmark study on biomedical text generation and mining with ChatGPT. Bioinformatics 2023;39(9):btad557. DOI: 10.1093/bioinformatics/btad557.
109. Younis HA, Eisa TAE, Nasser M, et al. A systematic review and meta-analysis of artificial intelligence tools in medicine and healthcare: Applications, considerations, limitations, motivation and challenges. Diagnostics (Basel) 2024;14(1):109. DOI: 10.3390/diagnostics14010109.
110. Hicks SA, Strumke I, Thambawita V, et al. On evaluation metrics for medical applications of artificial intelligence. Sci Rep 2022;12(1):5979. DOI: 10.1038/s41598-022-09954-8.
111. Wang C, Liu S, Yang H, et al. Ethical considerations of using ChatGPT in health care. J Med Internet Res 2023;25(2023):e48009. DOI: 10.2196/48009.
112. Arshad HB, Butt SA, Khan SU, et al. ChatGPT and artificial intelligence in hospital level research: Potential, precautions, and prospects. Methodist Debakey Cardiovasc J 2023;19(5):77–84. DOI: 10.14797/mdcvj.1290.
113. Li F, Jin Y, Liu W, et al. Fine-tuning bidirectional encoder representations from transformers (BERT)-based models on large-scale electronic health record notes: An empirical study. JMIR Med Inform 2019;7(3):e14830. DOI: 10.2196/14830.
114. Shea YF, Lee CMY, Ip WCT, et al. Use of GPT-4 to analyze medical records of patients with extensive investigations and delayed diagnosis. JAMA Netw Open 2023;6(8):e2325000. DOI: 10.1001/jamanetworkopen.2023.25000.
115. Jhajj KS, Jindal P, Kaur K. Use of artificial intelligence tools for research by medical students: A narrative review. Cureus 2024;16(3):e55367. DOI: 10.7759/cureus.55367.
116. Brown W 3rd, Balyan R, Karter AJ, et al. Challenges and solutions to employing natural language processing and machine learning to measure patients’ health literacy and physician writing complexity: The ECLIPPSE study. J Biomed Inform 2021;113:103658. DOI: 10.1016/j.jbi.2020.103658.
117. Sallam M. ChatGPT utility in healthcare education, research, and practice: Systematic review on the promising perspectives and valid concerns. Healthcare (Basel) 2023;11(6):887. DOI: 10.3390/healthcare11060887.
118. Liu J, Wang C, Liu S. Utility of ChatGPT in clinical practice. J Med Internet Res 2023;25:e48568. DOI: 10.2196/48568.
119. Lewis JJ, Rosen CL, Grossestreuer AV, et al. Diagnostic error, quality assurance, and medical malpractice/risk management education in emergency medicine residency training programs. Diagnosis (Berl) 2019;6(2):173–178. DOI: 10.1515/dx-2018-0079.
120. Pais C, Liu J, Voigt R, et al. Large language models for preventing medication direction errors in online pharmacies. Nat Med 2024. DOI: 10.1038/s41591-024-02933-8.
121. Cinquin O. ChIP-GPT: A managed large language model for robust data extraction from biomedical database records. Brief Bioinform 2024;25(2):bbad535. DOI: 10.1093/bib/bbad535.
122. Yuan C, Ryan PB, Ta C, et al. Criteria2Query: A natural language interface to clinical databases for cohort definition. J Am Med Inform Assoc 2019;26(4):294–305. DOI: 10.1093/jamia/ocy178.
123. Rivas AR, Iglesias EL, Borrajo L. Study of query expansion techniques and their application in the biomedical information retrieval. Scientific World Journal 2014;2014:132158. DOI: 10.1155/2014/132158.
124. Walker HL, Ghani S, Kuemmerli C, et al. Reliability of medical information provided by ChatGPT: Assessment against clinical guidelines and patient information quality instrument. J Med Internet Res 2023;25:e47479. DOI: 10.2196/47479.
125. Jin Q, Leaman R, Lu Z. PubMed and beyond: Biomedical literature search in the age of artificial intelligence. EBioMedicine 2024;100:104988. DOI: 10.1016/j.ebiom.2024.104988.
126. Ray PP. ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations and future scope. Internet of Things and Cyber-Physical Systems 2023;3:121–154. DOI: 10.1016/j.iotcps.2023.04.003.
127. NLM-Support-Center. How do I obtain an API Key through an NCBI account? Bethesda, MD: National Library of Medicine; 2024. Available from: https://support.nlm.nih.gov/knowledgebase/article/KA-05317/en-us.
128. biopython.org/. Biopython version 1.83: Python Tools for Computational Molecular Biology 2024. Available from: https://biopython.org/.
129. Sayers E. A General introduction to the E-utilities Bethesda (MD): National Center for Biotechnology Information (US) 2022. Available from: https://www.ncbi.nlm.nih.gov/books/NBK25497/.
130. Anderson N, Belavy DL, Perle SM, et al. AI did not write this manuscript, or did it? Can we trick the AI text detector into generated texts? The potential future of ChatGPT and AI in Sports & Exercise Medicine manuscript generation. BMJ Open Sport Exerc Med 2023;9(1):e001568. DOI: 10.1136/bmjsem-2023-001568.
131. Salvagno M, Taccone FS, Gerli AG. Can artificial intelligence help for scientific writing? Crit Care 2023;27(1):75. DOI: 10.1186/s13054-023-04380-2.
132. Pirciu T. Proofreading and editing your writing using ChatGPT: A course on using AI to write — LESSON 5 2023. Available from: https://teodorapirciu.medium.com/chatgpt-teaches-writing-proofreading-and-editing-your-writing-using-chatgpt-271b14e0d38f.
133. Gao CA, Howard FM, Markov NS, et al. Comparing scientific abstracts generated by ChatGPT to real abstracts with detectors and blinded human reviewers. NPJ Digit Med 2023;6(1):75. DOI: 10.1038/s41746-023-00819-6.
134. Zhang X, Zhong Y, Jin C, et al. Medical image Generative Pre-Trained Transformer (MI-GPT): Future direction for precision medicine. Eur J Nucl Med Mol Imaging 2024;51(2):332–335. DOI: 10.1007/s00259-023-06450-7.
135. Umer F, Adnan N. Generative artificial intelligence: Synthetic datasets in dentistry. BDJ Open 2024;10(1):13. DOI: 10.1038/s41405-024-00198-4.
136. Del Pra M. Generative adversarial networks 2023. Available from: https://medium.com/@marcodelpra/generative-adversarial-networks-dba10e1b4424.
137. Wenzel M. Generative Adversarial Networks and Other Generative Models. In: Colliot O, editor. Machine Learning for Brain Disorders. New York, NY: Humana; 2023. pp. 139–192.
138. Singh A, Ogunfunmi T. An overview of variational autoencoders for source separation, finance, and bio-signal applications. Entropy (Basel) 2021;24(1):55. DOI: 10.3390/e24010055.
139. Chen K, Shao A, Burapacheep J, et al. Conversational AI and equity through assessing GPT-3’s communication with diverse social groups on contentious topics. Sci Rep 2024;14(1):1561. DOI: 10.1038/s41598-024-51969-w.
140. Hoshi I, Shimobaba T, Kakue T, et al. Single-pixel imaging using a recurrent neural network combined with convolutional layers. Opt Express 2020;28(23):34069–34078. DOI: 10.1364/OE.410191.
141. geeksforgeeks.org/. Why Python is called Dynamically Typed?: geeksforgeeks.org/; 2018. Available from: https://www.geeksforgeeks.org/why-python-is-called-dynamically-typed/.
142. Foundation PS. gc — Garbage Collector interface 2024 Version 2: Available from: https://docs.python.org/3/library/gc.html#module-gc.
143. geeksforgeeks.org/. Introduction of Programming Paradigms geeksforgeeks.org/; 2024. Available from: https://www.geeksforgeeks.org/introduction-of-programming-paradigms/.
144. Okura Y. [[Fundamentals] 4. Visualization in Python Programming: How to Use Matplotlib and seaborn]. Nihon Hoshasen Gijutsu Gakkai Zasshi 2023;79(7):723–731. DOI: 10.6009/jjrt.2023-2228.
145. Weiler R, Diachenko M, Juarez-Martinez EL, et al. Robin’s Viewer: Using deep-learning predictions to assist EEG annotation. Front Neuroinform 2022;16:1025847. DOI: 10.3389/fninf.2022.1025847.
146. Plotly Technologies I. Dash by Plotly 2020. Available from: https://dash.plotly.com/.
147. Wickham H, Chang W, Henry L. ggplot2: Create Elegant Data Visualisations Using the Grammar of Graphics 2024. Available from: https://cran.r-project.org/web/packages/ggplot2/ggplot2.pdf.
148. Gustavsson EK, Zhang D, Reynolds RH, et al. ggtranscript: An R package for the visualization and interpretation of transcript isoforms using ggplot2. Bioinformatics 2022;38(15):3844–3846. DOI: 10.1093/bioinformatics/btac409.
149. Davy A, Borycki EM. Business intelligence dashboards for patient safety and quality: A narrative literature review. Stud Health Technol Inform 2022;290:438–441. DOI: 10.3233/SHTI220113.
150. Doyal AS, Sender D, Nanda M, et al. ChatGPT and artificial intelligence in medical writing: Concerns and ethical considerations. Cureus 2023;15(8):e43292. DOI: 10.7759/cureus.43292.
151. Popel M, Tomkova M, Tomek J, et al. Transforming machine translation: A deep learning system reaches news translation quality comparable to human professionals. Nat Commun 2020;11(1):4381. DOI: 10.1038/s41467-020-18073-9.
152. Lee PY, Salim H, Abdullah A, et al. Use of ChatGPT in medical research and scientific writing. Malays Fam Physician 2023;18:58. DOI: 10.51866/cm0006.
153. Mode I. Pandas. 2024. Available from: https://mode.com/python-tutorial/libraries/pandas.
154. Salimyanova P, Schipper V. Exploring Business Intelligence: Looker vs Tableau vs PowerBI. 2024. Available from: https://www.crystalloids.com/insights/business-intelligence-looker-vs-tableau-vs-powerbi.
155. Ordak M. ChatGPT’s skills in statistical analysis using the example of allergology: Do we have reason for concern? Healthcare (Basel) 2023;11(18):2554. DOI: 10.3390/healthcare11182554.
156. Harris CR, Millman KJ, van der Walt SJ, et al. Array programming with NumPy. Nature 2020;585(7825):357–362. DOI: 10.1038/s41586-020-2649-2.
157. Virtanen P, Gommers R, Oliphant TE, et al. SciPy 1.0: Fundamental algorithms for scientific computing in Python. Nat Methods 2020;17(3):261–272. DOI: 10.1038/s41592-019-0686-2.
________________________
© The Author(s). 2024 Open Access. This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (https://creativecommons.org/licenses/by-nc/4.0/), which permits unrestricted use, distribution, and non-commercial reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.