Skip to main content
Cureus logoLink to Cureus
. 2023 Sep 4;15(9):e44658. doi: 10.7759/cureus.44658

Reimagining Healthcare: Unleashing the Power of Artificial Intelligence in Medicine

Javed Iqbal 1,, Diana Carolina Cortés Jaimes 2,3, Pallavi Makineni 4, Sachin Subramani 5, Sarah Hemaida 6, Thanmai Reddy Thugu 7, Amna Naveed Butt 8, Jarin Tasnim Sikto 9, Pareena Kaur 10, Muhammad Ali Lak 11, Monisha Augustine 12, Roheen Shahzad 13, Mustafa Arain 14
Editors: Alexander Muacevic, John R Adler
PMCID: PMC10549955  PMID: 37799217

Abstract

Artificial intelligence (AI) has opened new medical avenues and revolutionized diagnostic and therapeutic practices, allowing healthcare providers to overcome significant challenges associated with cost, disease management, accessibility, and treatment optimization. Prominent AI technologies such as machine learning (ML) and deep learning (DL) have immensely influenced diagnostics, patient monitoring, novel pharmaceutical discoveries, drug development, and telemedicine. Significant innovations and improvements in disease identification and early intervention have been made using AI-generated algorithms for clinical decision support systems and disease prediction models. AI has remarkably impacted clinical drug trials by amplifying research into drug efficacy, adverse events, and candidate molecular design. AI’s precision and analysis regarding patients’ genetic, environmental, and lifestyle factors have led to individualized treatment strategies. During the COVID-19 pandemic, AI-assisted telemedicine set a precedent for remote healthcare delivery and patient follow-up. Moreover, AI-generated applications and wearable devices have allowed ambulatory monitoring of vital signs. However, apart from being immensely transformative, AI’s contribution to healthcare is subject to ethical and regulatory concerns. AI-backed data protection and algorithm transparency should be strictly adherent to ethical principles. Vigorous governance frameworks should be in place before incorporating AI in mental health interventions through AI-operated chatbots, medical education enhancements, and virtual reality-based training. The role of AI in medical decision-making has certain limitations, necessitating the importance of hands-on experience. Therefore, reaching an optimal balance between AI’s capabilities and ethical considerations to ensure impartial and neutral performance in healthcare applications is crucial. This narrative review focuses on AI’s impact on healthcare and the importance of ethical and balanced incorporation to make use of its full potential.

Keywords: new insights, modern medicine, newer technology in healthcare, artificial intelligence in healthcare, ai & robotics in healthcare

Introduction and background

In the modern world, the doctor-patient encounter has rapidly risen due to the introduction of artificial intelligence (AI). The cost of healthcare and the uneasy accessibility form the primary issues when it comes to the availability of healthcare in rural and urban areas. Besides, the number of treatment providers and the inability to cope with rising technology is also becoming a major problem. Therefore, to cope with the rising spread of disease and to be able to provide a more budget-friendly and timely cure, AI is the next best step [1].

The tiresome task of drug development and medicine manufacture, which required heavy investments and human intervention earlier, has now been reduced to a click. Ranging from cure identification to execution of its use has all become an AI-monitored task. Hence, effectiveness and efficiency are considerably targeted when the role of machines and new technology comes into play [2].

With accumulated laboratory technological data combined with analysis of patient history, AI proved to be nothing short of a blessing during the COVID-19 pandemic. A complete review of the post-COVID-19 syndromes helped to improve vaccine development. Thus, AI is a multifaceted approach to not just curative but also preventive measures [2].

While there is no doubt that AI has revolutionized healthcare to its core and is continuing to do so, ethical and private concerns form a major part of the demerits associated with it. The first problem is the unavailability of credible data. The second problem is algorithm dysfunction. Safe storage of data and elimination of breach of information forms the basis of patient privacy. Ethically, genomic replication or even human clone formation is resented by many cultures and religions worldwide [3]. Therefore, this paper provides insights into personalized treatment and precision medicine, drug discovery, and how AI has made its way into several domains of society. From disease management to patient monitoring, AI forms the backbone of advanced medicine [4]. Therefore, this paper focuses on answering the question, “How can AI change the future of a standard patient-physician encounter while also optimizing procedures and diagnostics aspect of it?” [1].  

Review

Selection process

We defined the criteria for selecting articles, including publication date, relevance to the topic, study design, and geographic scope, and used multiple databases (PubMed, Google Scholar, Web of Science, etc.) and keywords to search for articles related to the topic and to ensure we capture all relevant studies. Subsequently, the authors reviewed titles, abstracts, and full texts to identify potential articles meeting inclusion criteria and excluded those that did not meet the criteria. We also checked reference lists of selected articles to find additional relevant sources that may have been missed initially. Some authors evaluated the quality of selected articles based on factors such as study design, methodology, sample size, and author credibility. Finally, we extracted the key information from each selected article, including findings, methodologies, and conclusions, and organized this information systematically.

Artificial intelligence-driven diagnostics and imaging

AI employs algorithms such as machine learning (ML) and deep learning (DL) to diagnose diseases in a time-efficient manner. This helps diagnose diseases early and reduces complications [5]. A clinical decision support system (CDS) aids clinicians in diagnosing and treating diseases. CDSs such as diagnostic CDS aid in diagnosing using rules generated automatically or manually. This has been successful, as it has been tested in electronic medical records (EMRs) [6]. Newer DL methods have been used to construct contrast images utilizing the details from non-contrast imaging. A two-dimensional cycle generative adversarial network (cycleGAN) was used for this reconstruction. Development and improvement in this model can avoid the complications associated with contrast use [7].

Time is crucial for the early identification and containment of epidemics. AI-based deep Q learning auxiliary diagnosis (DQAD) was used to diagnose epidemics, reducing the time of diagnosis compared to the current methods [8]. Using chest computed tomography (CT) imaging, a deep learning model called COVID-19 detection neural network (COVNet) was built to correctly identify COVID-19 from community-acquired pneumonia and other lung conditions [9].

A highly efficient shock advice algorithm (SAA) is essential for an automated external defibrillator to diagnose and function correctly. A new SAA with DL and ML algorithms had better sensitivities and specificities than the existing SAA. Researchers used magnetic resonance imaging (MRI) histogram peaks to design AI algorithms that accurately detect tumor volumes, with improved specificity, sensitivity, and interoperator repeatability. ML models combining radiomics and clinical indicators have successfully predicted molecular subgroups of medulloblastomas and differentiated skull base chondromas from chondrosarcomas [10].

Diagnosing stroke early is essential as a delay in endovascular thrombectomy (EVT) reduces effectiveness. MethinksLVO (large vessel obstruction), an ML software, uses non-contrast computed tomography (NCCT) to identify the LVO location to facilitate the transfer of patients to EVT facilities early, thus improving outcomes [11]. The use of fully convolutional neural networks (fCNN) in reconstructing a volumetric brain from CT head films has proven to be a time-saving application [12]. AI-driven applications are used on smartphones to track health-related information and diagnose different conditions [13]. Kardia mobile cardiac monitor (KMCM) used an AI algorithm to detect heart rhythm as normal, possible Afib, and unclassified. A study putting this to the test against 12-lead electrocardiography (ECG) found that the KMCM had similar sensitivities and specificities compared to the 12-lead ECG. For unclassified rhythm, the sensitivities and specificities increase if the KMCM rhythm is interpreted by a physician [14].

The role of AI in diagnosing certain types of cancers, such as pancreatic cancer, which is often diagnosed in late stages, could be life-changing for many patients. Many challenges occur in the diagnosis, such as late symptom onset, inconsistent symptoms, and lack of molecular markers. AI tools such as K-nearest neighbor (KNN), artificial neural network (ANN), and support vector machine (SVM) are studied to identify subtle abnormal imaging findings that aid in the early detection of pancreatic cancer [15]. The AI-powered ML algorithm was used to predict the probability of developing lung cancer after initial screening with the low-dose CT scan. It matched the accuracy of predicting with the Brock scoring system, but the AI-based system was better if the radiomic scores were extreme [16]. In breast cancer (BC) screening, lesion identification and classification are the two basic processes AI offers. Early CNNs are used to identify breast lesions. Whereas deep CNN, a DL algorithm, helps identify and classify lesions by studying the images [17].

Biopsy remains the gold standard method for diagnosing skin cancers. Using DL algorithms on fluorescence spectroscopy, basal cell carcinoma (BCC) was diagnosed in real time. DSL-1 was the device employed, which used the fluorescence spectra data from BCC lesions and normal skin as a base to identify the lesions [18]. A study was done to compare the CNN, a DL algorithm, and dermatologists’ diagnosis of melanoma and found that the former achieved higher sensitivities and specificities in diagnosing melanoma and nevi. It concluded that these algorithms could assist dermatologists in diagnosing melanoma given further studies [19].

Tumor borders that are difficult to identify become challenging during surgery, such as gliomas, depending on frozen sections for delineating the boundaries in real time but at the cost of time. Fluorescent light CNN (FLCNN) was studied using the near-infrared window (NIR-II) images as an alternative or in conjunction with the surgeon’s optical judgment. The study showed that the FLCNN captured the details of the image better than the surgeon [20].

AI’s role in diagnosing neurological diseases is being explored further [21]. Battineni et al. studied four different ML algorithms, namely, naïve Bayes (NB), ANN, KNN, and SVM, individually and in combination to predict Alzheimer’s disease (AD). The study concluded that the combined models, with limited features, predict AD at the initial stages with greater accuracy [22]. Various ML and DL algorithms are studied for their use in diagnosing other neurological diseases such as Parkinson’s disease (PD), major depressive disorder (MDD), attention-deficit hyperactive disorder (ADHD), autism spectrum disorder (ASD), and psychiatric disorders such as schizophrenia [23].

Currently, the diagnosis of obstructive sleep apnea (OSA) is made by polysomnography (PSG) which employs the apnea-hypopnea index (AHI). AI-based CNN algorithm uses a single-lead ECG signal to diagnose OSA with reasonable accuracy, sensitivity, and specificity. This method is easy and efficient but still needs PSG to confirm if the estimated risk is high [24]. AI’s role in diabetes care is to ease the aspect of monitoring. The self-monitoring system uses AI to detect hypoglycemia one hour prior and notifies the patient. AI is also used in retinal screening, diabetes diagnosis support, and risk categorization [25].

The use of AI in diagnosing hematologic disease is on the rise. Cellavision, a peripheral blood smear analyzer approved by the Food and Drug Administration (FDA), utilizes the ML-ANN algorithm to perform tests such as complete blood count, cell differential, identification, and location [26]. Morphogo, a CNN-based bone marrow analyzer, identifies and classifies hematopoietic tumor cells. It also identifies non-hematopoietic metastatic cells with greater accuracy and specificity [27]. Other systems such as Scopio, Mantiscope, and Vision Hema use different AI algorithms to diagnose hematologic conditions [28]. Cytogenetics remains the primary diagnostic testing for genetic disorders, including malignancies. A new AI-based tool called chromoEnhancer uses Generative adversarial networks (GAN) to produce good-quality karyograms. This helps detect the abnormality efficiently [29].

AI-driven diagnostics and imaging are proliferating and finding applications in every possible field of medicine. Further exploration into these algorithms can be assimilated to provide better patient care, considering the ethics and social regulations [30].

Personalized treatment and precision medicine

Artificial Intelligence in Genomics and Its Role in Personalized Treatment Plans

Personalized medicine is a method that considers an individual’s unique genetic makeup, environment, and lifestyle for treating and preventing diseases [31]. It shifts away from traditional diagnostic methods that rely on shared patient characteristics, aiming to transform healthcare through prevention, personalization, and precision [32]. To achieve this, various data sources such as genomics, biological data, transcriptomics, and proteomics play a crucial role [33]. Big data is essential in this process, and researchers are increasingly exploring the potential of AI for robust data analysis [34,35]. These advanced tools empower us to identify precision intervention targets and personalized medical approaches for each patient [36].

ML is a method that employs mathematical algorithms to construct models using datasets; these models can enhance their performance through experience [35]. Among the various ML techniques, DL utilizes neural networks, allowing end-to-end learning directly from raw data and seamless integration of different data types [37].

Functional genomics explores the intricate interplay between genetic attributes and environmental conditions. This endeavor has been significantly aided by the utilization of advanced deep architectures [37]. For instance, a notable example is DeepBind, a fully autonomous standalone software designed to predict the sequence specificities of DNA and RNA binding proteins [38]. Another noteworthy tool, DeepSea (a DL-based sequence analyzer), employs DL techniques to forecast the chromatin effects of sequence alterations at the single-nucleotide level. By learning from extensive chromatin profiling data, DeepSea effectively identifies the functional implications of non-coding variants. It is worth noting that the majority of disease-associated single-nucleotide polymorphisms (SNPs) reside within the non-coding regions of the genome. Additionally, the groundbreaking DeepMind AlphaFold method has revolutionized protein structure prediction, fundamentally altering a once-daunting challenge into a resolved matter [39,40].

Significant strides have been made in applying functional genomics to enhance precision medicine for prevalent non-cancerous conditions, such as kidney disease [32]. An example of an ML application using tissues, Liu et al. used in silico nano dissection, an ML algorithm, to study mRNA expression in glomeruli from IgA nephropathy patients [41]. They identified cell-specific genes differentially expressed compared to healthy controls. In cancer, AI applied to genomics, some classification models have been developed to help stratify individuals into high-risk and low-risk groups. Vural et al. implemented an unsupervised clustering approach to identify subgroups based on individual omics systems; the resulting model successfully identified three groups of breast cancer patients and associated specific genes with disease progression [42]. In another study by He et al., researchers focused on the identification of cancer-selective combinatorial therapies for ovarian cancer using supervised ML algorithms. They analyzed genomic and expression data from ovarian cancer patients and pan-cancer markers to predict drug targets and mutations in the study to identify top-ranked drug combinations with high accuracy, offering potential therapeutic options for personalized treatment [43].

ML models have the potential to analyze multimodal data obtained from electronic medical records and other curated sources to identify patients who could benefit from early treatment or participation in randomized control trials of innovative interventions [44]. For instance, in a study by Dong et al., extensive pharmacogenomics analyses were conducted across more than 1,000 cell lines. The aim was to gain insights into the mechanisms underlying the responses of anticancer drugs. In this research, a sophisticated SVM model was developed and evaluated, utilizing genomic data to accurately predict the sensitivity of anticancer drugs [45].

Drug Discovery and Development With Artificial Intelligence Assistance

AI is already deeply embedded in almost every facet of modern society. In exchange, it is undergoing rapid innovation. Because of this, it is rapidly gaining importance in the pharmaceutical industry, particularly in the realm of drug research and development [46].

Drug development is time-consuming and labor-intensive, predicated on iterative trial-and-error experimentation and high-throughput screening processes. On the other hand, using algorithms such as ML and natural language processing (NLP) will significantly speed up the entire process while also requiring substantial data inputs. Errors may be reduced, and the quality of analytical findings improved by streamlining the process [47].

Using the DL method makes it easier to ascertain the effectiveness of a substance. Using AI techniques, we can also determine if a medicine is toxic [48]. How AI may generate and synthesize novel molecules has been studied, opening the door to using AI to create new medicines [49]. Nonetheless, there may be ethical considerations related to medicine that limit its use [50].

AI makes use of a number of methods, including reasoning, knowledge representation, problem resolution, and a model of ML. This analysis will focus on the SVM classification model. SVM is a supervised machine learning module that generates analytical outputs by classifying two distinct sets of troublesome subgroups using a hyperplane between their extreme edges. Therefore, with the use of a decision boundary or hyperplane, the two sets can be organized into usable data [51].

ANNs are also used in the domain of ML known as DL. These can be described as a collection of high-tech computers that mimic the way in which the human brain sends and receives electrical signals by using perceptons that are analogous to biological neurons. An ANN is a collection of nodes that accept various inputs and convert them to the output, either individually or in a multi-link configuration, using algorithms to solve problems [52].

Anticipating global infectious disease hazards using AI-based surveillance models may be very helpful [53,54]. The key to building an effective surveillance system is to use an integrated modeling method that incorporates many kinds of individual data models, such as travel, cell phone location tracking, epidemiology, and behavioral pattern data [55]. A single person’s level and, more significantly, mass collecting awareness of the intended population is essential for this modeling technique to effectively control the transmission of infection [56]. Possible applications for such an integrated model-based platform include anticipatory surveillance and the identification of hazards posed by infectious illnesses of global importance. The use of mathematical models that can assess the transmission of infectious illnesses in mass gatherings and simulate the impact of public health measures at both the national and international levels is also encouraged by such integrated techniques [57].

Artificial intelligence-enabled patient monitoring and disease management

Public health surveillance faces data sourcing and analytics challenges in identifying reliable signals of health anomalies and disease outbreaks from data sources. AI helps tackle these challenges by enabling and enhancing public health surveillance through methods and techniques such as DL, reinforcement learning, knowledge graphs, Bayesian networks, and multiagent systems. AI enhances surveillance of various open novel or unexplored data, unstructured and semi-structured data, complex spatiotemporal data, and epidemics’ evolution, which is difficult with traditional surveillance methods [55].

Remote patient monitoring (RPM) or telemonitoring uses digitally transmitted health-related data to improve patient care through education, early disease decompensation detection, intervention, and enhanced patient-physician relationship [58]. Telemonitoring technology uses digital health platforms to enable patient evaluation outside a typical clinical visit [59]. A meta-analysis found that type 2 diabetes patients followed through mobile technologies for RPM and clinical advice delivery had improvement in glycemic control by 0.8% compared to standard care [60]. The application of AI in RPM includes monitoring vital signs, physical activities, diabetes, and mental health. AI can build patient-specific models using ML techniques for patients in the emergency department and ICU [61]. In a retrospective study by Taylor et al., a local big data-driven ML approach had better in-hospital ED patient mortality than existing clinical decision rules and traditional analytical techniques [62].

An ML approach predicted cardiac arrest within 24 hours more accurately than the traditional modified early warning scores for critically ill patients in the emergency department [61]. A retrospective cohort study by Antunes et al. showed that the new chronic liver failure Consortium Acute Decompensation Score (CLIF-C ADs) machine has statistically significant superiority in predicting 30-day mortality over traditional models [63]. RPM is being increasingly adopted following the COVID-19 pandemic [59].

Vipin et al. proposed an edge AI-enabled Internet of Things (IoT) healthcare monitoring system for the real-time scheduling of patients and provision of resources on a priority basis depending on the patient’s condition. This tool collects and transmits data and triggers appropriate action on other integrated devices, simplifying the process of monitoring and assistance provision, which is particularly useful in the elderly or disabled, and in pandemic situations [64].

The FDA issued Emergency Use Authorization for wearable and mobile ECG technologies to record QT intervals in patients taking hydroxychloroquine or azithromycin. However, AI-ECG technologies need further evaluation for external validity in diverse populations to be applied in routine clinical practice [65]. AI-based surveillance can be used to identify and disperse need-based support. However, unequal representation and opportunities can lead to malfunctioning and unpredicted health risks [66]. For example, biased distribution of health surveillance technology between countries can impact the efficacy of the tools, resulting in unfairness [67]. Potential biases can be introduced willingly or unconsciously. For instance, including objective perceptions and opinions can lead to bias in the incorporation of data and can affect the conclusions [68]. Algorithmic design development with the right fairness for a context poses a challenge [69]. An example is disparities in the distribution and access to COVID-19 vaccines in the United States due to disharmonized priority settings in the algorithm [70]. Data privacy and security, transparency, fairness of the analytical AI models, and limited computing power of sensing processors are a few other technical and regulatory challenges with wearable sensors with AI-at-the-edge functionality [71].   Phenotypic personalized medicine (PPM) aims to provide a quantitative representation of an individual patient’s overall health, while directly translating this knowledge into clinical care helps to find a suitable drug combination quadratic phenotypic optimization platform (QPOP) and dosing (CURATE.AI) based on data from previously treated patients. PPM-based approaches have produced superior results with limited side effects and maximized desired output than standard care [72]. Applications empowered by AI hold the capacity to assist in devising treatment plans for individuals with valvular heart conditions. AI-driven clinical decision systems enhance procedure planning of aortic valve replacement by integrating anatomical data, aiding valve size/type determination with speed and accuracy. It is beneficial for high-volume sites where manual imaging analysis is time-consuming and variable [73,74].

Artificial intelligence for remote and telemedicine

Telemedicine refers to delivering healthcare services using information and communications technologies, considering distance and accessibility as critical factors. Initially envisioned for tech-savvy patients, the COVID-19 pandemic propelled telemedicine into mainstream healthcare globally [75]. Its scalability enabled essential health and safety protocols during the crisis and established it as a sustainable approach beyond the pandemic. DL is favored due to its capacity for autonomous learning and adaptation [75]. AI algorithms have been extensively studied and implemented in various medical fields to streamline exam interpretations, enhance diagnostic accuracy, and reduce time and workforce requirements [76]. In modern medicine, patient data includes binary and numerical information and vast image data. Traditional computer programs often struggle to interpret such detailed reports. However, modern AI has demonstrated its efficiency in processing and analyzing complex image data [76].

A typical deep neural network utilized for processing and analyzing visual images is the CNN [76]. This network comprises multiple processing layers, allowing it to extract meaningful patterns and features from visual data [77]. In ophthalmology, diabetic retinopathy (DR) is a common complication of diabetes mellitus (DM) and a leading cause of preventable blindness [77]. The World Health Organization (WHO) recommends annual screening with dilated funduscopy for DM patients to prevent vision loss [78]. However, the significant number of DR patients poses challenges regarding screening, timely referral, and treatment due to the availability of human assessors and long-term financial sustainability. As a solution, developing DL and digital technology is advocated to facilitate the DR management process [76].

Similarly, age-related macular degeneration (AMD) is a significant cause of vision loss among the elderly population globally. However, the growing number of AMD patients and frequent follow-ups create a pressing need for a robust automatic mechanism. DL and telemedicine offer potential solutions for AMD management, including initial screening, subsequent monitoring, and treatment prediction. DL algorithms can assess visual acuity (VA) and optical coherence tomography (OCT) findings, aiding in determining AMD treatment strategies [79].

In the mental health domain, a smartphone-sensing wearable device was proposed to assess behavior and identify depressive and manic states in patients with bipolar disorder [80]. Patient monitoring is also advanced using AI [75]. In 2016, a research project focused on detecting fetal ECG signals and other physiological data of a fetus to enhance the monitoring and identification of fetal conditions during pregnancy [81]. Similarly, in 2017, the introduction of Wi-Mon, a wireless continuous patient monitoring system, addressed the need for ongoing patient monitoring by employing the Wireless Body Area Network (WBAN) concept, particularly for individuals requiring it, such as dengue patients [82]. Furthermore, a mobile cyber-physical system for healthcare is recommended using wearable devices to enable continuous patient monitoring and provide real-time updates to doctors or family members on the patient’s status [83].

Telemedicine, especially in dermatology, has rapidly embraced AI due to increasing demand, the necessity for high-quality images, and the availability of advanced technology [84]. While in-person diagnosis remains more accurate than teledermatology, telemedicine offers the potential to enhance dermatological care by providing convenience, reduced waiting times, and diagnostic support during case reviews [85]. In face-to-face assessments, AI delivers real-time diagnostic support, significantly enhancing dermatologists’ accuracy in identifying skin disorders [86]. For non-dermatologists, who may have varying expertise in skin conditions, AI-assisted diagnosis can be particularly valuable, especially in cases such as diagnosing melanoma [87].

In neurosurgery, telemedicine demonstrates its potential with applications such as telehealth stroke triage and ML-based prognostics for neuro-trauma [88]. Recently, a pediatric neurosurgery team utilized ML to screen for craniosynostosis by analyzing digital photographs using semiautomated image analysis software [89]. The ML algorithms accurately identified common types of craniosynostosis, highlighting the effectiveness of this AI-augmented telemedicine approach. This innovative method enhances access to screening for head shape abnormalities, enables early diagnosis, and reduces the burden of ambulatory appointments for families. Primary care providers and non-specialists can also use this approach to identify patients who may require surgical referral [90].

In diabetes management, several projects have emerged utilizing new information and communication technologies and Web 2.0 technologies for automatic data transmission and remote interpretation of patient information [91-93]. A notable aspect of these telemedicine projects is the widespread incorporation of ML, which significantly enhances their capabilities and effectiveness [91-93]. AI algorithms enable more efficient and accurate data analysis, empowering healthcare professionals to make informed decisions and provide personalized care to diabetes patients.

Artificial intelligence-driven behavioral health interventions

Expanding access to digital mental health therapies, the increasing need for psychological services, along with the development of AI, has led to the rise of digital mental health interventions (DMHsI) in recent years. Patient evaluation, symptom evaluation, modifying behaviors, and information dissemination are just a few of the areas where AI-powered chatbots are currently being employed in DMHIs. Chatbots may be as basic as algorithms based on rules such as ELIZA or as complex as AI models that use ML and language processing methods [94].

Support, screening, training, intervention, surveillance, and avoiding recurrence are just a few of the chatbot-based services offered inside DMHIs. Using chatbots for mental health diagnosis and triage might reduce the burden on healthcare workers and expedite the treatment of individuals who need it the most [95]. Chatbots have been used to test for and diagnose a wide range of mental health conditions, including dementia, drug addiction, stress, depression, attempted suicide, anxiety disorders, and symptoms of post-traumatic stress disorder. In this context, the chatbot is used as a conversational counterpart. After the user enters the symptoms they are experiencing into the chatbot, it gives them a medical diagnosis, prognosis, and therapeutic plan (as well as information concerning the diagnosis) [96-105].

Overall, 51% of mental health professionals surveyed thought using chatbots for diagnostic purposes was a challenge. AI-assisted diagnostics, on the other hand, might aid in the early identification of those at risk, allowing for more effective intervention and problem avoidance [106,107]. To properly predict the onset of certain mental health diseases, other kinds of AI may also be able to analyze trends in behavior and information supplied by the person. For instance, ML algorithms possess a 79% success rate in predicting the start of psychosis and a success rate of 96% in identifying the first signs of attention-deficit disorder (ADD)/ADHD and ASD [108].

For instance, certain chatbots leveraging NLP in nature may mimic a therapeutic conversational approach for the purpose of deploying and educating users about different therapies [109]. The most studied and widely used kind of chatbot is one based on cognitive behavioral therapy (CBT), as shown by a meta-analysis that found 10 out of 17 chatbots primarily used CBT [110]. One such chatbot, Woebot, employs NLP to mimic real physicians and social discourse to give CBT to customers through instant messaging [111]. Happify Health, an ad-funded digital mental health provider, added a chatbot this year. Anna is a chatbot powered by AI that simulates a therapist’s job and provides several Happify activities. Anna was designed to work in tandem with other aspects of the Happify program, as opposed to some of the other DMHIs chatbots which operate as separate applications (i.e., the chatbot itself is the intervention). Happify was designed to disseminate enhanced forms of evidence-based activities culled from many therapeutic modalities including CBT, stress reduction using mindfulness, and positive psychology. The activities are divided into tracks to assist users in zeroing in on a specific issue, such as stress management [112].

Anna may be found in these tracks. After selecting a route that includes Anna, users are welcomed by the chatbot with a personalized video message and explanation of the chatbot’s capabilities. Listeners may also be polled by Anna with specific questions aimed at collecting information for refining the personalized playlist. To increase user involvement and the likelihood that they will complete the activities as intended, Anna also provides particular assignments inside these tracks. So, Anna helps make the most of the benefits of that move. Anna tracks the elements that contribute most to successful outcomes for each activity, evaluates responses based on those elements, and then prompts participants to fill in any gaps in their initial remarks. Whenever a user’s first response to a task requiring gratitude does not demonstrate an adequate amount of gratitude, Anna will ask for clarification. Anna guides users on how to make better use of the platform and listens to their feedback to allay concerns that conversations with chatbots are not engaging enough [113].

Artificial intelligence in medical education and training

With the increased integration of AI in healthcare, electronic health records (EHRs) can be used for novel techniques such as data processing and enhanced decision-making, prompting accurate data input into EHRs by physicians to maximize the benefits of AI [114]. These shifts into the AI era can positively impact clinical practice and patient health outcomes by training healthcare personnel with better knowledge and skills [115,116]. A single error in AI can cause harm to thousands of patients with wide usage, highlighting the importance of accurate data extraction and input by clinicians to train the AI sufficiently [117]. Therefore, future physicians must understand data management and AI applications. Medical education should incorporate topics about AI, data science, EHR basics, and AI-related ethics into students’ and physicians’ curricula, progressively educating them in their academic journey [118]. Medical schools and healthcare organizations should invest in infrastructure and technological support and collaborate with AI regulatory institutions to efficiently utilize AI [119].

A scoping review by Nagi et al. on current applications of different AI methods such as ML, robotics training, and virtual reality (VR) in several domains of medical education has shown enhanced practical skills of medical students following the implementation of AI-based assistance. Through VR, students can train their decision-making skills in surgical and medical procedures in a controlled and safe manner [116]. Despite the investment requirement in costly hardware infrastructure, the training is still cost-effective compared to traditional manikin/actor-based training in the long run [120]. AI can significantly benefit microsurgical education and ophthalmic surgeon training simulations and video content analytics in a low-risk manner. VR head-mount displays offer several possibilities for operating room simulation [121].

Creutzfeldt et al. trained 12 Swedish medical students in cardiopulmonary resuscitation by avatars to better understand their reactions and experiences using a multiplayer virtual world. The positive aspects of learning cardiopulmonary resuscitation were confirmed after a data-driven approach in qualitative methodology. Further clinical performance should be analyzed to rule out erroneous self-belief bias. Participants noted insufficient psychomotor skills and lack of stress to be unrealistic, which could affect their real-world performance [122].

A meta-analysis by Zhang et al. showed that AI could also be applied in medical education at different stages, such as teaching implementation, evaluation, and feedback. The quality of teaching can be assessed with feedback and evaluation from AI. However, it is challenging to verify the effectiveness of AI implementation [123]. Teachers must be appropriately trained to utilize AI in teaching [124]. It is essential to review the ability of AI compared to physicians with expertise in making healthcare-related decisions. Chat Generative Pre-trained Transformer (ChatGPT), a server-contained language model, was evaluated with 350 US Medical Licensing Exam (USMLE) sample test questions without virtual assets. The questions were formatted and input in the sequence of open-ended prompting, multiple-choice single answer without and with forced justification prompting with a new chat session for every entry to eliminate memory retention bias. ChatGPT scored passing performance for all three steps of USMLE with high internal concordance and significant insight in explanations. The study concluded that large language models have significant potential to assist medical education and clinical decision-making [125].

Although ChatGPT has revolutionizing potential in medical education, it cannot completely replace hands-on clinical experience and mentorship. Moreover, algorithms use large databases, which may contain biases resulting in biased system output that could lead to a loss of fairness in treating minority or underrepresented groups. Therefore, AI systems output should be monitored to ensure the absence of bias and eliminate any biases present [126]. Further research can help with the issue of bias considerably [127].

The use of AI is not without its challenges, such as black box problems [128], data privacy, and liability issues [129]. Hence, physicians should know about the strengths, limitations, and appropriate usage of AI tools [130,131].

Artificial intelligence ethical considerations

Healthcare is a high-risk field in which mistakes can lead to severe consequences for the patient. Patients meet with physicians at a time when their overall health is already compromised, leaving them vulnerable [132]. Therefore, the use of AI in healthcare must undergo rigorous testing and study before becoming a significant part of clinical management. The four primary ethical issues concerning AI in medicine are safety and transparency, algorithmic fairness and biases, informed consent for the use of personal data and health records, and data privacy [124].

The main legal concern is the limited transparency of algorithms used in AI, referred to as the black box issue, where the algorithms responsible for the outputs are not disclosed [133,134]. Transparency, particularly accessibility and comprehensibility in AI design and governance, provides vital information that should be attainable and understandable. However, details about the functionality of these algorithms are often intentionally made difficult to obtain [135]. Successful scrutiny of results or outputs obtained from AI systems is hindered as the thinking process that led to them is concealed by modern computing approaches.

The FDA has already approved autonomous AI diagnostic systems based on ML. These ML-healthcare applications (ML-HCAs) create algorithms from large datasets and make predictions without requiring explicit programming to avoid biases and errors [133]. While this new technology has the potential to revolutionize healthcare, it must be properly designed with effective safety measures to ensure data privacy and prevent inaccuracies and breaches [134].

Privacy protection poses challenges to ML. ML necessitates large datasets for effective training, leading to high accuracy in results. However, when significant amounts of data are used to train ML applications from multiple sources, the data’s origin can ultimately be traced back to the patients, negating attempts at privacy [134]. Ownership and rights to datasets containing patient information are complex and vary by jurisdiction and the degree to which data has been de-identified or anonymized [136,137].

In Europe, patients own and have usage rights to all their data, whereas healthcare providers in North America may own the physical evidence related to patient data [138,139]. With more protective regulations such as the General Data Protection Regulation (GDPR) [135] in the European Union (EU), patients may be hesitant to allow their data to be used for training AI models. This could lead companies to seek less regulated environments with poorly defined and enforced data ownership regulations. The GDPR might discourage US companies from conducting business due to these restrictions, and EU citizens may not be able to enforce their privacy rights in other jurisdictions [140].

Accountability for AI errors remains a gray area, with no clear individual or entity held responsible. Consequently, AI severely limits the ability to assign blame and/or ownership of the decision-making process [140]. The lack of liability in such a scenario raises concerns regarding professional obligations and patient safety when using unverified AI systems in a clinical setting. To address this, verification and validation are essential when clinicians use AI systems. Managers overseeing AIS users should clarify that physicians cannot avoid accountability by blaming AI systems [132]. Automation bias can occur when a mostly accurate AI system leads medical practitioners to become complacent, causing individuals, both patients and doctors, to accept the system’s results without questioning its limitations. A study by Mirsky and colleagues demonstrated how malware and AI DL tools could mislead expert radiologists about the presence or absence of malignant lung lesions on CT scans [141].

In the study conducted by Buolamwini and Gebru, selection bias was evident in automated facial recognition and the associated datasets, resulting in reduced accuracy in recognizing darker-skinned individuals, particularly women. The datasets used by MLs are derived from specific populations. Consequently, when these AI systems are applied to underserved or underrepresented patient groups, they are more likely to produce inaccurate results [142]. Due to the availability of extensive datasets focusing on common diseases, AI might overlook rare or uncommon conditions that a radiologist could diagnose on X-rays or CT scans [143]. Using models based on training data that do not adequately represent the population, case mix, modalities, and acquisition protocols can cast doubt on the performance and confidence in its use [144].

The contentious debate surrounding the legality of AI systems has prompted governing bodies such as the European Parliament to pass the GDPR [135], aimed at proactively addressing morally challenging situations that may arise from implementing AI in healthcare [145,146]. Integrating AI into medical practice [147-149] requires a well-established governance framework that safeguards patients from harm [150,151]. AI must always adhere to ethical principles to ensure the well-being of patients and should adapt to continuously changing environments with frequent disruptions [151]. As AI becomes more integral to healthcare delivery, AIS and ML-HCAs must meet all ethical requirements and be free from unjust biases. Eventually, AI is likely to surpass humans in some healthcare fields, and refraining from its use would be both unethical and unscientific [152].

Challenges and future directions

AI is a rapidly evolving technical discipline that plays a crucial role in medicine. However, the implementation of AI faces several challenges, including technical, ethical, safety, and financial concerns. AI technology offers significant opportunities for diagnostic and treatment purposes in medicine, but it requires overcoming challenges, particularly related to safety [153]. One of the most significant challenges in medical AI applications, especially in computational pathology, is the lack of labeled data, which is exacerbated by the multi-gigapixel nature of images and high data heterogeneity [154].

The integration of AI into cancer research shows promising outcomes and is currently addressing challenges where medical experts struggle to control and cure cancer. AI provides tools and platforms that enhance our understanding and approach to tackling this life-threatening disease. AI-based systems aid pathologists in diagnosing cancer more accurately and consistently, thereby reducing error rates [155]. AI has also demonstrated high effectiveness in various aspects of drug design and development. However, significant challenges persist, including the acquisition of high-quality data [156].

Despite the considerable advantages and notable progress in contemporary methodologies, rule-based systems commonly employed in medical AI exhibit limitations. These encompass substantial expenses associated with development, constraints arising from the intricate representation of multifaceted connections, and the need for extensive medical expertise [155].

While AI’s involvement in direct patient care is currently limited, its expanding role in complex clinical decision-making processes is on the horizon. Establishing fundamental guidelines for AI’s scope, transparent communication with patients for informed consent, and comprehensive evaluation of AI’s implementation are crucial steps toward setting a universal standard. This emphasizes the importance of algorithmic transparency, privacy protection, stakeholder interests, and cybersecurity to mitigate potential vulnerabilities. Proactive leadership from professional organizations can play a vital role in building public trust in the safety and efficacy of medical AI, thereby driving further advancements in this promising domain [157].

AI has the potential to collaborate with other digital innovations such as telemedicine, enabling virtual consultations and the implementation of the Internet of Medical Things to enhance referral procedures. AI’s capabilities extend to precise risk stratification of cancer stages and the selection of suitable treatment paths. Shared factors are evident across multiple tiers during data analysis, enabling AI models to identify causal links between variables. By amalgamating this knowledge, AI research can achieve higher precision in examining cancer-related medical incidents. In the future, comprehensive databases will emerge, generating new data sets encompassing every aspect of human health. These repositories will empower highly intricate models, capable of personalizing therapy choices, precise dosage calculations, surveillance strategies, timetables, and other pertinent factors [155].

Conclusions

Prominent AI technologies such as ML and DL have immensely influenced diagnostics, patient monitoring, novel pharmaceutical discoveries, drug development, and telemedicine. Significant innovations and improvements in disease identification and early intervention have been made using AI-generated algorithms for clinical decision support systems and disease prediction models. AI has changed the face of healthcare through unique advancements in patient monitoring, diagnosis, and treatment planning. AI has made strides in optimally implementing clinical drug trials, telemedicine, and personalized therapeutic regimes. However, ethical issues regarding data protection and transparency of AI-driven algorithms remain a cause for concern. Although AI holds promise in mental health interventions, medical education, and virtual training, its role must be aligned with human expertise. The main challenge is to optimize AI’s transformative impact while keeping ethical and regulatory principles in line.

Acknowledgments

JA contributed substantially to the conception of the work, drafted it for intellectual content, approved the final version, and agreed to be accountable for the accuracy and integrity of the work. DCCJ designed the work, critically reviewed it, approved the final version, and ensured accountability. PM and SS played a key role in data acquisition, drafting, approval of the final version, and accountability. SH analyzed data, critically reviewed the work, approved the final version, and ensured accountability. TRT interpreted the data, drafted the work, approved the final version, and maintained accountability. ANB was involved in the conception and design, critical review, final approval, and accountability. JTS and PK contributed to the data acquisition, analysis, interpretation, drafting, final approval, and accountability. MAL contributed to the conception or design, critical review, and final approval and ensured that questions related to accuracy or integrity were resolved. MA, RS, and MA, respectively, played substantial roles in data acquisition, data analysis, and data interpretation, in addition to drafting, final approval, and accountability. These contributions align with the ICMJE recommendations, and each author has met all four criteria.

The authors have declared that no competing interests exist.

References

  • 1.AI-driven optimization in healthcare: the diagnostic process. Lyon J, Bogodistov Y, Moormann J. Eur J Manage Issues. 2021;29:218–231. [Google Scholar]
  • 2.Evolving scenario of big data and artificial intelligence (AI) in drug discovery. Tripathi MK, Nath A, Singh TP, Ethayathulla AS, Kaur P. Mol Divers. 2021;25:1439–1460. doi: 10.1007/s11030-021-10256-w. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 3.Drawbacks of artificial intelligence and their potential solutions in the healthcare sector. Khan B, Fatima H, Qureshi A, Kumar S, Hanan A, Hussain J, Abdullah S. Biomed Mater Devices. 2023:1–8. doi: 10.1007/s44174-023-00063-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 4.Artificial intelligence in breast cancer screening and diagnosis. Dileep G, Gianchandani Gyani SG. Cureus. 2022;14:0. doi: 10.7759/cureus.30318. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 5.A deep learning approach to generate contrast-enhanced computerized tomography angiograms without the use of intravenous contrast agents. Chandrashekar A, Shivakumar N, Lapolla P, et al. Eur Heart J. 2020;41:0. [Google Scholar]
  • 6.Assessing the accuracy of an automated atrial fibrillation detection algorithm using smartphone technology: the iREAD Study. William AD, Kanbour M, Callahan T, et al. Heart Rhythm. 2018;15:1561–1565. doi: 10.1016/j.hrthm.2018.06.037. [DOI] [PubMed] [Google Scholar]
  • 7.Using artificial intelligence to detect COVID-19 and community-acquired pneumonia based on pulmonary CT: evaluation of the diagnostic accuracy. Li L, Qin L, Xu Z, et al. Radiology. 2020;296:0–71. doi: 10.1148/radiol.2020200905. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 8.Deep learning based software to identify large vessel occlusion on noncontrast computed tomography. Olive-Gadea M, Crespo C, Granes C, et al. Stroke. 2020;51:3133–3137. doi: 10.1161/STROKEAHA.120.030326. [DOI] [PubMed] [Google Scholar]
  • 9.AI-driven decision making for auxiliary diagnosis of epidemic diseases. Lin K, Liu J, Gao J. IEEE Transact Mol Biol Multi-Scale Commun. 2022;8:9–16. [Google Scholar]
  • 10.The future of artificial intelligence in neurosurgery: a narrative review. Iqbal J, Jahangir K, Mashkoor Y, et al. Surg Neurol Int. 2022;13:536. doi: 10.25259/SNI_877_2022. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 11.Deep feature learning for sudden cardiac arrest detection in automated external defibrillators. Nguyen MT, Nguyen BV, Kim K. Sci Rep. 2018;8:17196. doi: 10.1038/s41598-018-33424-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 12.A survey on AI techniques for thoracic diseases diagnosis using medical images. Mostafa FA, Elrefaei LA, Fouda MM, Hossam A. Diagnostics (Basel) 2022;12:3034. doi: 10.3390/diagnostics12123034. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 13.AI-driven clinical decision support: enhancing disease diagnosis exploiting patients similarity. Comito C, Falcone D, Forestiero A. IEEE Access. 2022;10:6878–6888. [Google Scholar]
  • 14.Deep neural networks are superior to dermatologists in melanoma image classification. Brinker TJ, Hekler A, Enk AH, et al. Eur J Cancer. 2019;119:11–17. doi: 10.1016/j.ejca.2019.05.023. [DOI] [PubMed] [Google Scholar]
  • 15.Santosh K, Gaur L. Artificial Intelligence and Machine Learning in Public Healthcare. Singapore: Springer; 2021. AI solutions to public health issues; pp. 23–32. [Google Scholar]
  • 16.Computational radiology in breast cancer screening and diagnosis using artificial intelligence. Tran WT, Sadeghi-Naini A, Lu FI, et al. Can Assoc Radiol J. 2021;72:98–108. doi: 10.1177/0846537120949974. [DOI] [PubMed] [Google Scholar]
  • 17.Artificial intelligence-driven diagnosis of pancreatic cancer. Hameed BS, Krishnan UM. Cancers (Basel) 2022;14:5382. doi: 10.3390/cancers14215382. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 18.Extraction of brain tissue from CT head images using fully convolutional neural networks. [ Mar; 2018 ];Akkus Z, Kostandy PM, Philbrick KA, Erickson BJ. SPIE Medical Imaging. 2018 :1057420. [Google Scholar]
  • 19.Predicting malignant nodules from screening CT scans. Hawkins S, Wang H, Liu Y, et al. J Thorac Oncol. 2016;11:2120–2128. doi: 10.1016/j.jtho.2016.07.002. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 20.Preoperative AI-driven fluorescence diagnosis of non-melanoma skin cancer. Andreeva V, Aksamentova E, Muhachev A, et al. Diagnostics (Basel) 2021;12:72. doi: 10.3390/diagnostics12010072. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 21.Artificial intelligence for Alzheimer's disease: promise or challenge? Fabrizio C, Termine A, Caltagirone C, Sancesario G. Diagnostics (Basel) 2021;11:2146. doi: 10.3390/diagnostics11081473. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 22.A sleep apnea detection system based on a one-dimensional deep convolution neural network model using single-lead electrocardiogram. Chang HY, Yeh CY, Lee CT, Lin CC. Sensors (Basel) 2020;20:4157. doi: 10.3390/s20154157. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 23.A comprehensive machine-learning model applied to magnetic resonance imaging (MRI) to predict Alzheimer's disease (AD) in older subjects. Battineni G, Chintalapudi N, Amenta F, Traini E. J Clin Med. 2020;9:2146. doi: 10.3390/jcm9072146. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 24.Application of artificial intelligence in the MRI classification task of human brain neurological and psychiatric diseases: a scoping review. Zhang Z, Li G, Xu Y, Tang X. Diagnostics (Basel) 2021;11:1402. doi: 10.3390/diagnostics11081402. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 25.Machine learning and deep learning in medical imaging: intelligent imaging. Currie G, Hawk KE, Rohren E, Vial A, Klein R. J Med Imaging Radiat Sci. 2019;50:477–487. doi: 10.1016/j.jmir.2019.09.005. [DOI] [PubMed] [Google Scholar]
  • 26.Artificial intelligence in current diabetes management and prediction. Nomura A, Noguchi M, Kometani M, Furukawa K, Yoneda T. Curr Diab Rep. 2021;21:61. doi: 10.1007/s11892-021-01423-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 27.Performance evaluation of the CellaVision DM96 system: WBC differentials by automated digital image analysis supported by an artificial neural network. Kratz A, Bengtsson HI, Casey JE, et al. Am J Clin Pathol. 2005;124:770–781. doi: 10.1309/XMB9-K0J4-1LHL-ATAY. [DOI] [PubMed] [Google Scholar]
  • 28.Detection of metastatic tumor cells in the bone marrow aspirate smears by artificial intelligence (AI)-based Morphogo system. Chen P, Chen Xu R, Chen N, et al. Front Oncol. 2021;11:742395. doi: 10.3389/fonc.2021.742395. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 29.Artificial intelligence-assisted diagnostic cytology and genomic testing for hematologic disorders. Gedefaw L, Liu CF, Ip RK, Tse HF, Yeung MH, Yip SP, Huang CL. Cells. 2023;12:1755. doi: 10.3390/cells12131755. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 30.ChromoEnhancer: an artificial-intelligence-based tool to enhance neoplastic karyograms as an aid for effective analysis. Bokhari Y, Alhareeri A, Aljouie A, et al. Cells. 2022;11:2244. doi: 10.3390/cells11142244. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 31.Medicina di precisione e medicina di laboratorio. Cappelletti P. Riv Ital Med Lab. 2016;12:129–133. [Google Scholar]
  • 32.On the role of artificial intelligence in genomics to enhance precision medicine. Álvarez-Machancoses Ó, DeAndrés Galiana EJ, Cernea A, Fernández de la Viña J, Fernández-Martínez JL. Pharmgenomics Pers Med. 2020;13:105–119. doi: 10.2147/PGPM.S205082. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 33.The need for multi-omics biomarker signatures in precision medicine. Olivier M, Asmis R, Hawkins GA, Howard TD, Cox LA. Int J Mol Sci. 2019;20:4781. doi: 10.3390/ijms20194781. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 34.Reconciling evidence-based medicine and precision medicine in the era of big data: challenges and opportunities. Beckmann JS, Lew D. Genome Med. 2016;8:134. doi: 10.1186/s13073-016-0388-7. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 35.Predicting the future - big data, machine learning, and clinical medicine. Obermeyer Z, Emanuel EJ. N Engl J Med. 2016;375:1216–1219. doi: 10.1056/NEJMp1606181. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 36.Digital health is a cultural transformation of traditional healthcare. Meskó B, Drobni Z, Bényei É, Gergely B, Győrffy Z. Mhealth. 2017;3:38. doi: 10.21037/mhealth.2017.08.07. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 37.AI applications in functional genomics. Caudai C, Galizia A, Geraci F, et al. Comput Struct Biotechnol J. 2021;19:5762–5790. doi: 10.1016/j.csbj.2021.10.009. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 38.Predicting the sequence specificities of DNA- and RNA-binding proteins by deep learning. Alipanahi B, Delong A, Weirauch MT, Frey BJ. Nat Biotechnol. 2015;33:831–838. doi: 10.1038/nbt.3300. [DOI] [PubMed] [Google Scholar]
  • 39.Predicting effects of noncoding variants with deep learning-based sequence model. Zhou J, Troyanskaya OG. Nat Methods. 2015;12:931–934. doi: 10.1038/nmeth.3547. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 40.'It will change everything': DeepMind's AI makes gigantic leap in solving protein structures. Callaway E. Nature. 2020;588:203–204. doi: 10.1038/d41586-020-03348-4. [DOI] [PubMed] [Google Scholar]
  • 41.Transcriptomic and proteomic profiling provides insight into mesangial cell function in IgA nephropathy. Liu P, Lassén E, Nair V, et al. J Am Soc Nephrol. 2017;28:2961–2972. doi: 10.1681/ASN.2016101103. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 42.Classification of breast cancer patients using somatic mutation profiles and machine learning approaches. Vural S, Wang X, Guda C. BMC Syst Biol. 2016;10 Suppl 3:62. doi: 10.1186/s12918-016-0306-z. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 43.Network-guided identification of cancer-selective combinatorial therapies in ovarian cancer. He L, Bulanova D, Oikkonen J, et al. Brief Bioinform. 2021;22:272. doi: 10.1093/bib/bbab272. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 44.Machine learning in medicine. Rajkomar A, Dean J, Kohane I. N Engl J Med. 2019;380:1347–1358. doi: 10.1056/NEJMra1814259. [DOI] [PubMed] [Google Scholar]
  • 45.Anticancer drug sensitivity prediction in cell lines from baseline gene expression through recursive feature selection. Dong Z, Zhang N, Li C, Wang H, Fang Y, Wang J, Zheng X. BMC Cancer. 2015;15:489. doi: 10.1186/s12885-015-1492-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 46.Artificial intelligence in drug discovery and development. Paul D, Sanap G, Shenoy S, Kalyane D, Kalia K, Tekade RK. https://doi.org/10.1016/j.drudis.2020.10.010. Drug Discov Today. 2021;26:80–93. doi: 10.1016/j.drudis.2020.10.010. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 47.Artificial intelligence: a powerful paradigm for scientific research. Xu Y, Liu X, Cao X, et al. Innovation (Camb) 2021;2:100179. doi: 10.1016/j.xinn.2021.100179. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 48.Deep learning for drug discovery: a study of identifying high efficacy drug compounds using a cascade transfer learning approach. Zhuang D, Ibrahim AK. Appl Sci. 2021;11:7772. [Google Scholar]
  • 49.eToxPred: a machine learning-based approach to estimate the toxicity of drug candidates. Pu L, Naderi M, Liu T, Wu HC, Mukhopadhyay S, Brylinski M. BMC Pharmacol Toxicol. 2019;20:2. doi: 10.1186/s40360-018-0282-6. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 50.Rees C. Unimagined Futures - ICT Opportunities and Challenges. IFIP Advances in Information and Communication Technology. Cham: Springer; The ethics of artificial intelligence; pp. 55–69. [Google Scholar]
  • 51.Artificial intelligence and collusion. Beneke F, Mackenrodt MO. IIC. 2019;50:109–134. [Google Scholar]
  • 52.Steels L, Brooks R. London: Taylor and Francis; [ May; 2020 ]. 1995. The Artificial Life Route to Artificial Intelligence: Building Embodied, Situated Agents. [Google Scholar]
  • 53.Classification and specific primer design for accurate detection of SARS-CoV-2 using deep learning. Lopez-Rincon A, Tonda A, Mendoza-Maldonado L, et al. Sci Rep. 2021;11:947. doi: 10.1038/s41598-020-80363-5. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 54.Outbreak prediction of COVID-19 for dense and populated countries using machine learning. Khakharia A, Shah V, Jain S, et al. Ann Data Sci. 2020;8:1–19. doi: 10.1007/s40745-020-00314-9. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 55.Artificial intelligence-enabled public health surveillance—from local detection to global epidemic monitoring and control. Zeng D, Cao Z, Neill DB. Artificial Intelligence in Medicine. 2020:437–453. [Google Scholar]
  • 56.Infectious disease surveillance and modelling across geographic frontiers and scientific specialties. Khan K, McNabb SJ, Memish ZA, et al. Lancet Infect Dis. 2012;12:222–230. doi: 10.1016/S1473-3099(11)70313-9. [DOI] [PubMed] [Google Scholar]
  • 57.Mathematical modeling of infectious disease dynamics. Siettos CI, Russo L. Virulence. 2013;4:295–306. doi: 10.4161/viru.24041. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 58.Remote patient monitoring: a systematic review. Farias FA, Dagostini CM, Bicca YA, Falavigna VF, Falavigna A. Telemed J E Health. 2020;26:576–583. doi: 10.1089/tmj.2019.0066. [DOI] [PubMed] [Google Scholar]
  • 59.Strengthening healthcare delivery with remote patient monitoring in the time of COVID-19. Mantena S, Keshavjee S. BMJ Health Care Inform. 2021;28:0. doi: 10.1136/bmjhci-2020-100302. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 60.Effectiveness of mHealth interventions for patients with diabetes: an overview of systematic reviews. Kitsiou S, Paré G, Jaana M, Gerber B. PLoS One. 2017;12:0. doi: 10.1371/journal.pone.0173160. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 61.Remote patient monitoring using artificial intelligence: current state, applications, and challenges. Shaik T, Tao X, Higgins N, et al. WIREs Data Mining Knowledge Discovery. 2023;13:0. [Google Scholar]
  • 62.Prediction of in-hospital mortality in emergency department patients with sepsis: a local big data-driven, machine learning approach. Taylor RA, Pare JR, Venkatesh AK, Mowafi H, Melnick ER, Fleischman W, Hall MK. Acad Emerg Med. 2016;23:269–278. doi: 10.1111/acem.12876. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 63.Comparison of the prognostic value of Chronic Liver Failure Consortium scores and traditional models for predicting mortality in patients with cirrhosis. Antunes AG, Teixeira C, Vaz AM, et al. https://doi.org/10.1016/j.gastrohep.2017.01.001. Gastroenterol Hepatol. 2017;40:276–285. doi: 10.1016/j.gastrohep.2017.01.001. [DOI] [PubMed] [Google Scholar]
  • 64.An edge AI-enabled IoT healthcare monitoring system for smart cities. Rathi VK, Rajput NK, Mishra S, et al. Comput Electrical Eng. 2021;96:107524. [Google Scholar]
  • 65.Artificial intelligence-enhanced electrocardiography in cardiovascular disease management. Siontis KC, Noseworthy PA, Attia ZI, Friedman PA. Nat Rev Cardiol. 2021;18:465–478. doi: 10.1038/s41569-020-00503-2. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 66.Assessing fairness in AI-enabled public health surveillance. Hohma E. IEAI Research Brief. 2021 [Google Scholar]
  • 67.Ethical issues in public health surveillance: a systematic qualitative review. Klingler C, Silva DS, Schuermann C, Reis AA, Saxena A, Strech D. BMC Public Health. 2017;17:295. doi: 10.1186/s12889-017-4200-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 68.The frontiers of fairness in machine learning. Chouldechova A, Roth A. arXiv Preprint. 2018 [Google Scholar]
  • 69.Fair AI: challenges and opportunities. Feuerriegel S, Dolata M, Schwabe G. Bus Inf Syst Eng. 2020;62:379–384. [Google Scholar]
  • 70.Singer Singer, N N. Where do vaccine doses go, and who gets them? The algorithms decide. [ Aug; 2023 ]. 2023. https://www.nytimes.com/2021/02/07/technology/vaccine-algorithms.html https://www.nytimes.com/2021/02/07/technology/vaccine-algorithms.html
  • 71.Mirmomeni M, Fazio T, Cavallar SV, et al. Wearable Sensors. New York: Elsevier; 2021. Chapter 12 - From wearables to THINKables: artificial intelligence-enabled sensors for health monitoring; pp. 339–356. [Google Scholar]
  • 72.CURATE.AI: optimizing personalized medicine with artificial intelligence. Blasiak A, Khong J, Kee T. SLAS Technol. 2020;25:95–105. doi: 10.1177/2472630319890316. [DOI] [PubMed] [Google Scholar]
  • 73.Narrative review of the role of artificial intelligence to improve aortic valve disease management. Thoenes M, Agarwal A, Grundmann D, Ferrero C, McDonald A, Bramlage P, Steeds RP. J Thorac Dis. 2021;13:396–404. doi: 10.21037/jtd-20-1837. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 74.Patient-specific computer simulation for transcatheter cardiac interventions: what a clinician needs to know. de Jaegere P, Rocatello G, Prendergast BD, de Backer O, Van Mieghem NM, Rajani R. Heart. 2019;105:0–7. doi: 10.1136/heartjnl-2018-313514. [DOI] [PubMed] [Google Scholar]
  • 75.AI in telemedicine: an appraisal on deep learning-based approaches to virtual diagnostic solutions (VDS) Oguine O, Oguine K. arXiv Preprint. 2022 [Google Scholar]
  • 76.The era of artificial intelligence-based individualized telemedicine is coming. Jheng YC, Kao CL, Yarmishyn AA, et al. J Chin Med Assoc. 2020;83:981–983. doi: 10.1097/JCMA.0000000000000374. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 77.Global prevalence and major risk factors of diabetic retinopathy. Yau JW, Rogers SL, Kawasaki R, et al. Diabetes Care. 2012;35:556–564. doi: 10.2337/dc11-1909. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 78.World Health Organization. WHO guidelines on managing diabetes in pregnancy. [ Aug; 2023 ]. 2013. https://apps.who.int/iris/bitstream/handle/10665/85975/9789241548441_eng.pdf;jsessionid=2D6B827978BBD2EDC8EE01D448D9864F?sequence=1 https://apps.who.int/iris/bitstream/handle/10665/85975/9789241548441_eng.pdf;jsessionid=2D6B827978BBD2EDC8EE01D448D9864F?sequence=1
  • 79.Applying artificial intelligence to disease staging: deep learning for improved staging of diabetic retinopathy. Takahashi H, Tampo H, Arai Y, Inoue Y, Kawashima H. PLoS One. 2017;12:0. doi: 10.1371/journal.pone.0179790. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 80.Smartphone-based recognition of states and state changes in bipolar disorder patients. Grünerbl A, Muaremi A, Osmani V, et al. IEEE J Biomed Health Inform. 2015;19:140–148. doi: 10.1109/JBHI.2014.2343154. [DOI] [PubMed] [Google Scholar]
  • 81.Detection of fetal electrocardiogram through OFDM, neuro-fuzzy logic and wavelets systems for telemetry. Kumar P, Sharma S, Prasad S. 2016 10th International Conference on Intelligent Systems and Control (ISCO) 2016:1–4. [Google Scholar]
  • 82.A wireless continuous patient monitoring system for dengue: Wi-Mon. Nubenthan S, Shalomy C. 6th National Conference on Technology and Management (NCTM) 2017:23–27. [Google Scholar]
  • 83.Mobile cyber-physical systems for health care: Functions, ambient ontology, and e-diagnostics. Costanzo A, Faro A, Giordano D, et al. 13th IEEE Annual Consumer Communications & Networking Conference (CCNC) 2016:972–975. [Google Scholar]
  • 84.Artificial intelligence in dermatology: a primer. Young AT, Xiong M, Pfau J, Keiser MJ, Wei ML. J Invest Dermatol. 2020;140:1504–1512. doi: 10.1016/j.jid.2020.02.026. [DOI] [PubMed] [Google Scholar]
  • 85.Teledermatology for the diagnosis and management of skin cancer: a systematic review. Finnane A, Dallest K, Janda M, Soyer HP. JAMA Dermatol. 2017;153:319–327. doi: 10.1001/jamadermatol.2016.4361. [DOI] [PubMed] [Google Scholar]
  • 86.Deep neural networks empower medical professionals in diagnosing skin cancer and predicting treatment options for general skin disorders. Han SS, Park I, Chang S, et al. J Invest Dermatol. 2019;139:171. doi: 10.1016/j.jid.2020.01.019. [DOI] [PubMed] [Google Scholar]
  • 87.Clinical recognition of melanoma in dermatologists and nondermatologists. Martinka MJ, Crawford RI, Humphrey S. J Cutan Med Surg. 2016;20:532–535. doi: 10.1177/1203475415623513. [DOI] [PubMed] [Google Scholar]
  • 88.Machine learning for predicting in-hospital mortality after traumatic brain injury in both high-income and low- and middle-income countries. Warman PI, Seas A, Satyadev N, et al. Neurosurgery. 2022;90:605–612. doi: 10.1227/neu.0000000000001898. [DOI] [PubMed] [Google Scholar]
  • 89.Machine learning-driven clinical image analysis to identify craniosynostosis: a pilot study of telemedicine and clinic patients. Paro M, Lambert WA, Leclair NK, et al. Neurosurgery. 2022;90:613–618. doi: 10.1227/neu.0000000000001890. [DOI] [PubMed] [Google Scholar]
  • 90.Telemedicine in neurosurgery and artificial intelligence applications. Paro MR, Lambert W, Leclair NK, Stoltz P, Martin JE, Hersh DS, Bookland MJ. World Neurosurg. 2022;163:83–84. doi: 10.1016/j.wneu.2022.04.078. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 91.[Telemedicine projects in the field of heart failure.: Update and focus of the E-care telemedicine 2.0 project. Perspectives in the field of diabetology] Andrès E, Hajjam M, Talha S, et al. Méd Malad Métabol. 2018;12:224–231. [Google Scholar]
  • 92.Telemonitoring and health counseling for self-management support of patients with type 2 diabetes: a randomized controlled trial. Lindberg I, Torbjørnsen A, Söderberg S, Ribu L. JMIR Diabetes. 2017;2:0. doi: 10.2196/diabetes.6884. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 93.Telemedicine in patients with chronic diseases: the time of maturity! Andresa E, Talha S, Hajjam M, et al. Open Access J Intern Med. 2018;1:1–4. [Google Scholar]
  • 94.AI in mental health. D'Alfonso S. Curr Opin Psychol. 2020;36:112–117. doi: 10.1016/j.copsyc.2020.04.005. [DOI] [PubMed] [Google Scholar]
  • 95.To what scale are conversational agents used by top-funded companies offering digital mental health services for depression. Alattas A, Teepe G, Leidenberger K, et al. Proceedings of the 14th International Joint Conference on BIOSTEC. 2021;5:801–808. [Google Scholar]
  • 96.Detection of dementia from responses to atypical questions asked by embodied conversational agents. Ujiro T, Tanaka H, Adachi H, et al. Proc Interspeech. 2018:1691–1695. [Google Scholar]
  • 97.Mirheidari B. Sheffield, UK: University of Sheffield; 2018. Detecting Early Signs of Dementia in Conversation. [Google Scholar]
  • 98.Development and validation of a virtual agent to screen tobacco and alcohol use disorders. Auriacombe M, Moriceau S, Serre F, et al. Drug Alcohol Depend. 2018;193:1–6. doi: 10.1016/j.drugalcdep.2018.08.025. [DOI] [PubMed] [Google Scholar]
  • 99.Huang J, Li X, Cheng T, et al. Health Information Science. HIS 2015. Lecture Notes in Computer Science. Cham: Springer; 2015. TeenChat: a chatterbot system for sensing and releasing adolescents’ stress; pp. 133–145. [Google Scholar]
  • 100.Response to a relational agent by hospital patients with depressive symptoms. Bickmore TW, Mitchell SE, Jack BW, Paasche-Orlow MK, Pfeifer LM, Odonnell J. Interact Comput. 2010;22:289–298. doi: 10.1016/j.intcom.2009.12.001. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 101.Usability and acceptability assessment of an empathic virtual agent to prevent major depression. Breso A, Martinez-Miranda J, Botella C, et al. Expert Syst Int J Knowl Eng Neural Netw. 2016;33:297–312. [Google Scholar]
  • 102.SimSensei kiosk: a virtual human interviewer for healthcare decision support. DeVault D, Artstein R, Benn GT, et al. https://dl.acm.org/citation.cfm?id=2617415 AAMAS '14. 2014:1061–1068. [Google Scholar]
  • 103.Reporting mental health symptoms: breaking down barriers to care with virtual human interviewers. Lucas GM, Rizzo AJ, Gratch S, et al. Front Robot AI. 2017;4:51. [Google Scholar]
  • 104.What really matters-an information gain analysis of questions and reactions in automated PTSD screenings. Wortwein T, Scherer S. 2017 Seventh International Conference on Affective Computing and Intelligent Interaction (ACII) 2017:15–20. [Google Scholar]
  • 105.A self-diagnosis medical chatbot using artificial intelligence. Divya S, Indumathi V, Ishwarya S, et al. https://api.core.ac.uk/oai/oai:ojs.matjournals.in:article/2334 J Web Dev Web Design. 2018;3:1–7. [Google Scholar]
  • 106.Can chatbots help support a person’s mental health?: perceptions and views from mental healthcare professionals and experts. Sweeney C, Potts C, Ennis E, et al. ACM Transact Comput Healthcare. 2021;2:1–16. [Google Scholar]
  • 107.Ecological momentary interventions for mental health: a scoping review. Balaskas A, Schueller SM, Cox AL, Doherty G. PLoS One. 2021;16:0. doi: 10.1371/journal.pone.0248152. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 108.Technology and mental health: the role of artificial intelligence. Lovejoy CA, Buch V, Maruthappu M. Eur Psychiatry. 2019;55:1–3. doi: 10.1016/j.eurpsy.2018.08.004. [DOI] [PubMed] [Google Scholar]
  • 109.A comparison of artificial intelligence and human doctors for the purpose of triage and diagnosis. Baker A, Perov Y, Middleton K, et al. Front Artif Intell. 2020;3:543405. doi: 10.3389/frai.2020.543405. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 110.An overview of the features of chatbots in mental health: a scoping review. Abd-Alrazaq AA, Alajlani M, Alalwan AA, Bewick BM, Gardner P, Househ M. Int J Med Inform. 2019;132:103978. doi: 10.1016/j.ijmedinf.2019.103978. [DOI] [PubMed] [Google Scholar]
  • 111.Artificially intelligent chatbots in digital mental health interventions: a review. Boucher EM, Harake NR, Ward HE, et al. Expert Rev Med Devices. 2021;18:37–49. doi: 10.1080/17434440.2021.2013200. [DOI] [PubMed] [Google Scholar]
  • 112.Correction: seeing the "big" picture: big data methods for exploring relationships between usage, language, and outcome in internet intervention data. Carpenter J, Crutchley P, Zilca RD, Schwartz HA, Smith LK, Cobb AM, Parks AC. J Med Internet Res. 2017;19:0. doi: 10.2196/jmir.8099. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 113.Using psychological artificial intelligence (Tess) to relieve symptoms of depression and anxiety: randomized controlled trial. Fulmer R, Joerin A, Gentile B, Lakerink L, Rauws M. JMIR Ment Health. 2018;5:0. doi: 10.2196/mental.9782. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 114.Bridging the health data divide. Celi LA, Davidzon G, Johnson AE, et al. J Med Internet Res. 2016;18:0. doi: 10.2196/jmir.6400. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 115.Medical education must move from the information age to the age of artificial intelligence. Wartman SA, Combs CD. Acad Med. 2018;93:1107–1109. doi: 10.1097/ACM.0000000000002044. [DOI] [PubMed] [Google Scholar]
  • 116.Applications of artificial intelligence (AI) in medical education: a scoping review. Nagi F, Salih R, Alzubaidi M, Shah H, Alam T, Shah Z, Househ M. Stud Health Technol Inform. 2023;305:648–651. doi: 10.3233/SHTI230581. [DOI] [PubMed] [Google Scholar]
  • 117.Artificial intelligence in medical education. Masters K. Med Teach. 2019;41:976–980. doi: 10.1080/0142159X.2019.1595557. [DOI] [PubMed] [Google Scholar]
  • 118.Introducing artificial intelligence training in medical education. Paranjape K, Schinkel M, Nannan Panday R, Car J, Nanayakkara P. JMIR Med Educ. 2019;5:0. doi: 10.2196/16048. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 119.Artificial intelligence in healthcare: transforming the practice of medicine. Bajwa J, Munir U, Nori A, Williams B. https://doi.org/10.7861/fhj.2021-0095. Future Healthc J. 2021;8:0–94. doi: 10.7861/fhj.2021-0095. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 120.Cost-utility analysis of virtual and mannequin-based simulation. Haerling KA. Simul Healthc. 2018;13:33–40. doi: 10.1097/SIH.0000000000000280. [DOI] [PubMed] [Google Scholar]
  • 121.The era of artificial intelligence and virtual reality: transforming surgical education in ophthalmology. Bakshi SK, Lin SR, Ting DS, Chiang MF, Chodosh J. Br J Ophthalmol. 2021;105:1325–1328. doi: 10.1136/bjophthalmol-2020-316845. [DOI] [PubMed] [Google Scholar]
  • 122.Cardiopulmonary resuscitation training by avatars: a qualitative study of medical students' experiences using a multiplayer virtual world. Creutzfeldt J, Hedman L, Felländer-Tsai L. JMIR Serious Games. 2016;4:0. doi: 10.2196/games.6448. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 123.AI in medical education: global situation, effects, and challenges [in press] Zhang W, Cai M, Lee HJ, et al. Educ Inf Technol. 2023 [Google Scholar]
  • 124.Ethical and legal challenges of artificial intelligence-driven healthcare. Gerke S, Minssen T, Cohen G. Artificial Intelligence in Healthcare. 2020:295–336. [Google Scholar]
  • 125.Performance of ChatGPT on USMLE: potential for AI-assisted medical education using large language models. Kung TH, Cheatham M, Medenilla A, et al. PLOS Digit Health. 2023;2:0. doi: 10.1371/journal.pdig.0000198. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 126.ChatGPT and the future of medical education. Feng S, Shen Y. Acad Med. 2023;98:867–868. doi: 10.1097/ACM.0000000000005242. [DOI] [PubMed] [Google Scholar]
  • 127.Application of artificial intelligence in medical education: current scenario and future perspectives. Mir MM, Mir GM, Raina NT, et al. http://\ J Adv Med Educ Prof. 2023;11:133–140. doi: 10.30476/JAMP.2023.98655.1803. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 128.Can we open the black box of AI? Castelvecchi D. Nature. 2016;538:20–23. doi: 10.1038/538020a. [DOI] [PubMed] [Google Scholar]
  • 129.Thomas S. Artificial intelligence and medical liability (Part II). Bill of health. [ Aug; 2023 ]. 2017. http://http:/​/blogs.​harvard.edu/​billofhealth/​2017/​02/​10/​artificial-intelligence-and-medical-liability-part-ii/​#more-20718 http://http:/​/blogs.​harvard.edu/​billofhealth/​2017/​02/​10/​artificial-intelligence-and-medical-liability-part-ii/​#more-20718
  • 130.Physicians' perceptions of chatbots in health care: cross-sectional web-based survey. Palanica A, Flaschner P, Thommandram A, Li M, Fossat Y. J Med Internet Res. 2019;21:0. doi: 10.2196/12887. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 131.Chatbots and conversational agents in mental health: a review of the psychiatric landscape. Vaidyam AN, Wisniewski H, Halamka JD, Kashavan MS, Torous JB. Can J Psychiatry. 2019;64:456–464. doi: 10.1177/0706743719828977. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 132.Clinical AI: opacity, accountability, responsibility, and liability. Smith H. AI Soc. 2020;36:535–545. [Google Scholar]
  • 133.Identifying ethical considerations for machine learning healthcare applications. Char DS, Abràmoff MD, Feudtner C. Am J Bioeth. 2020;20:7–17. doi: 10.1080/15265161.2020.1819469. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 134.Privacy and accountability in black-box medicine. Ford RA, Price II WN. https://repository.law.umich.edu./ Mich Telecomm Tech L Rev. 2016;23:1. [Google Scholar]
  • 135.Report on the proposal for a regulation of the European parliament and of the council on the protection of individuals concerning the processing of personal data and on the free movement. [ Aug; 2023 ]. 2012. https://www.europarl.europa.eu/doceo/document/A-7-2013-0402_EN.html https://www.europarl.europa.eu/doceo/document/A-7-2013-0402_EN.html
  • 136.Ownership of medical information. Hall MA, Schulman KA. JAMA. 2009;301:1282–1284. doi: 10.1001/jama.2009.389. [DOI] [PubMed] [Google Scholar]
  • 137.Health information ownership: legal theories and policy implications . Smith LC, Gray E, Thorpe JH. https://scholarship.law.vanderbilt.edu/jetlaw/vol19/iss2/1 Vanderbilt J Entertain Technol Law. 2020;19:207. [Google Scholar]
  • 138.Canadian Association of Radiologists white paper on artificial intelligence in radiology. Tang A, Tam R, Cadrin-Chênevert A, et al. Can Assoc Radiol J. 2018;69:120–135. doi: 10.1016/j.carj.2018.02.002. [DOI] [PubMed] [Google Scholar]
  • 139.Ethics of artificial intelligence in radiology: summary of the Joint European and North American Multisociety statement. Geis JR, Brady AP, Wu CC, et al. Radiology. 2019;293:436–440. doi: 10.1148/radiol.2019191586. [DOI] [PubMed] [Google Scholar]
  • 140.There is no techno-responsibility gap. Tigard DW. Philos Technol. 2020;34:589–607. [Google Scholar]
  • 141.DICOM images have been hacked! Now what? Desjardins B, Mirsky Y, Ortiz MP, Glozman Z, Tarbox L, Horn R, Horii SC. AJR Am J Roentgenol. 2020;214:727–735. doi: 10.2214/AJR.19.21958. [DOI] [PubMed] [Google Scholar]
  • 142.Gender shades: intersectional accuracy disparities in commercial gender classification. Buolamwini J, Gebru T. https://proceedings.mlr.press/v81/buolamwini18a.html Proceedings of the 1st Conference on Fairness, Accountability and Transparency. 2018;81:77–91. [Google Scholar]
  • 143.Andrew Ng: why AI is the new electricity. [ Aug; 2023 ];Ng A. https://www.gsb.stanford.edu/insights/andrew-ng-why-ai-new-electricity RSNA Spotlight Course. 2019 31 [Google Scholar]
  • 144.Methodologic guide for evaluating clinical performance and effect of artificial intelligence technology for medical diagnosis and prediction. Park SH, Han K. Radiology. 2018;286:800–809. doi: 10.1148/radiol.2017171920. [DOI] [PubMed] [Google Scholar]
  • 145.Legal and human rights issues of AI: gaps, challenges, and vulnerabilities. Rodrigues R. J Respons Technol. 2020;4:100005. [Google Scholar]
  • 146.The ethics of AI in health care: a mapping review. Morley J, Machado CC, Burr C, Cowls J, Joshi I, Taddeo M, Floridi L. Soc Sci Med. 2020;260:113172. doi: 10.1016/j.socscimed.2020.113172. [DOI] [PubMed] [Google Scholar]
  • 147.A review of the role of augmented intelligence in breast imaging: from automated breast density assessment to risk stratification. Arieno A, Chan A, Destounis SV. AJR Am J Roentgenol. 2019;212:259–270. doi: 10.2214/AJR.18.20391. [DOI] [PubMed] [Google Scholar]
  • 148.Evaluation of a machine learning algorithm for up to 48-hour advance prediction of sepsis using six vital signs. Barton C, Chettipally U, Zhou Y, et al. Comput Biol Med. 2019;109:79–84. doi: 10.1016/j.compbiomed.2019.04.027. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 149.Clinically applicable deep learning for diagnosis and referral in retinal disease. De Fauw J, Ledsam JR, Romera-Paredes B, et al. Nat Med. 2018;24:1342–1350. doi: 10.1038/s41591-018-0107-6. [DOI] [PubMed] [Google Scholar]
  • 150.Evaluation of a digital consultation and self-care advice tool in primary care: a multi-methods study. Cowie J, Calveley E, Bowers G, Bowers J. Int J Environ Res Public Health. 2018;15:896. doi: 10.3390/ijerph15050896. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 151.Artificial intelligence in hospitals: providing a status quo of ethical considerations in academia to guide future research. Mirbabaie M, Hofeditz L, Frick NR, Stieglitz S. AI Soc. 2022;37:1361–1382. doi: 10.1007/s00146-021-01239-4. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 152.Addressing bias in artificial intelligence in health care. Parikh RB, Teeple S, Navathe AS. JAMA. 2019;322:2377–2378. doi: 10.1001/jama.2019.18058. [DOI] [PubMed] [Google Scholar]
  • 153.Application of artificial intelligence in the health care safety context: opportunities and challenges. Ellahham S, Ellahham N, Simsekler MC. Am J Med Qual. 2020;35:341–348. doi: 10.1177/1062860619878515. [DOI] [PubMed] [Google Scholar]
  • 154.Artificial intelligence in computational pathology - challenges and future directions. Morales S, Engan K, Naranjo V. Digit Signal Process. 2021;119:103196. [Google Scholar]
  • 155.Artificial intelligence in cancer research: trends, challenges and future directions. Sebastian AM, Peter D. Life (Basel) 2022;12:1991. doi: 10.3390/life12121991. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 156.Artificial intelligence to deep learning: machine intelligence approach for drug discovery. Gupta R, Srivastava D, Sahu M, Tiwari S, Ambasta RK, Kumar P. https://link.springer.com/article/10.1007/s11030-021-10217-3. Mol Divers. 2021;25:1315–1360. doi: 10.1007/s11030-021-10217-3. [DOI] [PMC free article] [PubMed] [Google Scholar]
  • 157.The future of artificial intelligence in medicine: medical-legal considerations for health leaders. Jassar S, Adams SJ, Zarzeczny A, Burbridge BE. Healthc Manage Forum. 2022;35:185–189. doi: 10.1177/08404704221082069. [DOI] [PMC free article] [PubMed] [Google Scholar]

Articles from Cureus are provided here courtesy of Cureus Inc.

RESOURCES