Abstract
Background
Nowadays, healthcare systems are coping with the challenge of countering the exponential growth of healthcare costs worldwide, to support sustainability and to guarantee access to treatment for all patients.
Methods
Artificial Intelligence (AI) is the technology able to perform human cognitive functions through the creation of algorithms. The value of AI in healthcare and its ability to address healthcare delivery issues has been a subject of discussion within the scientific community for several years.
Results
The aim of this work is to provide an overview of the primary uses of AI in the healthcare system, to discuss its desirable future uses while shedding light on the major issues related to implications within international regulatory processes. In this manuscript, it will be described the main applications of AI in various aspects of health care, from clinical studies to ethical implications, focusing on the international regulatory framework in countries in which AI is used, to discuss and compare strengthens and weaknesses.
Conclusions
The challenges in regulatory processes to facilitate the integration of AI in healthcare are significant. However, overcoming them is essential to ensure that AI-based technologies are adopted safely and effectively.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
Introduction
Healthcare systems are coping with the challenge of countering the exponential growth of healthcare costs worldwide, which has far outpaced the growth rate of Gross Domestic Product (GDP) in individual states. These constrained finances, combined with an aging population and an increase in chronic diseases, could jeopardize their sustainability [1, 2]. The value of Artificial Intelligence (AI) in healthcare and its ability to address healthcare delivery issues has been a subject of discussion within the scientific community for several years [3,4,5]. Expectations for the use of AI in healthcare are high, as these technologies have demonstrated their potential as valuable support in various medical specialties, including mental health [6], radiology [7], oncology [8] and ophthalmology [9], as well as in improving waitlist reduction [10], treatment adherence [11] and therapy personalization [12]. Key stakeholders are pushing for a concrete integration of AI into clinical practice, transcending the theoretical or experimental boundaries discussed so far [13, 14]. However, although the rapidly growing list of AI-based clinical algorithms approved by the Food and Drug Administration (FDA), their real-life usage remains limited [14]. The challenges encountered on the path to integrating AI into clinical practice go well beyond the initial development and evaluation phase. As a result, it is crucial for the scientific community to have a clear understanding of medical technologies based on AI and how they are regulated, to assess their access in the toolkit available to healthcare professionals. The aim of this manuscript is to provide an overview of the primary uses of AI in the healthcare system and to discuss its future while shedding light on the major issues related to implications and gray areas within international regulatory processes. Figure 1 illustrates the main applications of AI in various aspects of health care.
Definition
AI is the technology able to perform human cognitive functions through the creation of algorithms, such as learning and problem solving [23]. The recent Executive Order 15 U.S.C. 9401 [3] from the White House provided a very concise definition of AI: “a machine-based system that can, for a given set of human-defined objectives, make predictions, recommendations or decisions influencing real or virtual environments.” AI is not a ubiquitous universal technology, but is composed of different subfields that, individually or in combination, add intelligence to applications. The different subgroups are illustrated in Table I.
Clinical Research and Drug Discovery
Effective preclinical preparation is essential for quality clinical studies, involving the identification of promising active ingredients and their targets, as well as the delineation of the experimental strategy for regulatory approval. Errors during this stage can condemn clinical studies to failure. Machine learning can assist researchers in reducing the inefficiencies in the preclinical process. The processes where AI can be used in clinical research are summarized in Fig. 2. The main ones are further detailed below:
Identification of the Pharmacological Target and Mechanism of Action
AI is able to accelerate the process and enhance the identification of active ingredients’ targets and generating new molecules by synthesizing large amounts of current research, elucidating medicine mechanisms of action, and predictively shaping protein structures and potential target interactions [36]. For instance, in the case a drug under investigation acts differently in vivo compared with what expected, AI is able to generate and analyse large amounts of data to improve the understanding of the medicine’s mechanism. Greater knowledge of the mechanism of action and interaction with the target can help increase the likelihood of testing drugs in populaces who take advantages from them [37].
Management of Clinical Trial Participants
Managing participants in clinical experimentation involves selecting target patient populations, recruiting patients, and retaining participants. Participant dropouts and non-adherence to protocol standards may bring studies to overcome allowed timing or costs or abort to generate usable data. In fact, it was assessed that among 33.6% and 52.4% of phase 1–3 clinical studies fail to proceed to the next experimental phase, reducing the probability of a drug tested in phase I gaining approval by 13.8%. ML approaches may ease the more effective and equitable recognition, enrolling, and retention of participants through rapid analysis of extensive databases of prior research [38]. Moreover, machine learning can identify patterns which can be employed to choose phenotypes of patient with a higher likelihood of benefiting from the tested drug. Unstructured data is condemnatory for phenotyping and selecting cohorts more representative, leading that including further patient data is a critical step toward selecting strong and cross-Sect. [39, 40]. Two general overtures are available to increase retention of participants and the adherence of protocol utilizing machine learning-assisted procedures. The first is to employ ML to identify and intercede with people who highly risk to be not compliant with the study [41]. The second overtures involve using machine learning to reduce the study burden on participants, for instance, during data collection, and thus improving their journey. In fact, patients generate critical content for clinical studies even outside the clinical trial context, such as reporting side effects, which ML can process and incorporate into the clinical study [42].
Recruitment
When a specific cohort is identified, there has been promising evidence for natural language processing identifying patients who match the desideratum phenotype, which or else would require a substantial human resource investment [19].
Identification of Endpoints
AI can further be used on data processing for outcome selection. The current approach is performed manually by a committee of clinicians and involves a high-intensity sorting and classification of events, which aligns perfectly with the capabilities of artificial intelligence, resulting in reduced time and costs [43].
Data Analysis
Data gathered in clinical studies, registries or clinical practices are important sources for generating hypotheses and risk assessment where machine learning is particularly suited.
Ethical Implications
The application of AI in healthcare raises a series of ethical and legal issues. In the past, healthcare decisions were made by healthcare professionals themselves. The fact that intelligent technologies autonomously make decisions or assist in various processes raises questions of responsibility, transparency, consent, and confidentiality. The most significant challenge to address is certainly transparency, as most AI algorithms are difficult to understand or interpret. The second aspect is the moral obligation, i.e., the obligation to accept responsibility for one’s actions, which is hard to define in the case of AI usage. Moral obligation could be attributed to both the developer and the operator who decides to use artificial intelligence. Based on the discussion, in order to mitigate risks and maximize the potential of AI in healthcare, the World Health Organization (WHO) has provided principles as a basis for control and governance, summarized in Fig. 3 [44].
Main Challenges
The main challenges for the correct use of AI in healthcare are linked to critical aspects arising from current ML approaches, which heavily rely on training data. Even in unsupervised learning, theoretically independent from direct human monitoring, a connection to supervision processes emerges. This raises significant concerns regarding data quality, security, and privacy, reinforcing the need for in-depth analysis [46]. Regarding public health research and practice, big data poses three key issues: the risk of unintentional disclosure of personally identifiable information (e.g., through the use of online tools), the potential increase in data dimensionality making it challenging to determine ‘deductive disclosure’ of personally identifiable information, and the challenge of identifying and maintaining ethical research standards in the face of emerging technologies that can alter generally accepted privacy norms (e.g., GPS, drones, social media, etc.) [47]. Data quality emerges as a crucial element to ensure the validity of ML models, as the reliability of results is closely connected to the quality of data used in the training process. Another area of concern is the risk associated with potentially biased or incorrect statistical models. It becomes necessary to conduct audits on the ‘black box,’ i.e., understand the internal functioning of models. Furthermore, there is a need to improve system interoperability, aiming to ensure harmonious collaboration and compatibility among them. This approach is crucial to mitigate risks and promote greater transparency in the use of ML/DL models in healthcare contexts [45].
Regulatory Framework in the United States
The regulatory process for AI-based medical devices in the United States is primarily managed by the Food and Drug Administration (FDA), the government agency responsible for regulating medical devices, medicines, and food in the United States. In April 2019, the U.S. Food and Drug Administration (FDA) proposed a regulatory framework for AI in medicine [46]. Here is how the regulatory process for AI-based medical devices in the United States works:
-
Device Classification: The first step involves the classification of AI-based medical devices. The FDA classifies medical devices into three main classes: Class I (low risk), Class II (moderate risk), and Class III (high risk). The class determines the level of required regulation.
-
Premarket Authorization Request: For Class II and Class III devices, the manufacturer must submit a Premarket Approval (PMA) or a 510(k) premarket authorization request to the FDA, depending on the level of risk. The request must include clinical data, evidence of device safety, and effectiveness [47].
-
FDA Evaluation: The FDA carefully evaluates the authorization request, examining the data presented by the manufacturer. This evaluation includes the analysis of clinical data and information on device safety and effectiveness. The FDA may request additional information or conduct inspections at the manufacturer’s facility.
-
Approval or Rejection: After completing the evaluation, the FDA can decide to approve the AI-based medical device, reject the approval, or request additional data or modifications. If approval is granted, the device can be marketed in the United States.
-
Post-Market Monitoring: After marketing, the FDA continues to monitor the safety and effectiveness of AI-based medical devices through adverse event reports, inspections, and periodic reviews.
It is noteworthy that the FDA is constantly working to develop specific guidelines and regulations for AI-based medical devices to address the unique challenges posed by these technologies. The agency has developed the “Artificial Intelligence/Machine Learning (AI/ML)—Based Software as a Medical Device (SaMD) Action Plan”, a new action plan which outlines activities for regulating AI-based software as Medical Devices (SaMD) [48]. The action plan emphasizes the need for specific regulatory attention given the rapid and constant evolution of these tools. The document illustrates the agency’s effort to reimagine a premarket review approach based on the total product lifecycle (TPLC). The FDA regulates products to ensure they are safe for consumers. In the healthcare sector, machine learning tools can pose safety concerns, as neglecting potential negative impacts can increase the risk and danger for already marginalized and discriminated groups in healthcare. Considerations of health disparities should not be out of scope or an optional dimension to consider when developing machine learning tools for medicine. To that end, considerations of health disparities can be integrated into the FDA’s AI/ML regulation both in the pre-market and post-market phases. In addition to conventional approaches, the FDA has outlined a specific pathway for SaMD technologies in its Digital Health Innovation Action Plan, offering a streamlined path to precertification, certifying the developer themselves rather than the product [49]. The TPLC facilitates continuous monitoring of the SaMD product throughout its development and post-market performance, as periodic disclosure of performance data to the FDA has been mandated. As a result, the agency aims to strike a balance between reducing the overall regulatory burden while ensuring efficacy and safety.
Regulatory Framework in the European Union
In European Union (EU) AI-based medical devices are not approved by a centralized agency. The regulatory process is as follows [50, 51]:
-
Low-risk devices (Class I): For devices considered low-risk (Class I), the manufacturer is responsible for ensuring that the product complies with regulations. Pre-market approval by a central regulatory authority is not required.
-
High-risk devices (Classes IIa, IIb, and III): Devices considered high-risk are overseen by private Notified Bodies, which are accredited and authorized independent organizations responsible for evaluating and approving such devices. These bodies verify that the devices meet European regulatory requirements before they can be placed on the market.
This European system relies on third-party assessment by Notified Bodies to ensure compliance with regulations before devices can be placed on the market. Therefore, EU delegates a significant portion of approval responsibility to accredited private entities, while the FDA in the United States plays a central role in the approval of medical devices. The differences between the European and American regulatory processes are illustrated in Table II.
The European regulatory system for AI-based medical devices has some critical issues and challenges, including:
-
Variation in Notified Body procedures: since Notified Bodies are private and operate in different European countries, procedures and standards may vary between countries. This can lead to a lack of uniformity in the assessment and approval of medical devices in Europe.
-
Need for adequate training and resources: Notified Bodies must have the necessary resources and expertise to conduct accurate assessments and approve medical devices. The quality of assessments may vary depending on the involved body.
-
Risks to patient safety: proper evaluation of medical devices is crucial to ensure patient safety. Any errors or deficiencies in the assessment procedures could pose risks to patients.
-
Delayed regulation: the European approval process involving Notified Bodies may be slower than that in the United States, potentially resulting in delays in European patients’ access to new medical technologies.
-
Coordination and standardization: variation in processes and standards among different European countries can create complexity and challenges in marketing medical devices throughout the European Union.
-
Post-market monitoring: after marketing, adequate monitoring of medical devices is essential to ensure long-term safety and effectiveness.
Overall, these challenges are related to the balance between the need to ensure stringent regulation for the safety and effectiveness of medical devices and the need to provide timely access to new medical technologies for patients. EU is addressing these challenges by improving coordination among Notified Bodies and establishing clearer guidelines for the evaluation and approval of AI-based medical devices [52].
Discussion and Conclusions
As illustrated in Table II, the regulatory processes for AI-based medical devices in the USA and the EU exhibit both similarities and differences, reflecting distinct approaches to ensure the safety and effectiveness of healthcare technologies. In the USA, the FDA governs the regulatory landscape for AI-based medical devices. The process involves classifying devices based on risk, with subsequent premarket authorization requests for Class II and III devices. The FDA evaluates clinical data and safety information, either approving or rejecting the device. Post-market monitoring is crucial, ensuring continued safety and effectiveness. The FDA is actively adapting to the challenges posed by AI, emphasizing a lifecycle approach, continuous monitoring, and efforts to streamline regulatory pathways. In contrast, the EU has a decentralized approach. Low-risk devices undergo self-assessment by manufacturers, while high-risk devices rely on private Notified Bodies for assessment and approval. Challenges arise from potential variation in procedures and standards among different countries, necessitating coordinated efforts. Potential delays in approval and risks to patient safety are concerns, prompting the need for robust post-market monitoring and improved coordination among Notified Bodies. The key distinction lies in the central role of the FDA in the USA, acting as the primary regulatory authority, while the EU delegates approval responsibilities to accredited private entities. This difference may contribute to variations in regulatory speed and potential challenges in uniformity across Europe. Both regulatory systems cope with the balance between stringent safety measures and the timely introduction of innovative medical technologies. The FDA’s emphasis on continuous monitoring aligns with the EU’s acknowledgment of post-market surveillance’s importance. Addressing these challenges requires ongoing efforts in both regions. The FDA is developing specific guidelines for AI-based devices to ensure rapid adaptation to evolving technologies. In the EU, there is a need for improved coordination among notified bodies and clearer guidelines. Furthermore, overall the integration of artificial intelligence into clinical practice presents many challenges. Acceptance and trust among healthcare professionals are pivotal for the successful integration of AI in healthcare. In addition, ensuring interoperability is a critical aspect. AI systems must seamlessly integrate with existing electronic health record systems, ensuring a smooth flow of data and facilitating easy access for professionals to the insights generated by AI. In the ever-evolving landscape of data privacy and security, AI algorithms must adhere to stringent regulations and robust information security practices. It is also becoming increasingly imperative to establish reliable evaluation criteria, to ensure that AI solutions meet clinical effectiveness and safety standards before being deployed for patient care. Furthermore, addressing bias in both data and decisions is a crucial ethical consideration for AI systems. Moreover, the implementation of AI in healthcare environments demands substantial investments in infrastructure and training, posing challenges for facilities with limited resources. The ethical complexities surrounding AI in medicine extend to the evolving landscape of legal responsibility. Clarity in determining legal responsibility for AI errors or incorrect decisions is still under development, and defining regulations and standards in this domain is crucial. To achieve all these aims is crucial to balance collaboration between regulatory bodies, healthcare professionals, and technology developers to create a robust and adaptive regulatory framework. Addressing these challenges requires a combination of technical expertise, resources, international collaboration, and a smart and balanced regulatory approach [53].
Data Availability
Full availability of data and materials. All stated data can be provided on request to the reader.
Code Availability
Not applicable.
References
Snowdon A. Digital health: a framework for healthcare transformation. 2020. Available online: https://www.gs1ca.org/documents/digital_health-affht.pdf. Accessed 23 Jan 2023.
Williams OD. COVID-19 and private health: market and governance failure. Development (Rome). 2020;63(2–4):181–90. https://doi.org/10.1057/s41301-020-00273-x.
Desai AN. Artificial intelligence: promise, pitfalls, and perspective. JAMA. 2020;323(24):2448–9. https://doi.org/10.1001/jama.2020.8737.
Reddy S, Fox J, Purohit MP. Artificial intelligence-enabled healthcare delivery. J R Soc Med. 2019;112(1):22–8. https://doi.org/10.1177/0141076818815510.
Gampala S, Vankeshwaram V, Gadula SSP. Is artificial intelligence the new friend for radiologists? a review article. Cureus. 2020;12(10):e11137. https://doi.org/10.7759/cureus.11137.
Renn BN, Schurr M, Zaslavsky O, Pratap A. Artificial intelligence: an interprofessional perspective on implications for geriatric mental health research and care. Front Psychiatry. 2021;12:734909. https://doi.org/10.3389/fpsyt.2021.734909.
Oduoye MO, Fatima E, Muzammil MA, Dave T, Irfan H, Fariha FNU, Marbell A, Ubechu SC, Scott GY, Elebesunu EE. Impacts of the advancement in artificial intelligence on laboratory medicine in low- and middle-income countries: challenges and recommendations-A literature review. Health Sci Rep. 2024;7(1): e1794. https://doi.org/10.1002/hsr2.1794.
Yang J, Huang J, Han D, Ma X. Artificial intelligence applications in the treatment of colorectal cancer: a narrative review. Clin Med Insights Oncol. 2024;5(18):11795549231220320. https://doi.org/10.1177/11795549231220320.
Cai Y, Zhang X, Cao J, Grzybowski A, Ye J, Lou L. Application of artificial intelligence in oculoplastics: a review. Clin Dermatol. 2024:S0738-081X(23)00271-7. https://doi.org/10.1016/j.clindermatol.2023.12.019.
Rana HS, Umer M, Hassan U, Asgher U. A novel multi-criteria decision-making approach for prioritization of elective surgeries through formulation of “weighted MeNTS scoring system.” Heliyon. 2022;8(8). https://doi.org/10.1016/j.heliyon.2022.e10339.
Gebremichael LG, Champion S, Nesbitt K, Pearson V, Bulamu NB, Dafny HA, Sajeev S, Pinero de Plaza MA, Ramos JS, Suebkinorn O, Gulyani A, Bulto LN, Beleigoli A, Hendriks JM, Hines S, Clark RA; NHMRC CHAP Project Team. Effectiveness of cardiac rehabilitation programs on medication adherence in patients with cardiovascular disease: a systematic review and meta-analysis. Int J Cardiol Cardiovasc Risk Prev. 2023;20:200229. https://doi.org/10.1016/j.ijcrp.2023.200229.
Marchiò C, Criscitiello C, Scatena C, Santinelli A, Graziano P, Malapelle U, Cursano G, Venetis K, Fanelli GN, Pepe F, Berrino E, Angelis C, Perrone G, Curigliano G, Fusco N. Think, “HER2” different: integrative diagnostic approaches for HER2-low breast cancer. Pathologica. 2023;115(6):292–301. https://doi.org/10.32074/1591-951X-942.
Feng J, Phillips RV, Malenica I, Bishara A, Hubbard AE, Celi LA, et al. Clinical artificial intelligence quality improvement: towards continual monitoring and updating of AI algorithms in healthcare. NPJ Digit Med. 2022;5(1):66. https://doi.org/10.1038/s41746-022-00611-y.
Park Y, Jackson GP, Foreman MA, Gruen D, Hu J, Das AK. Evaluating artificial intelligence in medicine: phases of clinical research. JAMIA Open. 2020;3(3):326–31. https://doi.org/10.1093/jamiaopen/ooaa033.
Kim HE, Kim HH, Han BK, Kim KH, Han K, Nam H, et al. Changes in cancer detection and false-positive recall in mammography using artificial intelligence: a retrospective, multireader study. Lancet Digit Health. 2020;2(3):e138–48. https://doi.org/10.1016/S2589-7500(20)30003-0.
Koçak B, Cuocolo R, dos Santos DP, Stanzione A, Ugga L. Must-have qualities of clinical research on artificial intelligence and machine learning. Balkan Med J. 2023;40(1):3–12. https://doi.org/10.4274/balkanmedj.galenos.2022.2022-11-51.
Wu H, Chan NK, Zhang CJP, Ming WK. The role of the sharing economy and artificial intelligence in health care: opportunities and challenges. J Med Internet Res. 2019;21(10). https://www.jmir.org/2019/10/e13469.
Ayers JW, Poliak A, Dredze M, Leas EC, Zhu Z, Kelley JB, et al. Comparing physician and artificial intelligence chatbot responses to patient questions posted to a public social media forum. JAMA Intern Med. 2023;183(6):589–96. https://doi.org/10.1001/jamainternmed.2023.1838.
Zhang X, Xiao C, Glass LM, Sun J. DeepEnroll: patient-trial matching with deep embedding and entailment prediction. In: The Web Conference 2020 - Proceedings of the World Wide Web Conference, WWW 2020; 2020.
Dikshit A, Pradhan B. Interpretable and explainable AI (XAI) model for spatial drought prediction. Sci Total Environ. 2021;801:149797. https://doi.org/10.1016/j.scitotenv.2021.149797.
Walsh CA, Cahir C, Tecklenborg S, Byrne C, Culbertson MA, Bennett KE. The association between medication non-adherence and adverse health outcomes in ageing populations: a systematic review and meta-analysis. Br J Clin Pharmacol. 2019;85(11):2464–78. https://doi.org/10.1111/bcp.14075.
Ramgopal S, Sanchez-Pinto LN, Horvat CM, Carroll MS, Luo Y, Florin TA. Artificial intelligence-based clinical decision support in pediatrics. Pediatr Res. 2023;93(2):334–41. https://doi.org/10.1038/s41390-022-02226-1.
Bernstein P. What is Artificial Intelligence (AI)? In: Machine Learning. 2022. https://www.taylorfrancis.com/chapters/mono/10.4324/9781003297192-3/artificial-intelligence-ai-phil-bernstein.
Rajkomar A, Dean J, Kohane I. Machine learning in medicine. N Engl J Med. 2019;380(14):1347–58. https://doi.org/10.1056/NEJMra1814259.
Lee SI, Celik S, Logsdon BA, Lundberg SM, Martins TJ, Oehler VG, et al. A machine learning approach to integrate big data for precision medicine in acute myeloid leukemia. Nat Commun. 2018;9(1):42. https://doi.org/10.1038/s41467-017-02465-5.
World Bank Group. Distributed Ledger Technology (DLT) and Blockchain Acknowledgments III. World Bank. 2017;(1). https://documents.worldbank.org/en/publication/documents-reports/documentdetail/177911513714062215/distributed-ledger-technology-dlt-and-blockchain.
Alruwaili FF. Artificial intelligence and multi agent based distributed ledger system for better privacy and security of electronic healthcare records. PeerJ Comput Sci. 2020;6:e323. https://doi.org/10.7717/peerj-cs.323.
IMDRF (International Medical Device Regulators Forum, IMDRF/AIMD WG/N67, 9 May 2022. https://ipi.academy/product/industry/25/medical-devices?utm_source=bing&utm_medium=cpc&utm_campaign=wrd-mads-medical-device&msclkid=4822df0402af10fce086bae707daac28.
Le Glaz A, Haralambous Y, Kim-Dufor DH, Lenca P, Billot R, Ryan TC, et al. Machine learning and natural language processing in mental health: systematic review. J Med Internet Res. 2021;23(5):e15708. https://doi.org/10.2196/15708.
Wren JD. The emerging in-silico scientist: how text-based bioinformatics is bridging biology and artificial intelligence. IEEE Eng Med Biol Mag. 2004;23(2):87–93. https://doi.org/10.1109/memb.2004.1310989.
Altman RB. Artificial intelligence (AI) systems for interpreting complex medical datasets. Clin Pharmacol Ther. 2017;101(5):585–6. https://doi.org/10.1002/cpt.650.
Trunfio M, Rossi S. Advances in metaverse investigation: streams of research and future agenda. Virtual Worlds. 2022;1(2). https://www.mdpi.com/2813-2084/1/2/7.
Yeung AWK, Tosevska A, Klager E, Eibensteiner F, Laxar D, Stoyanov J, et al. Virtual and augmented reality applications in medicine: analysis of the scientific literature. J Med Internet Res. 2021;23(2):e25499. https://doi.org/10.2196/25499.
Thorp HH. ChatGPT is fun, but not an author. Science. 2023;379. https://www.science.org/doi/10.1126/science.adg7879.
Sallam M. ChatGPT utility in healthcare education, research, and practice: systematic review on the promising perspectives and valid concerns. Healthcare (Basel). 2023;11(6):887. https://doi.org/10.3390/healthcare11060887.
Senior AW, Evans R, Jumper J, Kirkpatrick J, Sifre L, Green T, et al. Improved protein structure prediction using potentials from deep learning. Nature. 2020;577(7792):706–10. https://doi.org/10.1038/s41586-019-1923-7.
Madhukar NS, Khade PK, Huang L, Gayvert K, Galletti G, Stogniew M, et al. A Bayesian machine learning approach for drug target identification using diverse data types. Nat Commun. 2019;10(1). https://api.semanticscholar.org/CorpusID:208145918.
Wong CH, Siah KW, Lo AW. Estimation of clinical trial success rates and related parameters. Biostatistics. 2019;20(2):273–86. https://doi.org/10.1093/biostatistics/kxx069.
Glicksberg BS, Miotto R, Johnson KW, Shameer K, Li L, Chen R, et al. Automated disease cohort selection using word embeddings from Electronic Health Records. Pac Symp Biocomput. 2018;23:145–56. https://www.ncbi.nlm.nih.gov/pmc/articles/PMC5788312/pdf/nihms921845.pdf.
Liao KP, Cai T, Savova GK, Murphy SN, Karlson EW, Ananthakrishnan AN, et al. Development of phenotype algorithms using electronic medical records and incorporating natural language processing. BMJ. 2015;350:h1885. https://doi.org/10.1136/bmj.h1885.
Bain EE, Shafner L, Walling DP, Othman AA, Chuang-Stein C, Hinkle J, et al. Use of a novel artificial intelligence platform on mobile devices to assess dosing compliance in a phase 2 clinical trial in subjects with schizophrenia. JMIR Mhealth Uhealth. 2017;5(2). https://mhealth.jmir.org/2017/2/e18.
Gavrielov-Yusim N, Kürzinger ML, Nishikawa C, Pan C, Pouget J, Epstein LB, et al. Comparison of text processing methods in social media-based signal detection. Pharmacoepidemiol Drug Saf. 2019;28(10):1309–17. https://doi.org/10.1002/pds.4857.
Cascini F, Beccia F, Causio FA, Melnyk A, Zaino A, Ricciardi W. Scoping review of the current landscape of AI-based applications in clinical trials. Front Public Health. 2022;10. https://publires.unicatt.it/it/publications/scoping-review-of-the-current-landscape-of-ai-based-applications.
World Health Organisation. WHO Issues First Global Report on Artificial Intelligence (AI) in Health and Six Guiding Principles for Its Design and Use. 2021. Available online: https://www.who.int/news/item/28-06-2021-who-issues-first-global-report-on-ai-in-health-and-six-guiding-principles-for-its-design-and-use. Accessed 15 Dec 2021.
Beard RW. Decision making under uncertainty: theory and application [Bookshelf]. IEEE Control Syst. 2018;38(6). https://ieeexplore.ieee.org/document/8536567.
U.S. Food and Drug Administration. Proposed Regulatory Framework for Modifications to Artificial Intelligence/Machine Learning (AI/ML)-Based Software as a Medical Device (SaMD). https://www.fda.gov/medical-devices/software-medical-device-samd/artificial-intelligence-and-machine-learning-software-medical-device. Accessed 22 Sept 2021.
Overview of medical device classification and reclassification. Available at: https://www.fda.gov/about-fda/cdrh-transparency/overview-medical-device-classification-and-reclassification. Acccessed 25 July 2019.
U.S. Food and Drug Administration. Software as a Medical Device (SaMD). http://www.fda.gov/medical-devices/digital-health/software-medical-device-samd. Accessed 2 July 2019.
Wang T. Outpatient clinic process enhancement at children’s hospital: empoweredby information technology. Chin J Hosp Adm. 2018;34:753–5. https://api.semanticscholar.org/CorpusID:224250149.
Vokinger KN, Mühlematter UJ, Becker A, Boss A, Reutter MA, Szucs TD. Artificial intelligence und machine learning in der medizin. Aug 28, 2017. https://jusletter.weblaw.ch/juslissues/2017/903/artificial-intellige_da49225588.html__ONCE&login=false. Accessed 26 Dec 2020.
Swissmedic. Guide to the regulation of medical devices. Nov 6, 2017. https://www.swissmedic.ch/swissmedic/en/home/medical-devices/regulation-of-medical-devices/medical-device-regulation_onlineguide.html. Accessed 1 Oct 2020.
Pesapane F, Volonté C, Codari M, Sardanelli F. Artificial intelligence as a medical device in radiology: ethical and regulatory issues in Europe and the United States. Insights Imaging. 2018;9(5):745–53. https://doi.org/10.1007/s13244-018-0645-y.
Romagnoli A, Zovi A, Sabbatucci M, Ferrara F, Vitiello A. Therapeutic innovation and digital therapies in the world: Comparison between the American and European regulatory framework, with focus on Italy. J Interprof Educ Pract. 2023;32. https://www.sciencedirect.com/science/article/abs/pii/S2405452623000587.
Funding
No funding was received to conduct this study.
Author information
Authors and Affiliations
Contributions
AR: Conceptualization, Writing—original draft, Methodology
FF: Supervision, Validation
RL: Writing—review & editing, Supervision, Validation
AZ: Writing—review & editing, Supervision, Validation
Corresponding author
Ethics declarations
Ethical Approval
Not applicable.
Consent to Participate
Not applicable.
Consent to Publish
The authors consent to the publication of the manuscript.
Competing Interest
The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.
Other
The authors declare that the opinions expressed are of a personal nature and do not in any way commit the responsibility of the Administrations to which they belong.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Romagnoli, A., Ferrara, F., Langella, R. et al. Healthcare Systems and Artificial Intelligence: Focus on Challenges and the International Regulatory Framework. Pharm Res 41, 721–730 (2024). https://doi.org/10.1007/s11095-024-03685-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11095-024-03685-3