Abstract
Artificial intelligence (AI) is a growing phenomenon that will facilitate wide-scale changes in many professions such as healthcare and medical education. Artificial intelligence and machine learning aid medical professionals in managing patient data and records, developing new medical procedures, and diagnosing and treating chronic diseases. This literature review examines the history of artificial intelligence in healthcare settings and medical schools. The review opens by defining AI and providing an overview of the various classifications. Next, the literature summarizes developments of AI in healthcare from the 1950s to the present date. Finally, the literature review ends with a discussion of the ethical implications and limitations of AI in healthcare.
Access provided by Autonomous University of Puebla. Download chapter PDF
Similar content being viewed by others
Keywords
Introduction
Although a precise, universally accepted definition of artificial intelligence (AI) does not exist yet, AI is typically associated with a branch of computer science focused on building algorithm-based machines capable of performing tasks that generally require human decision-making and intelligence (Stone et al., 2016). AI can be categorized based on its capabilities and functionalities. Based on capabilities, AI can be classified as Narrow AI, General AI, or Super AI. Narrow AI, sometimes also called weak AI, focuses on a single or predefined subset of cognitive abilities, and cannot perform outside of these limitations (Kaplan & Haenlein, 2019). Narrow AI enables virtual assistants such as Siri and Alexa to understand human speech and respond when queries are within their breadth of abilities. Narrow AI is also the basis of services such as Google Translate, Facebook’s facial recognition abilities, and Tesla’s self-driving capabilities. General AI, also known as Strong AI, is the second generation of AI in development that allows machines to apply knowledge and skills within contexts within which they were not designed to function. Such systems can reason, plan, and solve problems independently for tasks that extend beyond their planned uses. Finally, Super AI refers to machine-based systems that are truly self-aware and capable of social skills, scientific creativity, and general wisdom (Kaplan & Haenlein, 2019). Through singularity, these systems can surpass human intelligence and have the ability to perform any task.
Several subfields exist under the umbrella of AI. Machine learning involves giving computers the ability to learn without being programmed and adapt based on previous experiences (Jordan & Mitchell, 2015). Neural networks are a branch of AI that attempts to replicate the human brain by using algorithms to discern elemental relationships across copious amounts of data. Robotics focuses on designing and constructing robots that can be deployed for tasks that may be laborious for humans to perform steadily. Finally, natural language processing facilitates the communication between computers and humans by natural language. This subfield enables computers to read and understand data by mimicking human language (Nadkarni et al., 2011).
This paper focuses specifically on the uses of AI in the field of healthcare. Applications of artificial intelligence in healthcare can be categorized as virtual or physical (Hamet & Tremblay, 2017). The virtual branch includes mathematically based algorithms that enable the systems to develop and learn through experience and use. In contrast, the physical branch of AI includes medical devices and sophisticated robots such as the da Vinci Surgical System capable of performing surgery or carebots that serve as companions for elderly individuals with limited mobility or with cognitive decline.
Methods
For this literature review, refereed online journals and other academic resources were searched for publications that described artificial intelligence applications in healthcare settings. This paper used the method of structured review to arrive at a comprehensive and reliable overview of research on artificial intelligence in the healthcare field. The selected articles were found by searching multiple academic databases, including EBSCO, ERIC, and Google Scholar, using one of the following keywords: Artificial Intelligence, Machine Learning, Neural Learning, in conjunction with the keyword, Healthcare or Patient Care. The criteria for articles included in the literature was being published in a peer-reviewed journal between 1950 and 2021, written in English, having online full-text accessibility, and searched keywords appearing in the title. This literature review used document analysis to screen articles and collect initial data. Content analysis was then used to identify research patterns and trends.
Discussion
Timeline of AI in Healthcare
1950s–1960s
Although storing digital information had its start in the early 1940s with large computer systems such as Vannevar Bush’s Memex, the drive of multimedia in the 1950s would also lead to significant developments in the creation and storage of digital information. During this time, the focus was on creating machines that could make human decisions as accurate as a human could. This interest included multiple areas of science, including medicine. It was during the 1950s that early experimentations of AI were being fabricated. In 1950 Alan Turing created a machine to test whether computers could be designed to think and referred to his creation as the imitation game (Turing, 1950). The goal of the experiment was to test humans to see if they could distinguish between computers and humans based on conversation text, and eventually became known as the Turing Test (Amisha et al., 2019). Since the mere mention of the term AI by John McCarthy in 1956, the world has looked into artificial intelligence and machine learningto not only solve human problems but also predict outcomes before problems occur.
The mid-1950s and 1960s were known for major advancements during the information explosion age due to the innovations in the computer and mechanical industries (Hiagh, 2009). Creating accompanying software to talk to the machines was led by pioneers like Grace Hopper, who is widely accredited for codeveloping the standardized computer language COBOL (Abbate, 2011). The developments of programming languages were considered just as important as the hardware developments; one simply did not advance without the other. It was because of the synchronicity of the advancements that the auto industry could capitalize on AI. In 1961, General Motors built a system named Unimate, which was able to automate die casting steps by following a series of instructions (Kaul Enslin, & Gross, 2020). Soon after, in 1964, Dr. Joseph Weizenbaum used natural language processing (NLP) to create what was considered the first chatterbot named ELIZA, which was designed to simulate conversions between humans in the field of psychology and communication (Kaul et al., 2020). Two years later, in 1966, what was referred to as the “first electronic person” was created by a team at the Stanford Research Institute (now known as SRI International). The robot was nick-named Shakey and was primarily programmed to solve problems on its own by mapping out its surroundings and completing mobile-based tasks. The hope was that Shakey would lead to robots carrying out and improving complicated tasks in factories (Szondy, 2015).
1970s–1980s
In 1970, William B. Schwartz published an article titled “Medicine and the Computer: The Promise and Problems of Change” (Swartz, 1970). Schwartz was a medical doctor with great interest in using computers in healthcare. He documented his views on future healthcare system developments and the benefits from such advances in technology. His team developed their own computer program that would consult with a physician by attempting to diagnose a patient based on details provided by the user in what he referred to as “branching of the decision tree” (Schwartz, 1970, p. 1259). Schwartz also issued warnings over some problems that are of concern in our current healthcare system, including high costs, social economics, legal demands, and technical considerations.
As with most technological advancements in history, there were years in which progress lacked technological developments. In the world of artificial intelligence and machine learning, these are known as “Winters.” The First AI Winter that occurred in the mid-1970s continued through most of the 1980s. Exceptions to this Winter included William Clancey and Edward Shortliffe, who brought to light the uses of AI applications in multiple healthcare settings. Their 1984 publication (Clancey & Shortliffe, 1984) defined what medical AI meant and considered the potentials of AI applications with the needs of healthcare professionals to improve overall documentation and communication in medicine.
1990s–2000s
The Second AI Winter occurred in the late 1980s and early 1990s where the main goals were to improve medical data online accessibility. But by the late 1990s there was regained interest in part due to the American National Library of Medicine awarding contracts to medical institutions to explore new ways to incorporate AI into healthcare, improving overall infrastructures (Kaul et al., 2020). AI took shape in the form of electronic medical records (EMRs), telemedicine, and other areas of information sharing.
2010–present
Mintz and Brodie (2019) define several focuses of AI in healthcare: image processing, computer vision, artificial neural network (ANN), machine learning(ML), convolutional neural network (CNN), and deep learning (DL). Deep learningis a subset of machine learning that is constructed to process the information on multiple levels, like the human brain (Mintz & Brodie, 2019). It was first investigated in the 1950s but had limitations mainly due to the absence of computer technology. It took many years, but technology finally advanced enough that DL could be re-investigated and built on. Cardiology, gastroenterology, oncology, radiology, and surgery are a few areas in healthcare exploring the uses of AI, ML, and DL (Kaul et al., 2020).
Jiang et al. (2017) performed a literature review on the topic of AI in healthcare. Their focus was to identify specific specialties in medicine utilizing artificial intelligence to improve clinical practices. They reviewed articles published during the time period 2013–2016 covering medical specialties, healthcare data, and AI categories (ML and NLP). Once the review was completed, the team found diagnostic imaging data to be the main interest in publication content, followed by genetics and electrodiagnosis. The top four diseases that resulted in publications during 2013–2016 included neoplasms, nervous, cardiovascular, and urogenital. A similar search of publications during late 2019–2020 yielded a dramatic shift of focus to the SARS-CoV-2, the novel coronavirus that causes COVID-19 (Bansal et al., 2020).
Healthcare Education and AI
Developments in artificial intelligence have also impacted medical schools and how future medical professionals are trained. The following section reviews common uses of AI in medical schools, and how students and instructors respond to the use of this technology.
Current Applications of AI in Medical Schools
Authors such as Wartman and Combs (2019) have suggested that the system for educating medical professionals requires a revolution because the amount of available medical knowledge exceeds the human mind’s organizing capacity and can lead to stress-induced mental illness among learners. Wartman and Combs (2019) suggest remediating this problem by re-engineering the medical school curricula so that it shifts from a “focus on information acquisition to an emphasis on knowledge management and communication” (p. 147). AI applications can aggregate vast amounts of data, generate diagnostic and treatment options, and assign confidence ratings to those recommendations that clinicians can then interpret and communicate to patients. Training in understanding AI systems’ recommendations can enable medical students to personalize treatment to the individual characteristics of each patient. Utilizing AI systems also reduces cognitive and information overload that medical students can experience earlier in their careers.
Some researchers such as Masters (2019) believe that AI will eventually affect every aspect of human life, including medicine and medical education. Masters (2019) asserts that surgical robotics, for instance, will continue to evolve until intelligent robots with AI software can perform surgery without humans. He cautions that medical schools that have not integrated AI into their curriculum and are not teaching robotic surgery will quickly fall behind this potential standard. Earlier researchers such as Moles et al. (2009) developed a program to teach and assess the development of otolaryngologic residents’ basic robotic surgical skills. Moles et al. (2009) suggest AI and robotic surgery training programs be formally established in residency programs because the skill will eventually become an integral part of surgical practice. Alonso-Silverio et al. (2018) conducted a similar study focusing on the laparoscopic surgical skills of ten medical students and six residents. They found that laparoscopic box training systems based on open-source hardware and artificial intelligence improved the learning curve and dexterity of the 16 participants.
Medical schools are increasing their use of artificial intelligence in simulation-based training designed to assess and train the psychomotor skills involved in the surgery. Mirchi et al. (2020) piloted their Virtual Operative Assistant, an educational platform that automatically assesses and provides user feedback, with 28 skilled participants and 22 novice participants who were required to perform a virtual reality-based subpial brain tumor resection task using the NeuroVR simulator. The Virtual Operative Assistant successfully classified novice and skilled participants with an accuracy of 92%, specificity of 82%, and sensitivity of 100%. The researchers concluded that the AI-based system was advantageous over traditional teaching methods because it enabled instructors to identify individual components of psychomotor expertise in tasks too multifaceted for instructors to observe normally. The system mimics real life by using anapprenticeship model with auditory feedback in natural human language in a clinical or operating environment.
Medical Student Attitudes Toward AI
Sit et al. (2020) surveyed medical students (n = 484) at 19 UK medical schools and found most medical school students surveyed (88%) believed that AI would play an essential role in healthcare. However, the students surveyed did not feel adequately prepared to work alongside AI and would find training on the subject beneficial. Some participants (n = 45) experienced AI training in their medical programs, but this group still reported a lack of confidence and understanding required for the critical use of healthcare AI tools.
Discussion of AI in healthcare in the news and lay media often express the concern that AI is replacing radiologists. Park et al. (2020) administered a web-based survey to American medical students (n = 152) to understand their perceptions about radiology and determine if such reports negatively impact medical students’ perceptions of radiology as a viable specialty. After analyzing results from the six-question survey, the authors concluded that more than 75% of participants believed AI would significantly impact medicine, with 66% believing the field of diagnostic radiology would be the most heavily influenced. Regarding radiology being a viable career, nearly half (44%) of those surveyed reported that AI made them less enthusiastic about radiology. Gong et al. (2019) conducted a similar study but focused on medical students (n = 322) at all 17 Canadian medical schools. In this study, 67.7% of participants believed AI would reduce the demand for radiologists, and 48.6% responded that AI caused anxiety when considering the radiology specialty. Pinto dos Santos et al. (2019) explored undergraduate medical students’ attitudes (n = 263) toward artificial Intelligence (AI) in radiology and medicine at three German universities. Participants in this study had more positive opinions about AI improving radiology and not making the profession obsolete. Results showed 77% agreed that AI would improve radiology, and 83% disagreed that AI technology will eventually replace human radiologists. A total of 77 percent of those surveyed indicated that AI training should be part of the radiology curriculum.
Instructor Attitudes Toward AI in Medical Curriculum
Much of the research on instructor attitudes toward AI in medical education has focused on the assessment of medical students. Gierl et al. (2014) explored the use of AI software to grade essays written by postgraduate medical students and found that scores generated by the AI systems aligned consistently with those of human scorers at a high level. Gierl et al. (2014) concluded that AI-based grading systems could improve medical education by providing consistency in scoring, reducing the time and cost required for assessment, and providing students with immediate feedback on constructed-response tasks. Kintsch (2002) addressed the use of latent semantic analysis (LSA), which is a form of machine learning, to compare student-constructed responses to a target or model essay. Medical school students are frequently assessed using standardized patients. This involves requiring students to compose a full-text written account of the encounter with the patient that is then evaluated using a rubric. Kintsch (2002) suggests that this practice is an effective teaching method, but grading these responses is challenging. LSA, however, could be used to devise automatic grading thatcould be supplemented with feedback from the instructor.
Promising AI Applications in Medicine
The following section explores several applications of AI in healthcare that are used by patients and healthcare professionals. This section represents a small selection of the AI tools that are currently available or in development.
Chatbots
AI-powered chatbots simulate human conversation with a patient to help assess their symptoms and determine the next steps for care. These automated systems are important in healthcare because they provide information to help patients cure some common illnesses without the intervention of a healthcare professional. Approximately 60% of doctor visits are due to simple health concerns, and 80% of these can be cured with home remedies, simple lifestyle changes, or over-the-counter medication (Bhirud et al., 2019). If additional medical help is needed, the chatbots can provide guidance.
One example of a widely used text-based chatbot is the Centers for Disease Control and Prevention’s (CDC) coronavirus self-checker. This AI-based program is designed for use by patients rather than medical professionals. After the user agrees to the terms of service and inputs demographic information relating to age, gender, race, and location, the chatbot proceeds to list potentially life-threatening symptoms. The user must respond using the on-screen “yes” or “no” buttons. Based on the response, the chatbot will either encourage the user to seek immediate medical attention or will continue asking more specific questions to gather more information. The chatbot can provide general recommendations for avoiding getting sick as well as help users determine if they should see a doctor. The AI-based software provides immediate information to the user from the comfort of their home.
Medical Imaging
Since the onset of the COVID-19 pandemic in 2019, researchers have been looking for ways to help track, diagnose and prevent the spread of the virus. Bansal et al. (2020) reviewed the potential ways in which AI and ML could be utilized to develop predictive models in the battle against COVID-19. Key highlights from their article outline how DL, NLP, sentiment analysis, and machine vision can help with predicting COVID-19 outbreaks. Diagnostic imaging can also harness the power of AI and ML for predicting COVID-19 by using algorithms to detect the virus on lung x-rays. Minaee et al. (2020) developed a deep learningalgorithm that used a dataset containing 5000 chest x-rays to teach the program how to identify COVID-19 on the images. The results of the experiment were considered promising by showing a sensitivity rate of 98% and specificity of 90%. The researchers were so impressed with the outcomes that they made the dataset of 5000 chest images publicly available for other researchers.
Several other areas in medicine arecurrently exploring the potential uses of artificial intelligence to detect and predict the disease. For example, AI is being used to detect different types of cancers. Medical imaging such as computed tomography (CT) colonography can identify the location and size of polyps, but it lacks the ability to help radiologists always differentiate between precancerous and noncancerous tissue. Grosu et al. (2021) demonstrated the ability of ML to distinguish between benign and premalignant colorectal polyps by identifying their differences on CT images with an area under the curve (AUC) of 0.91. In Costa Rica, scientists used AI to create an automated process of reviewing cervical images for the detection of cancerous tissue. Although image quality was a limitation in the study, Hu et al. (2019) proved the algorithm showed promise by having higher accuracy when compared to the traditional cervigram interpretations or conventional cytology (Fig. 1). Dermatology studies for leveraging AI for the identification of cancerous lesions on the skin are also being investigated. Soenksen et al. (2021) used deep convolutional neural networks (DCNN) in their study to distinguish between suspicious pigmented lesions and nonsuspicious lesions on photographs of patient’s backs. The results were considered promising by the researchers in that the algorithm could be used by physicians for providing rapid assessment of patient skin lesions.
Innovative uses of AI are also being used in the field of cardiology. Taylor et al. (2013) are leveraging DL and computational fluid dynamics (CFD) to create simulated models of the blood flow of the coronary arteries derived from coronary CT images. Physicians are then able to review 3D models and measurements of the coronary blood flow specific to each patient. The 3D representations also give doctors a communication tool to discuss coronary artery disease diagnosis with their patients. An overview of this simulation process can be seen in Fig. 2.
Han et al. (2021), Schock et al. (2020), and York et al. (2020) are all pioneering ways to use AI to aid in the area of skeletal radiology. Han et al. (2021) explored using neural symbolic learning (NSL) and deep neural learning (DNL) to help create autogenerated reports on spinal structures identified on magnetic resonance images (MRI). The team developed a two-step framework for autocreating reports with a system demonstrating 95.8% pixel accuracy. Schock et al. (2020) developed a convolutional neural network (CNN) to aid radiologists by automatically segmenting and measuring anatomy on long-leg radiographs (n = 225). The experimental AI system showed to be faster than radiologists at measuring anatomy by 33 seconds. Lastly, York et al. (2020) wanted to gain insight into how patients felt about the use of AI in diagnostic imaging of bones and joints. The team sent questionnaires to patients (n = 300) with a 72.2% completion rate (n = 216). Overall findings from the studysuggest that patients prefer a human doctor to read and assess their medical images instead of AI-based systems.
Ethics with AI in Medicine
Ethical Concerns
In 1970, medical doctor Willaim B. Schwartz published an article that provided valuable insights into the upcoming changes that would happen to the healthcare system due to advances in technology. Dr. Schwartz not only described the innovative uses that new technologies could deliver to the medical world, but he also understood the ethical considerations that would likely arise. The first item he listed as a potential threat was the ability to keep medical records confidential with the soon-to-be implementation of computer systems in hospitals and other medical institutions (Schwartz, 1970). The next issues that Schwartz questioned were the legal and liability of such things as data breaches and incorrect record keeping. Today, over 50 years later, some of those very worries are still a large concern for ethical, legal, privacy, and security reasons. As we bring AI and ML into the world of medicine, these concerns also need to be addressed with firm leadership.
To investigate these issues in today’s healthcare world, Hochheiser and Valdez (2020) reviewed articles published between 2017 and 2019 with a focus on biomedical informatics and ethics in research. The authors discovered three themes among the articles selected: current systems, system designs, and research conduct. They also highlighted the important role newer technologies such as social media could have on patient recruitment for research purposes, but responsible use is needed to maintain study integrity.
Patient Privacy
To protect patients’ medical records, the Health Insurance Portability and Accountability Act (HIPAA) of 1996 was introduced (U.S. Department of Health and Human Services, 2013). HIPAA provides a set of standards for medical institutions to follow in order to protect patient information. Prior to this act, the healthcare industry did not have a national set of requirements to adhere to. Most of the privacy protections addressed under HIPAA cover multiple forms of patient data. The use of AI in healthcare presents newer concerns that will need to be addressed properly even as new technology continues to emerge. Biosensors are one of such newer technologies in which the protection of patient data is causing alarm. Wearable biosensors are sensing devices attached to the human body that recognizes, measures, and/or records a biological element of that person (Kim et al., 2019). Examples of wearable biosensors documented by Kim et al. (2019) include contact lenses, watches, finger clips, and patches designed for blood glucose monitoring. Many of these sensors are approved by the Food and Drug Administration (FDA) and are contributing to patient care. However, the area of patient data and who has access is of great concern. Kim et al. (2019) highlighted the importance of designing information collection infrastructures with strict security and privacy while maintaining proper data management.
It is also important to understand thatdifferent countries comply with unique rules and regulations regarding protecting patient health information. Most nations have their own government entities that are responsible for data safety and compliance. Schönberger (2019) reviewed 300 articles covering legal and ethical issues on the topic of AI in healthcare in Europe and the United Kingdom. He concluded that current laws could be applied to AI technology in healthcare or could be adjusted to accommodate issues with modern technologies. This is a similar viewpoint shared by Vayena et al. (2018) as in their article, it was noted that the European General Data Protection Regulation (GDPR) imposes restrictions regarding the creation, storage, and communication of patient data.
Cyber Security
The important role of cyber security becomes very apparent when discussing the safeguarding of patient health information (PHI). Health information technology (IT) has the complex task of preventing data breaches and ensuring patient information is safe and used properly. In 2018, a team from Western Michigan University investigated cyber security threats in healthcare. Ronquillo et al. (2018) reviewed data breaches during the years of 2013–2017 and found there to have been 128 medical record breaches and 363 hacking occurrences reported in the United States. The team noted that because of the increased use of EMRs (electronic medical records), medical institutions had become targets for hackers to obtain private patient information illegally. The authors concluded that informatics and cyber security infrastructure in healthcare will have to evolve in order to maintain secure environments.
Limitations in AI
The greatest weakness ofartificial intelligence and machine learning is the process in which the learning result is underspecified because data points outnumber the parameters and are known as underspecification (D’Amour et al., 2020). Another term for this complicated issue is called overfitting and occurs when the ML algorithm is too focused. Reliability and credibility can then become problematic in machine learning because results may be incorrect due to the algorithm overlooking certain data or focusing only on certain data points. The team at Google decided to research this weakness in several areas, including computer vision, radiology images, NLP, clinical risk predictions, and medical genomics (D’Amour et al., 2020). The result of their deep dive into each category helped them conclude that creating training models with more trusted biases may help prevent inaccurate solutions producedby some ML algorithms.
Conclusion
In conclusion, the past few decades have shown fascinating new developments using AI in healthcare to improve patients’ health. Advances in AI show the potential for creating innovative solutions in the field of medicine. Both innovative and powerful, AI can be a turning point in solving medical problems and predicting outcomes. However, ethical, privacy, and security measures must be taken under thoughtful consideration when integrating AI into the healthcare setting. With proper and responsible use, AI can change the field of medicine by giving medical professionals an opportunity to learn and develop new pathways.
Although the intention of this literature review was to introduce and offer a historical overview of applications of AI in healthcare for those outside of the field, our study has some limitations that could be addressed by more in-depth future studies. First, the study was limited to articles written in English and with full-text accessibility, which may have excluded promising research. Second, our literature review took a broad approach to reviewing literature on AI in healthcare. Future studies could explore AI-based tools in specific areas of healthcare to produce results and strategies that can be implemented by healthcare professionals.
References
Abbate, J. (2011). Software’s founding mother. Metascience, 20, 215–218. https://doi.org/10.1007/s11016-010-9418-z
Alonso-Silverio, G. A., Pérez-Escamirosa, F., Bruno-Sanchez, R., Ortiz-Simon, J. L., Muñoz-Guerrero, R., Minor-Martinez, A., & Alarcón-Paredes, A. (2018). Development of a laparoscopic box trainer based on open source hardware and artificial intelligence for objective assessment of surgical psychomotor skills. Surgical Innovation, 25(4), 380–388. https://doi.org/10.1177/1553350618777045
Amisha, P. M., Pathania, M., & Rathaur, V. K. (2019). Overview of artificial intelligence in medicine. Journal of Family Medicine and Primary Care, 8(7), 2328.
Bansal, A., Padappayil, R. P., Garg, C., Singal, A., Gupta, M., & Klein, A. (2020). Utility of artificial Intelligence amidst the COVID 19 pandemic: A review. Journal of Medical Systems, 44(9), 1–6.
Bhirud, N., Tataale, S., Randive, S., & Nahar, S. (2019). A literature review on chatbots in healthcare domain. International Journal of Scientific & Technology Research, 8(7), 225–231.
Clancey, W., & Shortliffe, E. (1984). Readings in medical artificial intelligence: The first decade. Addison Wesley.
D’Amour, A., Heller, K., Moldovan, D., Adlam, B., Alipanahi, B., Beutel, A.,... & Sculley, D. (2020). Underspecification presents challenges for credibility in modern machine learning. arXiv preprint arXiv:2011.03395.
Gierl, M. J., Latifi, S., Lai, H., Boulais, A.-P., & Champlain, A. D. (2014). Automated essay scoring and the future of educational assessment in medical education. Medical Education, 48(10), 950–962. https://doi.org/10.1111/medu.12517
Gong, B., Nugent, J. P., Guest, W., Parker, W., Chang, P. J., Khosa, F., & Nicolaou, S. (2019). Influence of artificial intelligence on Canadian medical students’ preference for radiology specialty: A national survey study. Academic Radiology, 26(4), 566–577. https://doi.org/10.1016/j.acra.2018.10.007
Grosu, S., Wesp, P., Graser, A., Maurus, S., Schulz, C., Knösel, T., … Kazmierczak, P. M. (2021). Machine Learning–based differentiation of benign and premalignant colorectal polyps detected with CT colonography in an asymptomatic screening population: A proof-of-concept study. Radiology, 202363.
Hamet, P., & Tremblay, J. (2017). Artificial intelligence in medicine. Metabolism, 69, S36–S40. https://doi.org/10.1016/j.metabol.2017.01.011
Haigh, T. (2009). How data got its base: Information storage software in the 1950s and 1960s. IEEE Annals of the History of Computing, 31(4), 6–25.
Han, Z., Wei, B., Xi, X., Chen, B., Yin, Y., & Li, S. (2021). Unifying neural learning and symbolic reasoning for spinal medical report generation. Medical Image Analysis, 67, 101872.
Hochheiser, H., & Valdez, R. S. (2020). Human-computer interaction, ethics, and biomedical informatics. Yearbook of Medical Informatics, 29(1), 93–98. https://doi.org/10.1055/s-0040-1701990. (Links to an external site.) Retrieved from: https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7442500/
Hu, L., Bell, D., Antani, S., Xue, Z., Yu, K., Horning, M. P., Gachuhi, N., Wilson, B., Jaiswal, M. S., Befano, B., Long, L. R., Herrero, R., Einstein, M. H., Burk, R. D., Demarco, M., Gage, J. C., Rodriguez, A. C., Wentzensen, N., & Schiffman, M. (2019). An observational study of deep learning and automated evaluation of cervical images for cancer screening. Journal of the National Cancer Institute, 111(9), 923–932. https://doi.org/10.1093/jnci/djy225
Jiang, F., Jiang, Y., Zhi, H., Dong, Y., Li, H., Ma, S., Wang, Y., Dong, Q., Shen, H., & Wang, Y. (2017). Artificial Intelligence in healthcare: Past, present and future. Stroke and Vascular. Neurology, 2(4), 230–243. https://doi.org/10.1136/svn-2017-000101
Jordan, M. I., & Mitchell, T. M. (2015). Machine learning: Trends, perspectives, and prospects. Science, 349(6245), 255–260. https://doi.org/10.1126/science.aaa8415
Kaplan, A., & Haenlein, M. (2019). Siri, Siri, in my hand: Who’s the fairest in the land? On the interpretations, illustrations, and implications of artificial intelligence. Business Horizons, 62(1), 15–25. https://doi.org/10.1016/j.bushor.2018.08.004
Kaul, V., Enslin, S., & Gross, S. A. (2020). The history of artificial intelligence in medicine. Gastrointestinal Endoscopy, 92(4), 807–812.
Kim, J., Campbell, A. S., de Ávila, B. E. F., & Wang, J. (2019). Wearable biosensors for healthcare monitoring. Nature Biotechnology, 37(4), 389–406.
Kintsch, W. (2002). The potential of latent semantic analysis for machine grading of clinical case summaries. Journal of Biomedical Informatics, 35(1), 3–7. https://doi.org/10.1016/S1532-0464(02)00004-7
Masters, K. (2019). Artificial Intelligence in medical education. Medical Teacher, 41(9), 976–980. https://doi.org/10.1080/0142159X.2019.1595557
Minaee, S., Kafieh, R., Sonka, M., Yazdani, S., & Soufi, G. J. (2020). Deep-covid: Predicting covid-19 from chest x-ray images using deep transfer learning. Medical Image Analysis, 65, 101794.
Mintz, Y., & Brodie, R. (2019). Introduction to artificial intelligence in medicine. Minimally Invasive Therapy & Allied Technologies, 28(2), 73–81.
Mirchi, N., Bissonnette, V., Yilmaz, R., Ledwos, N., Winkler-Schwartz, A., & Maestro, R. F. D. (2020). The Virtual Operative Assistant: An explainable artificial intelligence tool for simulation-based training in surgery and medicine. PLoS One, 15(2), e0229596. https://doi.org/10.1371/journal.pone.0229596
Moles, J. J., Connelly, P. E., Sarti, E. E., & Baredes, S. (2009). Establishing a training program for residents in robotic surgery. The Laryngoscope, 119(10), 1927–1931. https://doi.org/10.1002/lary.20508
Nadkarni, P. M., Ohno-Machado, L., & Chapman, W. W. (2011). Natural language processing: An introduction. Journal of the American Medical Informatics Association, 18(5), 544–551. https://doi.org/10.1136/amiajnl-2011-000464
Park, C. J., Yi, P. H., & Siegel, E. L. (2020). Medical student perspectives on the impact of artificial intelligence on the practice of medicine. Current Problems in Diagnostic Radiology. https://doi.org/10.1067/j.cpradiol.2020.06.011
Pinto dos Santos, D., Giese, D., Brodehl, S., Chon, S. H., Staab, W., Kleinert, R., Maintz, D., & Baeßler, B. (2019). Medical students’ attitude towards artificial Intelligence: A multicentre survey. European Radiology, 29(4), 1640–1646. http://dx.doi.org.libproxy.library.unt.edu/10.1007/s00330-018-5601-1
Ronquillo, J. G., Erik Winterholler, J., Cwikla, K., Szymanski, R., & Levy, C. (2018). Health IT, hacking, and cybersecurity: national trends in data breaches of protected health information. JAMIA Open, 1(1), 15–19.
Schock, J., Truhn, D., Abrar, D. B., Merhof, D., Conrad, S., Post, M.,... & Nebelung, S. (2020). Automated analysis of alignment in long-leg radiographs using a fully automated support system based on artificial intelligence. Radiology: Artificial Intelligence, e200198.
Schönberger, D. (2019). Artificial Intelligence in healthcare: A critical analysis of the legal and ethical implications. International Journal of Law and Information Technology, 27(2), 171–203.
Schwartz, W. B. (1970). Medicine and the computer: The promise and problems of change. In Use and impact of computers in clinical medicine (pp. 321–335). Springer.
Sit, C., Srinivasan, R., Amlani, A., Muthuswamy, K., Azam, A., Monzon, L., & Poon, D. S. (2020). Attitudes and perceptions of UK medical students towards artificial intelligence and radiology: A multicentre survey. Insights Into Imaging, 11(1), 14. https://doi.org/10.1186/s13244-019-0830-7
Soenksen, L. R., Kassis, T., Conover, S. T., Marti-Fuster, B., Birkenfeld, J. S., Tucker-Schwartz, J., … Gray, M. L. (2021). Using deep learning for dermatologist-level detection of suspicious pigmented skin lesions from wide-field images. Science Translational Medicine, 13(581).
Stone, P., Brooks, R., Brynjolfsson, E., Calo, R., Etzioni, O., Hager, G., Hirschberg, J., Kalyanakrishnan, S., Kamar, E., Kraus, S., Leyton-Brown, K., Parkes, D., Press, W., Saxenian, A., Shah, J., Tambe, M., & Teller, A. (2016). Artificial intelligence and life in 2030: The one hundred year study on artificial Intelligence [Report]. Stanford University. https://apo.org.au/node/210721
Szondy, D. (2015). Fifty years of Shakey, the “world’s first electronic person”. New Atlas. Retrieved from: https://newatlas.com/shakey-robot-sri-fiftieth-anniversary/37668/
Taylor, C. A., Fonte, T. A., & Min, J. K. (2013). Computational fluid dynamics applied to cardiac computed tomography for noninvasive quantification of fractional flow reserve: scientific basis. Journal of the American College of Cardiology, 61(22), 2233–2241.
Turing, A. M. (1950). I.—Computing machinery and Intelligence. Mind, 59(236), 433–460. https://doi.org/10.1093/mind/LIX.236.433
US Department of Health and Human Services. (2013). Summary of the HIPAA security rule. Health Insurance Portability and Accountability.
Vayena, E., Blasimme, A., & Cohen, I. G. (2018). Machine learning in medicine: Addressing ethical challenges. PLoS Medicine, 15(11), 1–4. https://doi.org/10.1371/journal.pmed.1002689(EBSCOhost)
Wartman, S. A., & Combs, C. D. (2019). Reimagining Medical Education in the Age of AI. AMA Journal of Ethics, 21(2), 146–152. https://doi.org/10.1001/amajethics.2019.146
York, T., Jenney, H., & Jones, G. (2020). Clinician and computer: a study on patient perceptions of artificial intelligence in skeletal radiography. BMJ Health & Care Informatics, 27(3), e100233. https://doi.org/10.1136/bmjhci-2020-100233
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this chapter
Cite this chapter
Collinsworth, A., Benjamin, D. (2022). Uses of Artificial Intelligence in Healthcare: A Structured Literature Review. In: Albert, M.V., Lin, L., Spector, M.J., Dunn, L.S. (eds) Bridging Human Intelligence and Artificial Intelligence. Educational Communications and Technology: Issues and Innovations. Springer, Cham. https://doi.org/10.1007/978-3-030-84729-6_24
Download citation
DOI: https://doi.org/10.1007/978-3-030-84729-6_24
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-84728-9
Online ISBN: 978-3-030-84729-6
eBook Packages: EducationEducation (R0)