Keywords

1 Introduction

Artificial intelligence (AI) has become a buzzword of today, reminiscent of the concept of “globalization” and the relating debate on globalization two decades or so earlier (cf. James & Steger, 2014; Rosamond, 2003; Turenne Slojander, 1996). As in the case of globalization and the discourse surrounding it (however defined!), AI remains a concept both powerful and yet insufficiently explained, explored or understood by large sectors of society. AI is hailed by many as the panacea to all the ills of the current socio-economic model as well as a source of unimaginable opportunities (cf. Zhuang et al., 2017; Eager et al., 2020; Floridi, 2020). However, AI, especially in popular discourse, is also seen as a source of risks and threats. Clearly, it is too early to imagine and understand all the possible implications, positive and negative, that the use of AI-enhanced solutions may trigger for our societies, including their politics, economics, business sectors, competition, and innovation, as well as education and healthcare provision etc. Therefore, a thorough and unbiased conversation, including academics, researchers, and practitioners, on AI and AI-related topics is necessary. This chapter and thus the entire volume respond to this plea by offering an insight into diverse aspects of AI and its use across wide-ranging issues and domains, including security and military affairs, the decision-making process, and public administration. Interestingly, amidst the frequently alarming media accounts of AI and the alleged existential risks and threats to humanity that it bears, AI-based solutions, including services and applications, have been present in our lives for some time. Consider the autocorrect feature in smart phones, and how it picks the key language you use daily, and how it learns throughout the process to recognize which language you are using at a given moment when sending your messages or typing your emails. However, AI is much more than language and speech recognition applications in the smart phone. AI and AI-based techniques and applications, and, for example, AI-based solutions, have relevance across very many issues and domains. At present, when advances in AI and related technologies have been transferred from the field of research to application, the challenge is to create appropriate, resilient and flexible regulatory frameworks to ensure that the risks and threats that AI holds for our societies can be preempted and addressed, whilst ensuring that the development of AI and, thus, innovation centered on AI and its ecosystem are not suppressed. To address these complex issues, a thorough understanding of the nature of AI, its limitations and potential is necessary. This chapter aims to provide this.

The discussion in this chapter is structured as follows. In the next two sections, the “dry” i.e., quite technical, facets of AI are discussed. Here a case for an AI ecosystem is made and the notions of artificial neural networks and unsupervised machine learning are discussed. In what follows, technology-related limitations of AI as well as possibilities are outlined briefly. An overview of AI’s implications for the (global) economy and selected policies follows. The ethical concerns are discussed in the concluding section.

2 AI and Its Ecosystem

The popular understanding of AI and, for that matter, the key lines of the popular narrative on AI portray AI as a sort of crystal ball capable of foreseeing the future, a super-computer capable of resolving the most complex problems, a super-weapon of total destruction, and perhaps also as a super-machine waiting to replace human beings. This conception of AI, characterized by fear of the unknown, has been famously captured in Lem’s (1964, 1990) visionary robots’ fables, the thrust of which concerned the big question of the human–computer (aka robot) interaction. The onset of AI and, perhaps most importantly, the machine-learning-mediated possibility for people to enter a form of interaction with a computer/machine revives the basic question of how to conceive of the absolutely new form of communication, i.e., with a non-human. To be clear, the objective of this chapter and thus, also of the entire volume, is to showcase that, regardless of the potential inherent in AI and related technologies, AI should not be reified or, in other words, considered as ontologically distinct from the human being, even if valid fears exist that AI might replace the human (Archer, 2021; Rakowski et al., 2021; Sætra, 2020; Fiske et al., 2019). To understand why, it is necessary to shed light on the very essence of AI.

Contrary to popular (mis)understanding, AI is best conceived as an ecosystem of sophisticated technologies, techniques, and applications. Their emergence is a function, first and foremost, of the development of ever more efficient processors capable of handling ever more complex calculations in an ever-shorter time span, and thus delivering greater power necessary for our computers to operate and perform the tasks we require (cf. Khan et al., 2021). Central to the emergence of AI is the development of neural networks, i.e., several layers of interconnected processors that, similarly to neurons in a human brain, have the capacity to share impulses and, thus, data.

The emergence of AI is directly related to the onset of the big data paradigm in computer science, and thus the ability to process huge data sets in a short period of time, a task impossible before (Farami et al., 2021; Visvizi et al., 2021). The emergence of the big data paradigm changed the way data is conceived and valued. That is, now, huge datasets become a source of invaluable information across domains (Lytras et al., 2021). Yet the capacity to process the phenomenally huge and complex sets of data is only possible due to the emergence and application of such software technologies as Apache Hadoop and other. These technologies enable, on the one hand, a different and a more efficient way of processing data, and, on the other hand, distributed processing of large data sets across clusters of computers by means of simple programming models.

In other words, as Fig. 1 outlines, any discussion on AI needs to be based on the recognition that it is an AI ecosystem rather than a single super-technology. Specifically, the emergence, development and utilization of AI are driven by three major paradigms that shape the debate in the discipline of computer science, i.e., the big data paradigm, the quantum computing research, and green computing (Hu et al., 2021; Piattini et al., 2020; Lytras et al., 2021). These three broad debates, representing interconnected but paradigmatic shifts, shape the developments in the discipline and in the field. As such, they are consequential for the remaining items in the AI ecosystem, such as the (i) hardware-enabled big data processing frameworks and open-source software utilities, (ii) hardware- and-software enabled technologies; (iii) related techniques and applications already in application; (iv) emerging technologies and applications. Notably, the landscape of techniques, technologies, and applications, all fostered by advances in the three basic paradigms that shape advances in information and communication technology (ICT) today, changes rapidly. Thus, stories that hold truth today, are subject to contestation and revalidation tomorrow. Knowledge is emancipatory after all (Archer, 1998; Bhaskar, 1975).

Fig. 1
figure 1

Source The Author

AI and its ecosystem.

3 AI and the Neural Networks: From Supervised to Unsupervised Learning

To put it simply, AI is the—enhanced by super-processors—process of grouping, structuring and analyzing huge sets of data, especially of unstructured data, i.e., the kind of data that has not been previously “tagged”. Notably, AI is not a new term. Already in the 1950s, attempts were made to utilize the learning potential of computers of that period. At that time, however, the focus of the activity was directed at so-called supervised machine learning. The latter employs algorithms that require structured or tagged data. The machine learning process consists in this case of the classification of data based on a key, i.e., a key included in the algorithm, and drawing conclusions based on the classic probability tree. Even if supervised machine learning is still in use in some domains today (cf. Jiang et al., 2020), it displays fundamental weaknesses. Its efficiency is directly related to the possibility and ability of the human, in this case the system manager, to label or tag the data. Considering that the data sets are huge and that the supply of data is massive and continuous, tagging (or labelling) data is time-consuming, requires a lot of skill and thus is very expensive.

In contrast to supervised machine learning, unsupervised learning—and it is the essence of the concept of AI—allows unstructured data to be fed into the system and subsequently examined (cf. Alloghani et al, 2020; Rajoub, 2020). To explain the meaning of unsupervised learning, it is necessary to highlight three basic terms, i.e., artificial neural networks, algorithms, and deep learning. The artificial neural networks consist of an immense number of connected processors operating at the same time, which as per the design—are supposed to mirror the functioning of the human brain. Each of the processors that builds a part of each of the neural networks has access to the local memory and is fed by, on the one hand, huge data sets and, on the other hand, information about the relationships among the data. The neural networks work as a system, i.e., they may overlap—thus the talk of layers of the artificial neural networks—and co-operate.

The content and quality of the algorithms define the ways in which the neural networks and the layers operate and co-operate. A given algorithm defining the functioning of a neural network is responsible for the process of (i) recognizing the characteristics of data available in the system, a process called feature introspection; (ii) identifying similarities among chunks of data, a process called similarities detection; and (iii) grouping these unstructured data into clusters. In other words, algorithms establish the relationships between and among data, e.g., vanilla ice-cream (taste), and not, say, red ice-crem (color) (Strickland, 2019). Taking into account that neural networks can overlap, i.e. it is possible to use several layers of neural networks, and an algorithm enables the exchange of data/information among them, it is possible to talk about the learning process of a neural network. However, for a neural network to learn, the algorithm and the software that operationalize it must demonstrate to the neural network how it has to respond to external cues, e.g., to data introduced by the user.

AI is based on the utilization of the outcomes of this kind of multilayered co-operation of complex neural networks. Precisely this feature of unsupervised learning, consisting of the multilayered co-operation among neural networks, defines the thrust of what is otherwise referred to as “deep learning”. Viewed in this way, the process of deep learning is feasible due to specific algorithms and advanced processors. Under these conditions, in a complex multilayered neural network, based on co-operation among specific layers of the neural network, the following processes take place: introspection, similarities’ detection, clusters’ grouping, and, finally, delivering an answer to a question included in the algorithm.

Considering that the number of data available in the system increases, the artificial neural networks have to expand, the power/capacity of processors has to increase, the number of relationships between and among chunks of data becomes denser, and thus the learning capacity of a given neural network increases, always in line with the logic defined in the algorithm that defines its operation. It remains to be seen how and in which direction artificial neural networks will develop. Now, irrespective of the fact that AI tends to be viewed as an embodiment of unsupervised learning, and therefore also as a form of deep learning, the classic rules deriving from the science of probability are also at play in the process of the functioning of neural networks. Having said that, it is necessary to dwell on the limitations, possibilities and risks inherent in neural networks.

4 Limitations, Possibilities and Risks Inherent in Neural Networks

AI is directly correlated with the processing power of respective processors as well as with the quality of the algorithms that define a neural network’s operation and co-operation. From a different angle, regardless of arguments suggesting that neural networks can do everything, neural networks are not intelligent enough. In other words, a neural network cannot make decisions about and create the logic of thinking that subsequently will allow it to connect data in line with certain rules. The latter are quintessential for the processes of introspection, similarities detection, and clustering, i.e., for the process of building a logical whole out of a myriad of pieces. Notably, attempts to teach neural networks, e.g., by means of generating cookbooks led to quite anecdotal combinations such as “Take 250 gr of bones or fresh bread … ”, etc. In other words, the problem, or a challenge, rests in the fact that a neural network repeatedly committed mistakes suggesting it did not have any memory at all (Strickland, 2019). To put it differently, artificial neural networks can only work forward, i.e. in a range from T0 to T1. They cannot predict the future. Artificial neural networks can support real-time tasks and commands.

Another, by now classic, example that was meant to hail the success of AI relates to AlphaGo, i.e. the software and a game from the company DeepMind (cf. Granter et al., 2017; Li & Du, 2018; Holcomb et al., 2018). Go is a traditional, highly complex, Chinese board game, and its history goes back 3000 years. Attempts to utilize supervised machine learning to develop a software version of the game allowed for an amateur-level version of the game (cf. Lu et al., 2020). This is because the number of possible moves, the number of their combinations, and the evaluation of their relative values for the follow-up of the game were far more than the possibilities inherent in the traditional probability “trees”. The advent of neural networks and so, also, the onset of unsupervised learning made it possible to bypass this problem. That is, the software AlphaGo, developed by Deepmind (and purchased by Google for USD 500 m in 2014, (cf. Shu, 2014), employs an advanced probability equation to filter the layers of the neural network that is structured according to such values/functions as “politics”, “values” etc., i.e., items specific for the game (cf. Binder, 2021).

An interesting discovery relating to the AlphaGo is that the software was unable to define what “human being” is. Even more so, the software was unable to discern whether it was competing with a machine or with a human being. To put it differently, at present, the algorithms underpinning the functioning of neural networks demonstrate great precision whenever the problem/task at hand is narrow and very well defined. Thus, the more specific a given problem, the smarter AI appears to be. The opposite is also true. For instance, should a given algorithm be designed to generate pictures and, more specifically, pictures of birds, it will not be able to generate a picture of any other animal.

Discussion of possibilities inherent in AI needs to take into account the parallel advances in sophisticated technologies, applications, and techniques related to the process of the handling of big data. This includes the highly technical issues already at the level of the processor and its speed, as well as exchange of information between and among processors, etc. These factors, as mentioned earlier, directly influence the possibilities relating to deep learning, and especially the pace of deep learning and the condensation of information that it may lead to. Possibly, if only the processors’ capacity would increase as fast as the increasing quantity of data they are fed with, one could argue that the potential inherent in AI was unlimited. At this point, it is worth mentioning the case of deep learning accelerators (see also Fig. 1), i.e., a technology that seeks to mitigate some of the technical limitations specific to the limited overall speed of the existing processors. In this context, it is also important to highlight the case of quantum computing, i.e. an emerging paradigm that revolutionizes the way in which the very process of computing is conducted and that, at the same time and for precisely this reason, a myriad of possibilities and applications across fields and domains open up (Hassija et al., 2020; Lee, 2020; Outeiral et al., 2021).

With regards to AI’s limitations, several barriers exist that limit the feasibility of its utilization. These barriers can be identified at diverse levels and in several overlapping domains and issues specific to contemporary politics, society, and the economy. These barriers may include: digital illiteracy; lack of awareness and expertise regarding the use of AI-enhanced tools; lack of resources to purchase AI-based tools and solutions; and a lack of, or insufficient, infrastructure that would enable rightful access to data and their efficient use. The poor quality of data, i.e. data in an improper format, constitutes another obstacle to efficient AI utilization (cf. OECD, 2015; Hatani, 2020). As shall be discussed in the following section, obstacles and limitations to the use of AI-based solutions rest also in the regulatory field, including questions of privacy, personal data protection, safety and security, grey areas of regulations, as well as regulatory barriers, insufficient complementarity of regulation in the field of AI in the international context and many others. Considering the breadth of the discussion focused on contentious issues pertaining to the emergence, development and application of AI in modern world, the following section will merely outline the key areas of concern. The content of the volume (cf. Visvizi & Bodziany, 2021) will serve as a necessary addition to the brief introduction included in this chapter.

5 AI and the Society: Focus on the (Global) Economy and Selected Policies

From the perspective of what is technically feasible, the onset of AI and AI-based solutions, applications and techniques is a source of a great number of opportunities. These opportunities may be seen through several interpretive lenses. For example, viewed from the perspective of the global economy, it may be assumed that further automation and digitalization of production, this time employing AI-based solutions, would allow an aggregate global productivity increase. This would be due to the work substitution effect: the rise in innovativeness and innovation and the adoption of new technologies; increased competition; diversification of products; optimization of supply-chains as well as of global value chains, etc. From a different point of view, AI is a source of great hope in the field of healthcare in that the inroads of AI-based techniques and applications, always in connection with other technologies, revolutionize the way in which data is handled for the purposes of early detection and diagnosis of diseases, treatment, medicine development, the analysis of archived patients’ records, surgery and so much more (Lytras et al., 2021). AI, as the content of this volume amply demonstrates, is employed in public administration, the services sector, marketing (Huang & Rust, 2021; van Veenstra et al., 2020; Loukis et al., 2020) and, of course, in the field of the military. Viewed in this way, the hopes, prospects, and opportunities that AI-based solutions bring to our society are immense. However, as is the case with any new technology or an ecosystem of technologies, tools, techniques and applications, downsides always exist, and a period of adjustment is necessary. Consider the case of the automobile in the late nineteenth century (Berger, 2001), and the regulatory frameworks that we continue to put in place to strike a sound balance between the benefits of using a car and the externalities that the aggregate use of vehicles creates for the environment.

The case of AI proves that, even if its value added and its potential are immense, our societies, but perhaps more precisely the regulatory frameworks and the governance structures in place, are not fit enough to accommodate not only the plethora of regulatory challenges ahead but also the variety of resulting, and yet still unaccounted for, policy-making considerations that will have to be addressed (cf. Ulnicane et al., 2021). Substantial effort has been undertaken at the global (G20), international (United Nations, the Organization for Economic Co-operation and Development), regional (the European Union (EU)), and national levels to address the challenge (e.g., Chatterjee, 2020). In some ways, as more critical voices outline, the “race to AI” has turned into a global “race to AI regulation” (Smuha, 2021), in which a country’s capacity to influence international standards and regulations in the field of AI turns into competitive advantage (Dexe & Franke, 2020). Consider the case of the EU, a global player after all. 290 AI policy initiatives devised by the EU member states exist. The EU, especially the European Commission, makes an explicit attempt to promote a broad approach to AI (European Commission, 2020; European Commission, 2021). AI is thus seen as a tool, on the one hand, to help the EU compete globally and, on the other hand, to—even if this notion is quite implicit—allow the EU to reclaim its position as the “normative power of Europe” (Manners, 2002). In relation to the first objective, the European Commission places emphasis on innovative capacity at large, and in this context on the use of data and developing “critical computer capacity”. In relation to the second objective, the European Commission’s narrative is filled with references to “AI for the people”, “AI as a source of good”, and to “trustworthy AI” (European Commission, 2018).

It is difficult not to be critical of the approach to AI that the European Commission promotes. It seems like too little, too late, and not comprehensive enough. In other words, the Commission’s approach to AI and to the prospect of reaping the benefits of AI, resembles hand-steering rather than anything else. The point is that too much emphasis is placed on direct-support measures for, say, research and investment in AI, at the expense of letting the market mechanisms play their due part as well. This is certainly a nod to all those who deal with economic policy in specific EU member states and with economic policy coordination at the euro-area level. The notions of the EU competition policy and industrial policy complete the picture. Full stop. The point is that hand-steering alone will not allow small- and medium-sized enterprises (SMEs) in the EU, or elsewhere in the world, to pick up and adopt AI-based tools and solutions in their daily activities (cf. Troisi et al., 2021). For this, tax incentives and flexible labor markets are needed to enable SMEs to acquire talent, which is rare and highly expensive, and invest in AI-based solutions.

The downside of this story, as many observers note (cf. Naudé, 2021), is that, since AI-based tools are key to digital platforms and digital platform economy, and data is the key source of advantage, as the case of Google, Apple, Facebook, Amazon, and Alibaba (GAFAA) illustrates, we witness a major distortion of competition these days. The thrust of the problem is that the barriers to entry are so high that companies like GAFAA lack natural competitors. From a different perspective, though, the case of Microsoft and the way it was dealt with in the context of the EU competition policy more than a decade ago demonstrates that a thin line divides attempts to defend fair competition from those to create disincentives for innovation (Larouche, 2009; Manne & Wright, 2010). In brief, the case of the EU is illustrative of the scale of the challenge that the inroads of AI create for the economy and for the regulators.

Looking at this issue from a different perspective, it is necessary to mention the “first-mover advantage”, i.e., a situation when companies that were first in the field are capable of maintaining their dominant position, at the expense of other market players (cf. Park, 2021). This is already happening, cf. GAFAA. Consider, however, the global implications of the same process. That is, how will the inroads of AI shape and reshape the Global North-Global South relationship? In the context of the debate on the Global South, the “first-mover advantage” thesis suggests that companies from the Global North, i.e., current leaders, might be able to capture the opportunities first, effectively denying them to actors in the Global South. This might lead to exclusion and reproduction of the old global dependency patterns. Of course, the case of China adds to the complexity of the issue. In brief, with reference to AI and its impact on the global economy, time and coordination are crucial. It remains to be seen whether a consensus on these issues will be reached at G20 any time soon.

6 By Means of Conclusion

While the economy has been centrally featured in the discussion in the previous section, this chapter would be incomplete if some of the ethical considerations pertaining to the development and adoption of AI-based solutions were not mentioned. Several exist. The thrust of the concern is twofold, i.e., to what extent will the computer, the machine, be able to operate independently of the human being and regardless of the human being’s express consent, and how can it be ensured that it will be a force for good? There is a growing body of literature that deals with the question of ethics and AI (Jobin et al., 2019; Coeckelbergh, 2020; Vesnic-Alujevic et al., 2020; Roberts et al., 2021; Ryan & Stahl, 2021; Rakowski et al., 2021; Neubert & Montañez, 2020). International organizations, such as the UN, the EU, the World Health Organization, as well as national governments (Dexe & Franke, 2020) are central in shaping and developing the debate. A very useful typology of ethical considerations that the debate on AI and ethics revolves around has been introduced by Jobin et al. (2019) and further elaborated by Ryan and Stahl (2021, p. 65). In brief, the following ethical considerations are key for the debate in AI in our societies: transparency, justice and fairness, non-maleficence, responsibility, privacy, beneficence, freedom and autonomy, trust and trustworthiness, sustainability, dignity, and solidarity. What are the practical implications thereof? As the chapters included in this volume demonstrate, the use of AI-enhanced tools in the battlefield is one of the most contentious issues in this respect. This certainly applies to unmanned aerial vehicles (UAVs), or simply drones, and warfare. Consider non-maleficence. The case of facial features’ recognition in China raised very serious ethical concerns. Consider justice and fairness, privacy, freedom and autonomy and so on. The question of who, under which conditions and for what purpose, has the right to collect, or who gains access to, our personal data is another case that highlights sensitive issues surrounding AI-based tools and techniques. The General Data Protection Regulation (GDPR) implemented in the EU since 2018 is one of the strictest regulatory responses to this issue in the world.

The conversation on AI and ethics leads to another question, i.e., prospective directions of growth and development of AI and AI-based tools. In this context, it is necessary to return to the initial point that AI is bound to fundamentally change the relationship between the human being and the machine. This time, it is not a literary question. This is because current and future users of technology will have to learn a form of interaction with a non-human. Consider the case of education and technology-enhanced learning based on AI and cognitive computing. By means of enhancing the user’s experience, and thus amplifying the teaching and learning process (Lytras et al., 2018; Visvizi et al., 2020), already today, young students should acquire the necessary skills and ethical stance that would allow them to shape the future human–computer relationship in a responsible manner.

Another issue that should be mentioned here relates to the question of the impact of AI on inequality, exclusion, and, from a different perspective, non-growth. It is possible to imagine that the emerging economic system, to a large extent based on digital platforms (Alahmadi et al., 2020; Malik et al., 2021), will also lead to the consolidation of a new system of social relations, of wealth, and wellbeing, both locally and globally. Depending on the way of reading it, this scenario may be devastatingly daunting. To prevent it from happing, transparency understood as explainability, explicability, understandability, interpretability, communication, disclosure, and showing, is the key issue in the debate on AI. This chapter has sought to do just that.