1 INTRODUCTION

Artificial intelligence (AI) technologies are becoming increasingly widespread in the industrial and everyday life of society. The growing popularity and rapid development of technologies lead to the complication of tasks solved with the help of AI, and, as a result, to the fact that information processing on a standard user device is performed critically inefficiently, which is unacceptable for the end user. In this case, the solution is the use of cloud technologies (CT), which in turn have already gained wide popularity and have developed their methodology. It is also worth noting that AI and CT are of great interest in the scientific community. For example, projects such as [1, 2], have gained popularity in the daily field of activity, while they are scientific projects in the field of AI. When AI and CT methods are used together, a number of both standard and specific problems arise related to reliability, security and confidentiality. Considering that of all the functionality implemented by CT, AI uses mainly cloud computing (CC), for AI using CT, all the problems characteristic of CC occur. CC security threats can generally be divided into two categories: external threats and internal threats. External threats include various attacks by adversaries aiming to steal information (for example, hacking) or to damage information (for example, DDOS attacks) [3]. Internal threats in general are a combination of all possible ways to compromise the security system from the inside. To combat the latter, for example, secret sharing schemes are used [4]. However, the highest level of security is achieved when using methods that allow you to process data in encrypted form. In this case, the probability of compromising the system is reduced to a minimum. A possible solution to the problem of confidential data processing is homomorphic encryption. This paper presents a study of modern AI privacy methods and algorithms used in practice and a predictive study of methods that may be used in the future.

The work consists of 4 sections. Section 2 discusses the main aspects related to AI and cloud technologies. Section 3 discusses approaches to building a privacy-preserving AI. Section 4 presents the results of the analytical review.

2 ARTIFICIAL INTELLIGENCE AND CLOUD TECHNOLOGIES

2.1 Artificial Intelligence

Artificial intelligence (AI) originally developed within the framework of intelligent systems and is still considered one of their components, namely the ability to perform creative functions. Initially, the need for intelligent systems was due to the automation of decision-making. That is, a certain reaction to certain events was expected from the system. With the development of computer technology, methods and algorithms, as well as methods for developing systems and applications, these reactions became more complex and more flexible.

Historically, the first AI methods were machine learning (ML) methods. ML is a class of AI methods whose task is not to solve a problem directly, but to find it through learning based on the analysis of many solutions to similar problems [5]. The ML problem statement can be defined as follows. There is some unknown relationship between the two sets. Only the precedents are known, i.e. pairs of these two sets, which are called a training sample. Based on these data, the task is to restore dependence, that is, to build an algorithm that will create a new pair with a given accuracy. In fact, ML is tasked with approximating a function, but not necessarily by another function, but by some kind of algorithm.

AI methods can also be divided according to the method of learning. For example, there are reinforcement learning models [6], among such models, genetic algorithms can be distinguished. Genetic algorithms are heuristic search algorithms used to solve optimization and modeling problems by random selection, combination and variation of initial parameters, which are based on methods similar to natural selection in nature [7]. There is also teaching without a teacher, such methods are used to solve, for example, the clustering problem [8]. However, the largest group of methods consists of methods using teaching with a teacher, where for a set of precedents (known pairs of input and output data) it is necessary to build an algorithm that returns the required solution [9].

Artificial neural networks or simply neural networks (NN) are a mathematical model that is based on the principle of functioning of networks of nerve cells of a living organism [10]. The development of computer technology has made it possible to create NN models of great complexity. In this context, several events can be identified that have allowed the expansion of the use of AI, this is the emergence of various hardware and graphics accelerators [1113], as well as valve matrices [1417] for AI, which allow you to solve more a wide range of tasks previously unavailable. The work of such NN models is carried out through the use of so-called deep learning (DL). In fact, DL is a learning process for multi-layered NN. Theoretically, 2–3-layer NN is sufficient to solve a wide range of problems, however, DL is often used to solve complex problems, which shows good results [18]. DL includes methods such as the limited Boltzmann machine. Convolutional neural networks are also built on the basis of DL, which use different forms of convolutions on different layers [19]. They are often used for pattern recognition [20]. One of the most popular modern AI technologies is GPT technology. Even at the beginning of the development of GPT models required large training samples, which take up tens of gigabytes at the pre-training stage, and given the fact that the number of users of models with GTP exceeds 100 million, the amount of resources consumed exceeds the resources available to one device, if we consider the average office computer. The effective operation of such technologies requires the use of distributed computing systems.

2.2 Cloud Computing

Let’s take a closer look at cloud technologies (CT). In fact, CT and CC are synonymous, since any data processing in the cloud involves some kind of calculation. The situation is similar with cloud storage, since when processing information (for example, search) Certain calculations are also performed.

They are a development of the distributed computing model [21] with some exceptions. Distributed computing assumes the presence of parallelism in computing, namely the integration of computing resources into a parallel computing system. The implementation of such a system is also possible on a single physical device, for example, a server rack or a supercomputer [22]. The transition point between distributed computing and cloud computing can be considered grid computing [23]. The main difference between CC is its conceptualization [24]. Unlike distributed and grid computing, which are primarily a set of tools and methods for solving computationally complex problems, CC is primarily a service to provide opportunities for the implementation of elastic computing. It is customary to classify CC according to the types of models within which this service is provided:

\( \bullet \) Software as a Service (SaaS) – this model involves using a cloud infrastructure to implement CC by providing the user with an application software package. The control and management of the infrastructure is carried out exclusively by the service provider [25];

\( \bullet \) Platform as a Service (PaaS) – in this case, the user is presented with an infrastructure for hosting various basic software. These are usually tools for creating software and, for example, database management systems. The provider can also provide various environments for working with programming languages [26];

\( \bullet \) Infrastructure as a Service (IaaS) – in this case, the provider provides the user with the most complete rights to use the cloud. Here, the user is offered a basic infrastructure within which he independently organizes the processes of managing computing resources, as well as building a network and storing data. The user also independently controls the operating systems that are deployed in the cloud. Another difference from the previous categories is that the user is given limited control over the network services of the cloud space allocated to him [27, 28].

Based on these three models, various hybrid ones are also distinguished, such as Data Base as a Service [29], Monitoring as a Service [30], etc., however, hybrid types, in fact, are divided in terms of customer needs and are rather aimed at narrow specialization of the application of a particular type of CC. Based on the above information, it is possible to highlight the advantages that CC provides for AI. As already mentioned, AI methods are quite computationally complex. When conducting research and developing applications and services related to AI, the researcher/developer faces difficulties related to limited computing resources. In the case of a researcher, in theory, you can use the distributed computing system of your institution, but not every institution has its own powerful computing system, or access to it is difficult. CC’s allow a researcher/developer to rent computing resources from a vendor.

2.3 Cloud Computing and Artificial Intelligence

Let’s consider the applicability of various CC models for AI. In general, AI can be deployed within any of the three service delivery models. However, there are several nuances, in the case of SaaS, the supplier, in addition to computing resources, also supplies AI, in this case AI is also a service. An example of such a service is MLaaS [31]. In the case of PaaS, the provider provides the user, in addition to computing power, with AI development tools. Various cloud services for software development can be included in this category. A striking example is Google Colab [32]. Despite the fact that this solution is not specialized, it still provides the user with opportunities for AI development and research. At the same time, there are solutions aimed specifically at working with AI [33]. IaaS and AI are quite difficult to distinguish into a specialized category, since in this case the service provider provides exclusively computing resources and the infrastructure connecting them. In this case, specialization in AI is achieved due to two factors: due to the fact that the hardware of the technical component of CC is selected in such a way that AI works as efficiently as possible; as well as due to the very position of the service provider [34]. In the cases considered, AI acts as an object, either being part of a service, or being an actual consumer of computing power, however, there are works that consider other possibilities for using AI in CC [3538]. For example, AI can be used to build infrastructure and load balancing.

2.4 Data Privacy in Cloud Computing

If the projects, as in the cases [1, 2], are open, i.e. they do not initially contain confidential information, then the requirements for their security are low. However, any AI can be used in tasks related to the processing of confidential data. It could be medical information [3944], banking [4548], government [49, 50] and all kinds of personal data. In these cases, a situation arises in connection with which CC is often criticized both in society as a whole and in the scientific community. When processing any information in the cloud, both parties have access to it – the user and the service provider. In this case, the degree of data confidentiality is determined by an agreement between the user and the provider, and how this agreement is respected is the responsibility of the service provider. It is not uncommon for confidential data stored or processed by various service providers to be compromised. In this case, the user is forced to give preference to suppliers who have either proven themselves to be reliable in terms of ensuring the confidentiality of the processed data, or claim a low probability of compromise through the use of effective protection methods. This raises a number of questions. Which privacy method is effective for storing data? What methods are capable of providing the necessary level of confidentiality in the context of data processing using AI in general and Big Data in particular?

The purpose of this work is to answer the questions asked and provide the reader with the opportunity to familiarize themselves with current methods of ensuring the confidentiality of CC using AI.

For a more visual demonstration of the retrospective review of distributed computing technologies and AI, a table was compiled (Table 1). The above works are not presented in chronological order, Table 1 is designed to restore it. Note that some technologies were first introduced much earlier than they were developed and studied in full.

Table 1. Historical background on the technologies considered

Analyzing the data in Table 1 and the overview presented above, it can be noted that theoretical AI models have been actively developing since the middle of the XX century, but practical models were developed in the late XX–early XXI centuries. The table also shows that the development of practical AI models coincides with the late stages of the development of distributed computing (until the end of the 20th century) and the emergence of CC at the beginning of the 21st century. This coincidence is not accidental and is due to the fact that the development of distributed and cloud computing all this time has been on the path of aggregating more and more computing power, which was so lacking in large AI models. Thus, it is fair to say that distributed computing has made a significant contribution to the development of AI models, and CC allows us to develop and create modern computationally expensive AI models, the complexity of which is growing every day.

Given the above, it can be argued that the issue of CC privacy has the same importance for AI as for any CC “consumer,” which AI essentially is. Next, the methods of ensuring the confidentiality of AI in CC will be considered.

3 MODERN METHODS AND ALGORITHMS FOR ENSURING AI PRIVACY

3.1 Differential Privacy

The concept of confidentiality is somewhat lax. Depending on the situation, it can be viewed from different angles. For example, from the point of view of medicine, it is enough that the data is anonymous, i.e., the medical history is depersonalized. In [70], a study of DL methods for ensuring soft confidentiality is carried out. The authors achieve their goal by introducing differential confidentiality, i.e. by mixing private depersonalized data together with synthesized data. The authors have demonstrated the high accuracy of this method. However, this method has a number of disadvantages: the basis of confidentiality is the introduction of additional noise and distances between points during gradient descent. This not only increases the computational complexity of calculations, but also does not ensure proper data security, since if an attacker gets access to the data during processing, he will be able to remove unnecessary noise.

On the other hand, [71] considers a fundamentally different approach to ensuring data confidentiality from the above. The authors consider the possibility of training and using a neural network by a group of users without the need to disclose confidential data to each other. This possibility is provided by the feature of stochastic gradient descent, which allows it to be performed in parallel and asynchronously. The authors also claim that their solution allows participants to train the neural network independently on their own (confidential) sets by sharing subsets of key parameters. In general, this method allows you to ensure confidentiality among the group members, however, it turns out to be ineffective in the case of prior collusion of the participants or an external attack.

3.2 Secret Sharing Schemes

In the context of building confidential AI in CC, it is also necessary to consider secret sharing schemes (SSS) [4]. Here are briefly the main theoretical aspects related to SSS. The secret S is divided by dealer D between N participants in such a way that to decrypt any information, it will be necessary to combine the shares \({{S}_{i}}\) of all participants in the scheme back into a secret, i ∈ 1, 2, …, N, where N is the number of participants. If the secret is, for example, the key for the encryption scheme, the confidentiality of the system increases.

There are two types of secret separation schemes: full [4] and threshold [72]. The full ones assume that all the fractions of the secret are needed for recovery, the threshold ones mean that a certain number of fractions are needed, but not all. Threshold secret sharing schemes are used much more often in practice due to their flexibility. Threshold SSS allow you to reduce the computational load, while the threshold is set so that an attacker cannot master the required number of shares of the secret, or so that a preliminary collusion of the required number of participants cannot take place.

In [73], the authors propose a distributed DL when participants train a neural network using their confidential sets. The authors cite the results of a study in which it was possible to maintain confidentiality with distributed DL in the cloud with untrusted participants. The work demonstrates the functionality provided by SSS when working with AI, and the very fact of the possibility of confidential training. The authors apply Shamir’s scheme [72]. During its existence, the scheme has proven its suitability from a security point of view. However, if we consider the application of SSS methods from the CC point of view, then security problems are also overlaid with reliability problems, additional corrective codes impose additional loads on the system. To offset the need for additional correction codes, SSS based on the residue number system (RNS) [74] can be used, for example, such as SSS Asmut-Bloom [75] and SSS Mignotte [76].

The paper [77] considers the organization of federated training of neural networks in cloud systems using SSS Asmut-Bloom to ensure confidentiality [78]. The authors also consider the SSS Mignotte, but there are works proving its unsuitability for security.

In [79], the authors propose a protocol for confidential data transmission in conditions of working with neural networks. The transmission protocol is based on SSS. The main focus of the article is on the speed of data processing while maintaining their confidentiality. The authors show the effectiveness of their solution in comparison with some methods of homomorphic encryption, which is also applicable when building confidential AI in CC.

3.3 Homomorphic Encryption

The first works on homomorphic encryption (HE) appeared several decades ago. It allows the processing of encrypted data without the need for a decryption operation. Homomorphic encryption schemes can be homomorphic in addition, multiplication, or both of these arithmetic operations simultaneously. This feature is typical for various asymmetric ciphers. For example, homomorphic addition is supported by schemes presented in [80, 81], and homomorphic multiplication by schemes from [82, 83]. It is worth noting that such encryption schemes are still being used, for example, in [84] the authors use a modified El Gamal scheme to train a neural network, based on the fact that the El Gamal scheme is homomorphic in multiplication. The authors use a linear approximation of the activation function. This method is quite narrowly focused and difficult to scale.

In 2009, Gentry developed a fully homomorphic encryption scheme (FHE) [85]. This scheme was not computationally efficient enough, but further research by its followers allowed to increase efficiency to a level sufficient for real practical application of FHE [8691]. FHE allows you to perform both homomorphic addition and multiplication (however, it has a limit on the number of multiplications with one key), in addition, there are schemes that allow you to perform polynomial operations on ciphertexts [92].

Such a set of operations allows you to implement a large number of AI algorithms, which was quickly discovered by many AI researchers. Let’s consider the methods they developed.

CryptoNets [93]. In this solution, the authors propose the use of GS in the operation of confidential cloud storage with the possibility of processing by neural networks. The main result achieved is the accuracy of the neural network for pattern recognition, equal to 99%.

In the case of SEALion [94], a solution based on TensorFlow [95] and SEAL [96] is proposed. TensorFlow implements calculations in tensors, SEAL itself GS. With the help of the developed solution, the authors implement convolutional neural networks (CNN), the method of support vectors is used as a learning method. The paper presents several NN models, the accuracy is determined based on the quality of digital image recognition. The authors show that their solutions are more effective than, for example, in [93]. A similar solution is TenSEAL [97], this library provides opportunities to work with ML, NN, and CNN in conjunction with GS. As a disadvantage, difficulties can be identified when learning from encrypted data, while the trained NN demonstrates performance and gives accurate results.

In [98], variations of confidential CNNs with GS are considered. The authors propose a CNN, which uses the integer scheme FHE – BGV [99]. The main focus is on the possibility of calculating the value of the ReLU activation function from ciphertexts obtained according to the BGV scheme. The authors defined three requirements for NN – accuracy, confidentiality and efficiency, the result of the work is a polynomial approximate ReLU function, which allowed to reduce the multiplicative depth, thereby increasing the efficiency of the network. This work can be considered quite important for several reasons: firstly, the very fact of the possibility of completely confidential calculations is shown; secondly, the possibility of building and using neural networks with FHE is shown.

In [100], the confidentiality-preserving DL is considered. The authors contrast their research with many of the works that were reviewed earlier, focusing on eliminating the shortcomings that were voiced above. The main result is based on the application of approximation methods for various activation functions, for example, such as ReLU and Softmax. To provide a large number of homomorphic multiplications in the CKKS scheme, the authors propose their modification of the bootstrapping procedure, which is based on the fact that the CKKS scheme [92] uses a system of residual classes to speed up arithmetic calculations. This result is interesting because it increases the speed of GS methods for specific tasks, whereas increasing the speed and applicability of GS in general is the main focus of research in this area at present.

GS allows you to solve the security problems of CC and CT in general by creating a transparent environment for free and confidential computing in the clouds. The analysis of GS methods in AI allows us to state that many researchers hold this opinion. The above FHE schemes are considered, which are more based on cryptographic lattices, however, there are FHE schemes that are built on SSS. This approach allows you to expand the capabilities of FHE in CT. Next, hybrid models that combine FHE and SSS methods are considered.

3.4 Hybrid Systems of Homomorphic Encryption and Secret Sharing Schemes

Considering the results in the direction of increasing AI privacy obtained through the use of SSS and FHE, the idea of creating a hybrid FHE-SSS system looks promising. Such a hybrid system will enhance the security of SSS by processing information in encrypted form using FHE in the space of a single local node. SSS-based security algorithms are used to transfer data between nodes. Thus, it becomes possible to adjust the balance, shifting the focus towards safety or productivity.

In [101], a solution was proposed that uses SSS and FHE to implement a confidentiality-preserving NN. The authors describe various algorithms used in this system, for example, security algorithms, key generation, SSS operation, etc., and also analyze in detail how NN works with this approach to security. The main focus is on demonstrating the effectiveness, safety, accuracy of the NN obtained and its comparison with other solutions, however, the data obtained are theoretical. The considered work is remarkable in that, despite the lack of specifics in terms of the FHE and SSS used, it shows the very possibility of implementing such a system and describes its characteristics, encouraging other researchers to engage in developments in this field.

4 ANALYSIS OF THE RESULTS OBTAINED

In the course of the study, three groups of methods for ensuring the confidentiality of artificial intelligence were analyzed (Table 2). The data presented below are based on information provided in the reviewed works.

Table 2. Results of analytical review of methods

Analyzing the results obtained (Table 2), we can summarize the following.

Differential confidentiality methods are primarily based on changes in the training sample. It is proposed to add unnecessary noise to the data, due to which the degree of confidentiality increases, while at the same time seriously reducing the efficiency of calculations. In lines 1 and 2, the values of computational complexity are highlighted in bold. In the first case, ε – means additional noise. In the second case, δ is the number of asynchronous nodes. Taking into account the level of confidentiality achieved in this case, additional security measures are required when transferring data. During processing, the degree of confidentiality is average – in order to steal data, an attacker will either need to filter the information from noise, or take control of several nodes in order to track the data flow. In addition, the need to apply corrective codes to avoid collisions or data loss will have an additional impact on the effectiveness of the system.

In the case of SSS, RNS-based circuits can improve the reliability of the system by using the self-correcting properties of RNS. However, confidentiality during data processing must be ensured by additional encryption methods. Also, in this case, to intercept data, an attacker will need to take control of the threshold number of nodes. When building SSS, the threshold for secret recovery is calculated so that the time spent compromising the shares of the secret is greater than the time of data relevance.

FHE allows you to ensure complete confidentiality. However, it has a significant drawback. Despite the fact that the computational complexity of encryption is low, compared with other algorithms, the computational complexity of data processing is much higher, which the authors of the schemes do not hide.

The last of the categories considered were hybrid SSS-FHE methods. At the moment, they are represented only by theoretical models, however, analyzing them, it is possible to assess the possible characteristics of the system.

The conducted research has shown that different research groups are interested in developing an AI with a high level of confidentiality. However, at the moment there are no relevant methods to achieve this level. In the future, it is planned to study the most promising methods based on SSS-FHE from the point of view of ensuring confidentiality. In particular, it is planned to develop NN using the FHE CKKS scheme, as well as Asmut-Bloom SSS. Justifying the choice of these algorithms, it can be noted that among the many FHE schemes, it is CKKS that is of the greatest interest to researchers, and among the many SSS based on RNS, it is Asmut-Bloom SSS that has the best characteristics in terms of security. The use of RNS in CKKS will increase the efficiency of the solution, and in Asmut-Bloom’s SSS – the reliability of data processing. An important part of the study will be to determine the accuracy of the calculation results, due to the integer nature of RNS, as well as the approximation error in CKKS.

5 CONCLUSIONS

In this paper, the methods of constructing a confidentiality-preserving AI in CC were investigated. At the first stage of the study, an analytical review of both AI and CC methods was conducted. Based on the results of the review, the criteria for AI safety in CC were determined. Next, the second stage of the analytical review was performed, namely, a review of AI privacy methods, on the basis of which 4 groups of methods were identified:

\( \bullet \) differential privacy;

\( \bullet \) secret sharing schemes;

\( \bullet \) homomorphic encryption;

\( \bullet \) hybrid methods based on SSS-FHE.

Based on the results of the study, the positive and negative sides of the considered methods were identified, an idea of the current state of the problem of ensuring AI confidentiality in CC was formed, and approaches to its solution were outlined. An analytical review has shown that there is currently no relevant solution. Solutions that provide the highest level of confidentiality have low efficiency due to the need to perform complex calculations. FHE supports addition and multiplication operations on encrypted values, such operations as determining the sign of a number, division, and matrix operations are not fully implemented. SSS does not allow you to process information in encrypted form. Confidentiality is ensured by not disclosing the source of a part of the data set of one participant to others. Differential confidentiality ensures anonymity during the operation of the neural network, while it has no protection against data interception. A hybrid method based on SSS-FHE is theoretically established as a possible solution, but it requires detailed research.

In future works, SSS-FHE research will be carried out, namely, the development of a prototype and the study of its characteristics.