Keywords

1 Introduction

Intelligent manufacturing (IM), which has been driven by Information and communication technologies (ICT), greatly improves the automation level, production quality and efficiency of manufacturing industry. These information and communication technologies (ICTs) provide reduced Capital Expenditure (CAPEX) and Operating Expense (OPEX) with higher efficiency and effectiveness. However, owing to the inherent limitations and complex network protocols in traditional industrial networks, traditional industrial Ethernets cannot manage distributed resources flexibly. Moreover, collected raw data are becoming increasingly more granular and voluminous.To address these drawbacks, bringing the cloud computing resources nearer to the underlying networks is attractive and promising.

In [1], the authors proposed a low latency mobile edge computing (MEC) framework based on the SDN architecture. Security and privacy are two of the main challenges to the IoT; hence, [2] proposed solutions and models for securing IoT devices and communications using the SDN architecture. Meanwhile, the performance of SDIN using in data offloading and edge computing in several scenarios including cloud server and mobile tasks was discussed in [3,4,5]. Computation offloading technology as discussed in [6, 7]. In [8] and [9], mobile devices can extend the standby time by computation offloading.

Other previous research addressed special purposes such as energy saving and real-time communication; however, they did not focus on the details of the entire system architecture and operation. Therefore, we present a new software defined industrial network (SDIN) architecture that is the combination of Software Defined Networking (SDN) and IIoT. Industry network intelligence and control are logically centralized in SDIN to provide greater processing performance and avoid the majority of the aforementioned drawbacks. Particularly, our contributions are as follows:

  • We propose a new SDIN architecture, and analyze the control and management process.

  • Based on the SDIN architecture and the characteristics of industrial computing tasks, we propose a hybrid centralized edge computing offloading strategy.

The remainder of this paper is organized as follows. Section 2 identifies the architecture of SDIN. Section 3 analyses the application of SDIN in solving the computing offloading problem and the unique features. The system model, problem formulation, and solutions are provided in Sect. 4. We present a performance simulation in Sect. 5. Finally, Sect. 6 concludes the work.

Fig. 1.
figure 1

Proposed software defined industrial network architecture

2 Software Defined Industrial Network

We propose an SDIN architecture as displayed in Fig. 1. The SDIN architecture contains four layers: field devices layer, data transport layer, distributed control layer and cloud platform. Field devices include the basic infrastructures such as robot arms, conveyor belts, lathes and deployed sensors, etc.

(1) Data Transport Layer

This layer is composed of SDN switches, wireless access points (APs). The APs emphasize authentication where IoT devices access the network and for data transporting. Data plane devices receive commands from control plane through southbound interface, such as computing decisions.

(2) Distributed Control Layer

The distributed SDN controllers are responsible for the centralized management of the edge computation servers and the authorization of multiuser access. Controllers receive requests from devices in the data transport plane and execute the offloading decision algorithm considering both the mission requirements and status of the edge computation servers. As displayed in Fig. 1, two-tier heterogeneous controller structure (domain controller and super controller) is one of the typical deployment solutions in large scale manufacture enterprise.

(3) Cloud Platform

The cloud platform includes a series of cloud service applications composed of industry application systems (Such as MES (Manufacturing Execution System), EPR (Enterprise Resources Planning)). The super controller places emphasis on resource management and the authorization of the distributed SDN controllers.

Besides, edge computing servers are more powerful than the local computing nodes in field devices. Due to the edge servers are located near the factories and production lines, it can provide lower and more stable latency than cloud computing server.

3 Computing Task and Offload in Industrial Scenario

In the concept of the intelligent factory, there are many computing tasks during the production process, such as Automatic Guided Vehicle (AGV) navigation, operation control of mechanical arms, and the identification of product imperfection. These services and applications can require significant computation resources and constrained time delay. However, the computational capabilities of the field devices are limited owing cost and size limitations.

3.1 Industrial Computing Task

After investigation, the industrial computing task has following unique features:

(1) Stringent computing delay tolerant

In industrial internet, the distributed sensors, actuators, machines, and other computing devices need to collaborate together to achieve real-time operation or complete the production tasks. In order to minimize the influence to the production line, the latency-sensitive industrial applications require delay from tens of milliseconds to hundreds of milliseconds.

(2) Diverse computing factors

As mentioned before, typical computing tasks in modern factory could be classified into following types with diverse characteristics and QoS requirements.

  • Image or video recognition (such as quality inspection).

    Characteristic: huge amounts of raw input data, small size of computing result data.

  • Localization and mapping (such as welding robot positioning guide).

    Characteristic: high computing accuracy, huge computing resource.

  • Production planning and scheduling.

    Characteristic: multiple data sources, complex computing, low frequency.

To study the effects of the computation task characteristics on the design of offloading schemes, we classify the tasks into I types. Each type of computing task has various QoS requirements. According to computation task types, the field devices can be correspondingly classified into I types. Let us assume that one field device only generates one type of computing task.

(3) Regular task pattern

Generally, the computing tasks originate from the production line which has a fixed production cycle. For instance, the cycle of one product line is five products per minute, therefore the cycle of product imperfection identify task is also five times per minute.

To simplify the analysis, the arrival of single industrial computing task is modeled as regular arrival. However, considering the asynchronism of different field devices and production lines, the arrival of tasks at the edge server could be treated as poisson flow.

3.2 Computing Offload Procedure Based on the SDIN Architecture

The decision of whether local computing or offloading the computing task to the edge server is an important and difficult procedure. Figure 2 displays the computing decision process. The normal working process includes two stages: maintenance and update. In the maintenance stage, domain controllers broadcast the domain offloading strategy (the offloading probability of each computing task type, denoted as \(\xi _{i}\), \(i \in I=[1,2,...I]\)) to field devices (\(\textcircled {1}\) in Fig. 2). Field devices generate a random number between 0 and 1. If this number is less than \(\xi _{i}\), then the computing will be offloaded to the edge server. Otherwise, the computing will be implemented in local computing unit. Update stage is trigged by the change of manufacturing environment and other factors, such as the increasing of computing task frequency, adjust of production scheduling. It contains three parts: requests collection, mode decision, computing, as follows:

Fig. 2.
figure 2

The computing offloading decision process

(1) Update Request

The devices collect update data (information of computing tasks) and send it to the data transport plane through the APs. Then the update request message is send to its domain SDIN controller by southbound interface (\(\textcircled {3}\) in Fig. 2).

(2) Mode Decision

The domain controller exacts the computing capability parameters from the latest update request message and edge computing server (from the load report message, as \(\textcircled {2}\) in Fig. 2), then it will execute the decision algorithm described in Sect. 4 and return the domain offloading strategy (the offloading probability of each computing task type, \(\xi _{i}\)).

(3) Computing

Similar with the maintenance stage, field devices decide whether offloading the task to the edge server or not after received the domain offloading strategy. For instance, if the device chooses the remote computing, field device will upload the necessary data and code to the edge server (\(\textcircled {4}\) in Fig. 2), and receive the computing result from the server afterward (\(\textcircled {5}\) in Fig. 2).

4 System Model and Offloading Algorithm

Considering most of computing devices in current factory are powered with electricity instead of batteries, the energy consumption of computing and data transfer is not a great issue. Therefore, the computing offload policy only considers the goal of minimizing computing delay.

The computing latency is divided into the following five aspects for decision analysis: local computing delay \(D_{Local}\), data transmission delay from the local to the edge server \(T_{Trans}\), task queuing delay for the edge computing server \(D_{queue}\), edge computing server calculation delay \(D_{Remote}\), and computing result return delay from the edge server to the local devices \(D_{Result}\).

We denote the computing tasks of type i by \(I_{i} = (D_{i},C_{i},T_{i})\), where \(D_{i}\) denotes data size, \(C_{i}\) represents the size of computation data involved in the number of CPU cycles required to complete the type-i task, and \(T_{i}\) represents the maximum delay tolerance of the type-i task.

According to their computation task types, the proportion of the field devices with type-i tasks is given by \(\pi _{i}\), where \(i \in I\), and \(\sum _{i \in I}^{}\) \(\pi _{i} = 1\).

4.1 System Model

The local computing unit capability is defined as \(f^{l}_{i}\), and \(f^{r}_{i}\) represents the CPU computation cycles per second that the edge server can provide. We assume that the transmission bandwidth is not constrained. Infinite buffer exists in the edge servers, and at most one computing task is served by the edge server simultaneously. Then the latency components could be calculated as (1)–(4).

$$\begin{aligned} D_{Local}(i) = \frac{C_i}{f_i^l} \end{aligned}$$
(1)
$$\begin{aligned} D_{Remote}(i) = \frac{C_i}{f_r^l} \end{aligned}$$
(2)
$$\begin{aligned} T_{Trans}(i) = \frac{D_i}{r_i} \end{aligned}$$
(3)
$$\begin{aligned} D_{Result}(i) = \frac{D_i^r}{r_i} = \frac{K_{i}C_{i}}{r_{i}} \end{aligned}$$
(4)

where \(r_{i}\) represents the transmission rate between the local node and the edge server node (either uplink or downlink), the unit is Kbps; \(D_{i}^{r}\) represents the size of computation result for type-i task. In general, its size is proportional to the amount of computation data \(C_{i}\), the proportion factor is a constant \(K_{i}\).

The queue delay \(D_{queue}\) is estimated by multiple class M/D/1 queue theory. Here we consider two typical cases: FIFO with equal priority (EP in short) and Non-preemptive priority queue (NPP in short). Denoted the arrival rate of all type-i tasks by \(\lambda _{i}\), which is usually a known parameter at the domain controller. Therefore, the real arrival rate of type-i tasks at the edge server is \(\xi _{i}\lambda _{i}\).

The mean service time of type-i \(E(S_{i})\) is deterministic, and equals to \(D_{queue}(i)\). We define the probability that the server is busy and busy with a type-i task as \(\rho \) and \(\rho _{i}\) respectively. Obviously

$$\begin{aligned} \rho _{i} = \xi _{i}\lambda _{i}E(S_{i}) = \frac{\xi _{i}\lambda _{i}C_{i}}{f_{i}^{r}} \end{aligned}$$
(5)
$$\begin{aligned} \rho = \sum _{i \in I} \rho _{i} = \sum _{i \in I} \frac{\xi _{i}\lambda _{i}C_{i}}{f_{i}^{r}} \end{aligned}$$
(6)
$$\begin{aligned} E(S) = \sum _{i \in I} \frac{\xi _{i}\lambda _{i}}{\sum _{i \in I} \xi _{i}\lambda _{i}} E(S_{i}) = \sum _{i \in I} \frac{\xi _{i}\lambda _{i}}{\sum _{i \in I} \xi _{i}\lambda _{i}} \frac{C_{i}}{f_{i}^{r}} \end{aligned}$$
(7)

Case A: Equal Priority (EP):

Based on Little theory and PASTA (Poisson arrivals see time averages) property, the average queueing latency is equal to the average queueing latency of each class, which could be estimated by

$$\begin{aligned} E(D_{queue}) = \frac{\sum _{i \in I} \rho _{i} \frac{E(S_{i})}{2}}{1-\rho } = \frac{\sum _{i \in I} \frac{\xi _{i}\lambda _{i}(C_{i})^{2}}{2(f_{i}^{r})^{2}}}{1-\sum _{i \in I} \frac{\xi _{i}\lambda _{i}C_{i}}{f_{i}^{r}}} \end{aligned}$$
(8)

Case B: Non-preemptive priority (NPP)

If type 1 has non-preemptive priority over type 2, then a type 2 task cannot be preempted once it enters service. Type 1 task still have priority over any type 2 task that are waiting but not being served. Let us assume that if \(i < j\), then type i has non-preemptive priority over type j.

For type 1 task,

$$\begin{aligned} E(D_{Queue(1)}) = \frac{ \sum _{i \in I}\rho _i \frac{E(S_i)}{2}}{1-\rho _1} = \frac{\sum _{i \in I}\frac{\xi _i \lambda _i (C_i)^2}{2*(f_i^r)^2} }{1-\frac{\xi _1 \lambda _1 C_1}{f_1^r}} \end{aligned}$$
(9)

For type \(i >1\), again using little and PASTA,

$$\begin{aligned} \begin{aligned} E(D_{Queue}(i))&= \frac{ \sum _{i \in I}\rho _i \frac{E(S_i)}{2}}{(1-\sum _{k=1}^{i-1}\rho _k)(1-\sum _{k=1}^{i}\rho _k)} \\&= \frac{ \sum _{i \in I}\rho _i \frac{E(S_i)}{2}}{(1-\sum _{k=1}^{i-1}\frac{\xi _k \lambda _k C_k}{f_k^r})(1-\sum _{k=1}^{i} \frac{\xi _k \lambda _k C_k}{f_k^r})} \end{aligned} \end{aligned}$$
(10)

Therefore, the average computing latency for each type of task could be obtained as:

$$\begin{aligned} \begin{aligned} E(L_i) = (1-\xi _i)&D_{Local}(i) + \\&\xi _i(D_{Remote}(i)+T_{Data}(i)+E(D_{Queue}(i))+T_{Result}(i)) \end{aligned} \end{aligned}$$
(11)

4.2 Offloading Policy

The optimization goal is minimizing the total computing latency of all computing units, including the edge offloading units and local computing units under the constraints of allowed maximum delay tolerant of each unit \(T_{i}\), the optimization problem is mathematically modeled as:

$$\begin{aligned} \begin{aligned}&\underset{\xi _i}{min}\sum _{i \in I}\pi _i E(L_i) \\&\texttt {s}.\texttt {t}. E(L_{i}) \le T_{i}, 0 \le \xi _{i} \le 1, i \in I \end{aligned} \end{aligned}$$
(12)

It is known from [8] that the optimization problem above is convex optimization. Thus, one can use the block coordinate descent (BCD) approach to deal with it as in the following iterative algorithm.

figure a

5 Simulation and Result Analysis

In this section, we use MATLAB simulation to evaluate the performance of proposed edge computation offload scheme. The computation tasks of field devices are classified into four types with the probabilities \(\pi :\{ 0.1,0.3,0.4,0.2 \}\).

The incoming computing flow of each type obeys poisson distribution of parameter \(\lambda \). Other parameters are listed in Table 1. In addition, the computational capability \(f_{i}^{l} = 2\) GHz, \(f_{i}^{r} = 10\) GHz, \(r_{i} = 20\) Mbps, \(K_{i} = 10^{-5}\). The link bandwidth bottleneck and transmission error are ignored.

Table 1. Parameters of various computation tasks

Figure 3 evaluates the percentage of various types of the tasks that are off-loaded under different computing task density \(\lambda \) in equal priority case. Because the tasks of higher priority are more sensitive to the delay constraints, the equal priority case cannot improve the probability of processing for high-priority tasks. We found that the offloading percentage of the type-1 and the type-4 tasks is nearly \(0\%\) and \(100\%\) in all the \(\lambda \) values, respectively. With the increase of task density, the offloading percentage decrease due to higher queuing latency at the edge server.

Fig. 3.
figure 3

EP offloading percentage

Fig. 4.
figure 4

NPP offloading percentage

Figure 4 shows the offload probability for each type of tasks under different computing task density \(\lambda \) in non-preemptive priority case. Compared with EP case, type-2 tasks will increase the offload probability slightly due to higher priority in high load region. Priority is given to high-priority tasks, which have less impact on the delay of subsequent tasks and easier to be flexibly choreographed.So Non-preemptive priority case can improve the service rate of high-priority tasks and further reduce average delay overall service.

Fig. 5.
figure 5

Average delay

Fig. 6.
figure 6

Outage probability

Figure 5 depicts the average delay of proposed central offloading scheme, All-local computing and All-remote computing scheme in EP case. Figure 6 shows the probability of outage (The probability that the computing latency larger than the maximal allowed latency) of those schemes in EP case. The delay and outage probability prove the feasibility of proposed offloading scheme. With the growing of computing load, the performance of All-remote computing solution will grow worse due to the increasing queuing delay. The offloading scheme proposed in the paper can greatly reduce the computing latency and improve the computing QoS for different users.

6 Conclusion

In this paper, we propose a new SDIN architecture and a kind of centralized computing offloading strategies based on our SDIN architecture. The simulation results have indicated that the proposed SDIN architecture is feasible and effective in computing offloading. And to a certain extent, our architecture can provide traditional industries a better resource management solution and more flexible production scheme which means the production efficiency can possibly be improved.