1 Introduction

The Wireless Sensor Network (WSN) paradigm is an emerging and fast growing technological platform in myriad working environments. To name just a few, WSNs can be used in military and agriculture applications, harsh environment monitoring, emergency search-and-rescue operations. One of the important characteristics of WSNs is their ability to densely deploy in ad-hoc manner. For successful operation of WSNs, sensors should effectively cover the required region of interest for a long period of time. Energy-aware sensing and communication mechanisms have been substantially pursued by the research community in order to prolong the lifespan of WSNs. Energy saving techniques can generally be classified in the following categories; (1) Energy-efficient data aggregation, gathering and routing; (2) power management by adjusting the transmission and/or sensing range of sensor nodes; and (3) sensor wake-up scheduling to alternate between active and idle state.

In this paper, we consider the third approach to achieve energy-efficient WSN that monitors all the targets at all times. In this approach, sensors are divided into disjoint sets of covers (DSC) such that every set completely cover a set of targets being located with known locations. We assume that the lifetime of a WSN is divided into intervals and at each interval only one set is active with while the remaining set covers are in the low-energy sleep mode. The sensors from the active set are in active state, monitoring all targets within their sensing ranges. Once the active set cover runs out of energy, another set cover will be selected to enter the active mode and provide the functionality continuously. As all targets should be monitored by every set cover, the goal of this approach is to determine a maximum number of DSC. The more set covers we can find, the longer sensor network lifetime is prolonged. It has been proven that this problem (also called the SET \(k\)-cover problem) is a generalization of the minimum cover problem [1] and proves its NP-completeness [2, 3].

Many attempts in the literature have been made to solve DSC problem in WSNs using either heuristic or meta-heuristic methods (like genetic algorithms). Different scheduling rules determine when sensors change to be active or sleep mode. In localized and distributed realizations, sensors periodically investigate their neighborhood and decide whether to change their operation modes [411].

In [2], a heuristic approach called the “most constrained-minimally constraining covering (MCMCC)” is proposed to select and successively activate mutually exclusive sets of covers, where every set completely covers the entire monitored area. This method gives priority to sensors which cover a high number of uncovered fields, cover sparsely covered fields and do not cover fields redundantly. This method achieves energy savings by increasing the number of disjoint covers. Also, the DSC problem and its extension to the maximum set cover (MSC) problem have been solved in [12] and [3] using integer programming, linear programming, and greedy heuristics. In [12], the DSC problem is reduced to a maximum flow problem and solved using mixed integer programming. By a branch and bound method, the maximum covers based on mixed integer programming algorithm (MC-MIP) acts as an implicit exhaustive search to guarantees finding the optimal solution.

The definition of DSC problem has been re-formulated in [3, 13], and [14] to include additional coverage constraints. The definition of DSC problem has been generalized in [3] to a maximum set covers (MSC) problem and solved it using linear programming and greedy techniques. The extended problem in MSC lets the sensors to participate in multiple sets. In [13], the DSC problem has been extended to include connectivity constraint as well. Then, the Connected Set Covers (CSC) problem has as objective finding a maximum number of set covers such that each sensor to be activated should be connected to the base station. In [14], DSC problem has been extended to include sensor coverage-failure probability. Each sensor is associated with sensor’s failure probability (comes from several facts, e.g., manufacture, weather in the monitoring area, interferences to the sensors, or unexpected accidents). The proposed Maximum Reliability Sensor Covers (MRSC) problem has been solved in [14] using a heuristic greedy algorithm to compute the maximal number of set covers that satisfy a user specified coverage-reliability threshold.

The works in [1517] also provide solutions to the DSC problem in WSNs by employing the meta-heuristic framework of evolutionary and genetic algorithms. Like the previous mentioned heuristic methods, the single-objective genetic algorithms proposed in [1517] assume simple and common isotropic (i.e., disc) sensing model. Each sensor in this model is associated with a sensing area which is represented by a circle with the same radius. According to [18], the upper bound, \(ub\), of the maximum number of disjoint complete set covers depends on the size of the target area, the total number of sensors, the sensors’ locations, and their sensing ranges. \(ub\) can be estimated as the minimum number of sensors covering the most sparsely covered target, and thus can be computed by polynomial-time algorithms. It has been shown that for the same deployment of sensors and targets, \(ub\) can be increased by enlarging the sensing radius of each sensor [1518]. In reality, however, the sensing region of a sensor could be irregular, resulting in imprecise target coverage. The coverage in this case could be expressed in probabilistic terms [1921]. In probabilistic sensing model, there is a measure of uncertainty in sensor signal-detection being expressed by a value from 0 to 1. For reliable coverage with threshold \(c_{th}\), the detection uncertainty of each target should not exceed \(1-c_{th}\). This in turn should effect on the number of sensors reliably covering the most sparsely covered target. As more reliability degree to cover a target is required, \(ub\) becomes smaller. Thus, there should be a trade-off between the coverage reliability (also denoted throughout this paper as covers reliability) and \(ub\).

Like many other sensor network design problems, maintaining reliable coverage while maximizing the number of DSC (i.e., extending network lifetime) often require optimizing two conflicting decision variables. Due to the conflicting nature of many sensor network design problems, multi-objective optimization (MOO) has, recently, attracted several researchers in formulating multi-objective optimization problems (MOPs). In MOO, a set of alternative solutions (called non-dominated solutions) can simultaneously be obtained providing the decision maker with an optimal tradeoff between the conflicting objectives. As multi-objective evolutionary algorithms (MOEAs) possess several interesting characteristics for tackling MOPs, there are many MOEA approaches for solving different sensor design problems. For organizing energy efficient WSNs, the literature provides a collection of MOO schemes in which sensing coverage and network lifetime are the main design issues to be optimized. However, all of these methods fall in one of the first two categories; data aggregation, routing, and sensor clustering [2225], or sensors’ power management [2629]. For the reliable DSC design problem in WSNs, our work in [30] formulates a single objective evolutionary algorithm to tackle it. To the best of our knowledge, however, MOO concept has not been exploited. The contributions of this paper are as follows:

  1. 1.

    The paper turns the de-facto definition of single-objective DSC in WSN design problems to a multi-objective DSC. The goal of which becomes to cope with contradictory objectives. The formulated optimization problem will be directed to effectively maintaining reliable coverage with large number of cover sets in WSNs.

  2. 2.

    The paper presents two MOEAs with a new heuristic perturbation operator (instead of the traditional bi-perturbation of crossover and mutation operators) to solve the modeled problem. The results show that the Pareto-optimization can provide the decision maker with a set of non-dominated solutions that effectively render the tradeoff between the two conflicting objectives (i.e., network lifetime and coverage probability).

The rest of this paper is outlined as follows. Section 2 describes the traditional formulation of DSC problem in WSNs as well as the new formulation of multi-objective DSC by adding probabilistic coverage means to the original problem. Section 3 provides the preliminaries to the concept of multi-objective optimization and introduces the formulation of the proposed perturbation operator. The results of multi-objective DSC optimizations are then compared with a related work of single-objective genetic algorithm in Sect. 4. Finally, Sect. 5 concludes the current work and hints some further ramifications.

2 Problem Definition

2.1 System Model and Assumptions

In order to model the system, we assume that the investigated WSNs have 2D sensing area \({\mathcal A}\) with known \(\hbox {size}(X_{max} ,Y_{max} )\). We also assume that \({\mathcal A}\) has a set \({\mathcal T}\) (i.e., target set) of \(n\) targets with known locations, i.e., \({\mathcal T}=\{\left( {x_{t1} ,y_{t1} } \right) ,\left( {x_{t2} ,y_{t2} } \right) ,\ldots ,(x_{tn} ,y_{tn} )\}\). There are \(m\) homogenous sensors \({\mathcal S}=\{\left( {x_{s1} ,y_{s1} } \right) ,\left( {x_{s2} ,y_{s2} } \right) ,\ldots ,(x_{sm} ,y_{sm} )\}\) having the same sensing range \(r_s \). All the sensors are dropped randomly in \({\mathcal A} \quad (1\le \forall i\le m|\left( {x_{si} ,y_{si} } \right) =([0,X_{max} ],[0,Y_{max} ]))\). Depending on the sensing range \(r_s \), each sensor is responsible for sensing and covering a part of \({\mathcal A}\). We consider a probabilistic sensing model [20, 21] to define the notion of the probabilistic coverage of a target \({\mathcal T}_j =\left( {x_{tj} ,y_{tj} } \right) \) by a sensor \(s_i \).

$$\begin{aligned} Coverage\left( {s_i ,t_j } \right) =\left\{ {{\begin{array}{ll} 0&{} {\quad if \quad r_s +r_u \le d(s_i ,t_j )} \\ {e^{-\lambda a^{\beta }}}&{} {\quad if \quad r_s -r_u <d(s_i ,t_j )} \\ 1&{} {\quad if \quad r_s -r_u \ge d(s_i ,t_j )} \\ \end{array} }} \right. <r_s +r_u \end{aligned}$$
(1)

where \(r_{u}\) is a measure of the uncertainty in sensor detection. \(d\left( {s_i ,t_j } \right) \) is the Euclidean distance \(\sqrt{(x_{si} -x_{tj} )^{2}+(y_{si} -y_{tj} )^{2}}\) between sensor \(s_i \) and target \(t_{j}\,.\,a=d(s_i ,t_j )-(r_s -r_u )\), and \(\lambda \) and \(\beta \) are probabilistic detection parameters to measure detection strength when a target point lies within the interval \(\{r_s -r_u ,r_s +r_u \}\). It causes coverage value to exponentially decrease as the distance increase. All points that lie within a distance of \(r_s -r_u \) from the sensor are said to be 1-covered. Beyond the distance \(r_s +r_u \), all the points have 0-coverage by this sensor.

To save energy and prolong WSN’s lifetime, sensors in the sensor set \({\mathcal S}\) should be divided into duty-cycling sensor cover (also called set cover) subsets, each of which can cover all the interested targets in \({\mathcal T}\). Thus, in the traditional Boolean sensing model, the definition of the sensor cover could be formulated as:

Definition 1

(Sensor Cover). Given a WSN consists of target set \({\mathcal T}\) and sensor set \({\mathcal S}\), where each sensor \(s_i \in S\) can be represented as a subset \({\mathcal T}_i \subset {\mathcal T}\), such that \(tj\in {\mathcal T}_i \) if and only if \(Coverage\left( {s_i ,t_j } \right) =1\). Any subset \({\mathcal S}_i \subset {\mathcal S}\) that can completely cover all the target set \({\mathcal T}\) is termed as a sensor cover.

However, considering probabilistic sensing model, the definition of the traditional sensor cover needs to be re-formulated as:

Definition 2

(Reliable Sensor Cover). Given a WSN consists of target set \({\mathcal T}\) and sensor set \({\mathcal S}\), where each sensor \(s_i \in S\) can be represented as a subset \({\mathcal T}_i \subset {\mathcal T}\), such that \(tj\in {\mathcal T}_i \) if and only if \(Coverage\left( {s_i ,t_j } \right) \ge c_{th} \). Any subset \({\mathcal S}_i \subset {\mathcal S}\) that can satisfy a user coverage constraint \(c_{th} \) to cover all the targets in \({\mathcal T}\) is termed as a reliable sensor cover or reliable set cover.

Now, the problem of finding the maximum number of disjoint set covers (DSC) could be turned into the problem of finding the maximum number of disjoint reliable sensor covers (DRSC), and can be formulated as:

Definition 3

(Disjoint Reliable Sensor Covers Problem—DRSC). Given a collection \({\mathcal S}\) of reliable subsets of a finite set \({\mathcal T}\), find the maximum number of disjoint reliable covers for \({\mathcal T}\). Every cover \({\mathcal S}_i \) is a subset of \({\mathcal S}\), \({\mathcal S}_i \subseteq {\mathcal S}\), such that every element \(tj\) of \({\mathcal T}\) belongs to at least one member of \({\mathcal S}_i \), and for any two covers \({\mathcal S}_i \) and \({\mathcal S}_j , {\mathcal S}_i \mathop \cap \nolimits ^ {\mathcal S}_j =\emptyset \). Now, the upper bound \(ub\) of the maximum number of disjoint reliable covers is determined by the most sparsely covered target and can be calculated as:

$$\begin{aligned} ub={ }_{i=1,2,\ldots ,n}^{min} (\left| {{\mathcal S}_i } \right| ) \end{aligned}$$
(2)

where \({\mathcal C}{\mathcal S}={\mathcal S}_i ,\left| {{\mathcal S}_i } \right| =ub\) denotes the set of sensors (called Critical sensors Set) covering target \(ti\) with certainty equal to or larger than a user specified threshold \(c_{th} \). As the most sparsely covered target only covered by \(ub\). Sensors, the maximum number of disjoint reliable covers \(N\) is no larger than \(ub\) i.e., \(N\le ub)\).

2.2 Problem Formulation

In this section, we formulate the problem of getting high coverage percentage with maximum number of DSC as multi-objective Disjoint Reliable Sensor Covers Problem:

Given:

\({\mathcal A}\) :

: 2-D plane area with \(\hbox {size}(X_{max} ,Y_{max})\).

\({\mathcal T}\) :

: set of \(n\) targets being uniformly distributed in \({\mathcal A}\).

\({\mathcal S}\) :

: set of \(m\) sensors with probabilistic sensing capability, being uniformly distributed in \({\mathcal A}\).

\(r_s\) :

: sensing range.

\(r_u,\) :

\(\lambda \) and \(\beta \): uncertainty-parameters adjusted according to the physical properties of the sensor.

Decision (Design) variables of a solution:

The scheduling sets of sensors in \({\mathcal S}\), i.e., \(\left\{ {{\mathcal S}_1 ,{\mathcal S}_2 ,\ldots ,{\mathcal S}_m } \right\} |{\mathcal S}_i \in [1,ub]\).

Objective:

Two objectives are considered as optimization functions. The first objective is to maximize the number of scheduling sets of the sensors in \({\mathcal S}\), such that each scheduling set \({\mathcal S}_i \) can form a complete reliable sensor cover (see Eq. 3). Thus, Eq. 3 computes the number of reliable covers (i.e., without and coverage holes) being formed by a particular solution. Detection of coverage holes is formulated in Eq. 4. The second objective is to maximize the reliability of the whole disjoint sensor covers (see Eq. 5). In each sensor cover, we consider only the highest coverage of the sensors belong to this cover to each target in set \({\mathcal T}\).

$$\begin{aligned} max\,DRSC=\mathop \sum \nolimits _{{\mathcal S}_i \in [1,ub]} Cover({\mathcal S}_i ,{\mathcal T}) \end{aligned}$$
(3)

where:

$$\begin{aligned} Cover\left( {{\mathcal S}_i ,{\mathcal T}} \right) =\left\{ \begin{array}{l} {1 \quad if\quad \forall t\in {\mathcal T}\rightarrow \exists s\in {\mathcal S}_i |Coverage\left( {s,t} \right) \ge c_{th} } \\ {0 \quad otherwise} \\ \end{array} \right. \end{aligned}$$
(4)
$$\begin{aligned} max\,Reliability=\mathop \sum \limits _{{\mathcal S}_{i|Cover\left( {{\mathcal S}_i ,{\mathcal T}} \right) =1} } Coverage({\mathcal S}_i ,{\mathcal T}) \end{aligned}$$
(5)

where:

$$\begin{aligned} Coverage\left( {{\mathcal S}_i ,{\mathcal T}} \right) =\frac{\mathop \sum \nolimits _{t\in {\mathcal T}} { }_{s\in {\mathcal S}_i }^{max} Coverage\left( {s,t} \right) }{n} \end{aligned}$$
(6)

3 Multi-objective Evolutionary Algorithms for DRSC Problem

3.1 Preliminaries

The general formalization of MOP is defined as [31, 32]: finding a vector \({\mathbb {X}}^{*}=[x_{1}^{*} ,x_{2}^{*} ,\ldots ,x_{n}^*]^{T}\) optimizing (in terms of domination) the vector function \({\mathbb {F}}({\mathbb {X}})=[{\mathbb {f}}_1 \left( {\mathbb {X}} \right) ,\mathbb {f}_2 \left( {\mathbb {X}}\right) ,\ldots ,\mathbb {f}_k \left( {\mathbb {X}} \right) ]^{T}\) where \({\mathbb {X}}=[x_1^ ,x_2,\ldots ,x_n]^{T}\) is the decision variables vector. Optimizing \({\mathbb {F}}\left( {\mathbb {X}} \right) \), here, considers the domination concept. Let us consider two solution \({\mathbb {U}}\) and \({\mathbb {V}}\) from the solution space \(P\). Then, solution \({\mathbb {U}}\) is said to dominate \({\mathbb {V}}\) if and only if the following two conditions hold:

  1. 1.

    The solution \({\mathbb {U}}\) is no worse than \({\mathbb {V}}\) in all objectives, or \(\mathbb {f}_i \left( {\mathbb {U}} \right) \ntriangleright \mathbb {f}_i \left( {\mathbb {V}} \right) \) for all \(i=1,2,\ldots ,k\). For example in maximization, the word “no worse” means \(\mathbb {f}_i \left( {\mathbb {U}}\right) \nless \mathbb {f}_i \left( {\mathbb {V}} \right) \).

  2. 2.

    The solution \({\mathbb {U}}\) is strictly better than \({\mathbb {V}}\) in at least one objective, or \(\mathbb {f}_i \left( {\mathbb {U}} \right) \triangleleft \mathbb {f}_i \left( {\mathbb {V}} \right) \) for at least one \(i\in 1,2,\ldots ,k\). For example in maximization, the word “strictly better” means \(\mathbb {f}_i \left( {\mathbb {U}} \right) >\mathbb {f}_i \left( {\mathbb {V}} \right) \).

The notation \(\left( {i\triangleleft j} \right) \) is used to denote that solution \(i\) is better than solution \(j\) regardless of the type of the optimization problem at hand (maximization or minimization). Also, the notation \(\left( {i\triangleright j} \right) \) is used in the same way to express that solution \(j\) is better than solution \(i\). Hence, a dominated set can be defined as: among a set of solutions \(P\), the non-dominated solutions set are those that are not dominated by any member of the set \(P\).

In this paper, we consider two multi-objective evolutionary algorithms (MOEAs). These are: multi-objective evolutionary algorithm with Tchebycheff decomposition (MOEA/D) and non-dominated sorting genetic algorithm (NSGA-II). In what follows, we only present these two algorithms focusing on the proposed perturbation operator. We refer interested readers to [3135] for more detailed description of the general framework of MOEA/D and NSGA-II.

3.2 The Proposed Meta-heuristic Framework

As MOEA/D and NSGA-II are population-based optimization algorithms, let us consider a population \(\rho \) of \(K\) solutions. The choice of a good solution representation is a critical issue for the applicability and performance of evolutionary algorithm. Solution representation is highly problem dependent and related to the evolution operations. In our algorithm design, each individual solution \({\mathbb {P}}_{1\le i\le K} \in \rho \) is represented as a fixed-length vector of size \(m\), where each element (i.e., gene) value indicates the index of the set cover that the corresponding sensor joins, specifying that the index does not exceed the upper limit \(ub\). Thus, \(\forall i\in \left\{ {1,\ldots ,N} \right\} \) and \(\forall j\in \left[ {1,m} \right] :{\mathbb {P}}_i =\left( {{\mathbb {P}}_{i1} ,\mathbb {P}_{i2} ,\ldots ,{\mathbb {P}}_{im} } \right) s.t.{\mathbb {P}}_{ij} \in [1,ub]\). Then, each of the implemented MOOs can be described as a process formulated in an iterative evolution function \({\Psi }:\left\{ {\rho ,{\textit{EP}}} \right\} \rightarrow \{\rho ^{\prime },{\textit{EP}}^{\prime }\}\) with \({\Psi }\left( {\rho _i } \right) =\rho _{i+1} \), where \(i\) is the iteration index and \(\rho _i \) is the population in iteration \(i\). \({\textit{EP}}\) is an external archive for the generated non-dominated solutions. The population starts with an initial random population \(\rho _0 \) and continues until a maximum number of iterations \(max_i \) has been reached. Each iteration \(i\) consists of a three main operators: parents selection, heuristic perturbation, and \({\textit{EP}}\) update.

figure e

Based on the chromosome representation, we adopted an intuitive and fast random initialization mechanism, where each gene can take a random value that does not exceed the upper limit \(ub\). Formally speaking:

$$\begin{aligned} (\forall i\in \left\{ {1,\ldots ,N} \right\} ,\forall j\in \left[ {1,m} \right] ):{\mathbb {P}}_{ij} \sim [1,ub] \end{aligned}$$
(7)

where \(\sim [1,ub]\) means a uniform random integer number between \(1\) and \(ub\).

For selection operator, we imitate similar functions to that found in the implementation of the traditional MOEA/D and NSGA-II. For the perturbation function, we proposed one evolution operator \({\Psi }\) that provides heuristic for improving a solution coverage reliability or number of disjoint covers. The process of the proposed heuristic perturbation operator is presented next (see Algorithm 1). It takes as input two parent individuals \({\mathbb {P}}_1 \), \({\mathbb {P}}_2\) being selected from \(\rho \) by the selection operator and perturbs them with probability \(p_{per} =1.0\) to return one child \({\mathbb {C}}\) with its decoded set covers \(C_1 , C_2 , \ldots , C_k\). For each child, the proposed heuristic will be directed (with equal probability) either towards maximizing the reliability of the generated set covers or towards maximizing the number of sensor covers \(k\). Thus, for each child \({\mathbb {C}}_i ,1\le i\le K\), the heuristic procedure also takes as input a flag \(F\) indicating whether to prefer maximizing reliability or maximizing number of sensor covers. The proposed heuristic also takes as input the set of the critical sensors \({\mathcal C}{\mathcal S}\).

The proposed heuristic recursively builds sensor covers in the child. For each cover \(i\), the set \(S\) (in line 6) collects sensors from both parents that have an index value equal to \(i\). The sensors in \(S\) can then participate in creating the cover \(i\) in the child. When sensors in \(S\) can cover all the targets without any coverage hole (in line 8), then (and according to the flag \(F)\), the procedure will select from \(S\) those sensors having the highest contribution to the heuristic (line 12–18). In order to maximize number of sensor covers towards its upper limit \(ub\), a heuristic process is maintained in line 15. Note that this process is guided to restrict the choice of sensors, for the current sensor cover \(i\) , to those that do not belong to the set of critical sensors (i.e., \(s\notin {\mathcal C}{\mathcal S}\) ) or at least bias the choice in their support. The biasing depends on the number of targets being covered by each sensor \(s\notin {\mathcal C}{\mathcal S}\). However, if there is still coverage hole in the current cover \(i\), the heuristic process turns the selection to the critical sensors \(s\in {\mathcal C}{\mathcal S}\). The remaining sensors can then contribute in the next covers (line 20). They will migrate to the last set cover \(k\) and used again (line 23) when there is a coverage hole (line 21) while forming a new cover \(i\). Finally, the last cover (line 25) may form a complete set cover or incomplete one. The generated child \({\mathbb {C}}\) may modify the contents of the external archive \({\textit{EP}}\) according to the non-domination condition described previously.

4 Simulation Results and Discussion

In this section, we evaluate the performance of the proposed MOO algorithms in terms of number of set covers obtained and coverage reliability. The results are obtained after setting WSNs and algorithm parameters into the most commonly used in the related literature. The simulation area is square-shaped with side length \(X_{max}\) = 500 m of 10 randomly distributed targets. Unless otherwise stated, the simulation is mainly divided into five groups according to five different settings of sensor density: 50, 60, 70, 80, and 90. For each group, we vary the sensing range of the sensor nodes \(R_s \) to three different values \(\{100,200,300\}\), to get a different test instance. Each test instance includes 10 random WSNs with different configurations. Thus the overall simulation examines a total of \(150\) random networks. As we have random WSN configurations, we may get different value for the upper bound of set covers \(ub\), for the same parameter settings. Thus in each test instance, the presented results also indicate the average number of upper bound of set covers (denoted by \( \overline{ub}\)). Uncertainty level \(R_u \) is set to \(R_s *0.5\) units, both \(\lambda \) and \(\beta \) are set to \(0.5\), and \(c_{th} \) is set to \(0.001\). The setting of the probabilistic coverage parameters also influences the overall network’s coverage reliability. Since studying the impact of varying these parameters is out of the scope of this paper, we fixed these parameters to one setting and evaluate the average coverage reliability (denoted by \(\bar{r}\)) of all network configurations in each test instance. Population size is set to \(50\) and will be allowed to evolve \(500\) times. Perturbation probability is fixed to \(1.\)0. For MOEA/D, the size of neighboring solutions that participate in the evolution of each individual solution is fixed to \(5\).

Both MOEA/D and NSGA-II have been compared with Genetic Algorithm for Maximum Disjoint Set Covers (GAMDSC). GAMDSC has been proposed in [15] to extend the lifetime of WSNs. GAMDSC is single-objective genetic algorithm that use integer representation for chromosome representation, traditional uniform crossover, creep mutation, and a scatter operator to evolve their chromosome solutions. The proposed fitness function in GAMDSC has been devoted to maximize number of complete set covers in each chromosome. The performance of GAMDSC has been compared in [15] against the most constrained-minimum constraining heuristic (MCMCC) [2] and the exhaustive heuristic of Maximum Covers using Mixed Integer Programming (MCMIP) [12] for solving the DSC problem in WSNs. Simulation results in [15] show that GAMDSC can get near-optimal solutions and improve the performance of MCMCC by 16 % in terms of the obtained sensor covers. As compared with the exhaustive search of MCMIP, GAMDSC can get results with acceptable computation time.

4.1 Comparison of Multi-objective Versus Single-objective DSC

First, results are presented in Tables 1 and 2 to compare the quality of results obtained by GAMDSC, MOEA/D, and NSGA-II. Table 1 presents number of set covers being obtained by the three algorithms. Table 2 presents the coverage reliability being obtained by the algorithms. Also, to examine the quality of solutions provided by the three algorithms, we present \(\overline{ub}\) and \(\bar{r}\) as reference values in Tables 1 and 2, respectively.

Table 1 Comparison of the maximum number of the generated set covers for \(10\) WSNs in each test case
Table 2 Comparison of the maximum covers reliability provided for \(10\) WSNs in each test case

For both MOEA/D and NSGA-II, Tables 1 and 2 present the two extreme non-dominated solutions (in terms of number of set covers and reliability, respectively). The solution providing the closest value to \(\overline{ub}\) in each test instance in Table 1 has been presented in bold. Also, the solution providing the maximum reliability in each test instance in Table 2 has been bolded. Note that in Table 1, closest solutions to \(\overline{ub}\) do not necessary imply best solutions. In terms of the algorithm’s independent single-objective measure of solution’s quality, closest solutions to \(\overline{ub}\) should result in high quality solutions. That is GAMDSC provides more sensor covers in four test instances while NSGA-II provides more sensor covers in almost all of the remaining test instances. However, seeking for tradeoff solutions that balance between the sensor covers and coverage reliability (see Table 2), the performance of GAMDSC in these four test instances is not good enough. The coverage reliability of GAMDSC gives the lowest score while both MOEA/D and NSGA-II outperform GAMDSC in all test instances. Similarly, in the third and tenth test instances (referred with asterisk), GAMDSC gives equal values of sensor covers to NSGA-II or MOEA/D, respectively (as presented in Table 1), however, both NSGA-II MOEA/D outperform GAMDSC in terms of reliability (as presented in Table 2). Thus, the combined results of Tables 1 and 2 signify the importance of utilizing the two objectives as optimizing criteria rather than maximizing only the number of sensor covers. Moreover, in all test instances, Table 2 presents that NSGA-II performs better than MOEA/D even in the case of providing an equal number of set covers as in the sixth test instance, referred with asterisk in Table 1.

Figure 1 compares the performance of GAMDSC, MOEA/D, and NSGA-II in terms of average number of disjoint sensor covers obtained while fixing sensing radius. In this simulation, we fixed sensing range \(R_s \) to \(500\) and vary number of sensors from \(50\) to \(90\) with an incremental value of \(5\). Each test instance of this simulation includes \(10\) random networks, and thus the total simulation is experimented under test-bed of \(90\) different network topologies. This simulation clarifies the impact of varying sensing ranges, of a given fixed number and fixed locations of sensors, on the total number of disjoint sensor covers. The results of this simulation summarize that for a fixed sensing radius, the number of sensor covers increases with increasing number of sensors. Also, from the figure we can that NSGA-II provides more number of sensor covers than both MOEA/D and GAMDSC.

Fig. 1
figure 1

Average number of sensor covers for 90 networks with \(R_s =500\) and number of sensors \( =\{50,55,\ldots ,90\}\). In each group of bars: left, middle, and right bar corresponds to NSGA-II, MOEA/D, and GAMDSC results, respectively

On the other hand, Fig. 2 clarifies the impact of varying sensing range \(R_s \) while fixing number of sensors. In this simulation, the number of sensors is set to \(70\) and let \(R_s \) to vary from \(100\) to 500 with an incremental value of \(50\). Also, in this simulation, each test instance includes \(10\) random networks with a total of \(90\) different network topologies. The results of the figure indicates that increasing sensing radius have a positive impact on the number of sensor covers, even when fixing the number of sensors to a certain value. Also, we can see that NSGS-II provides more number of sensor covers than both MOEA/D and GAMDSC.

Fig. 2
figure 2

Average number of sensor covers for 90 networks with number of sensors = 70 and \(R_s =\{100,150,\ldots ,500\}\). In each group of bars: left, middle, and right bar corresponds to NSGA-II, MOEA/D, and GAMDSC results, respectively

4.2 Quality and Diversity of Pareto Front Solutions

Let us now turn to quantitatively compare the performance of MOEA/D and NSGA-II. Here, the comparison will be according to the quality of the non-dominated solutions set being obtained by each algorithm. The performance of MOEA/D and NSGA-II can also be verified visually by depicting the Pareto Front \({\textit{EP}}\). According to [36], if \(A\) and \(B\) are two non-dominated sets resulted from two MOEAs A and B, respectively, then, the set coverage quality metric (C-metric) will express the fraction of solutions in one non-dominated solution set that are dominated by those solutions of the second algorithm. The C-metric is defined as:

$$\begin{aligned} C\left( {A,B} \right) =\left| {b\in B|\exists a\in A:a\succ b} \right| /\left| B \right| \end{aligned}$$
(8)

where \(\succ \) reads \(dominate\). Note that:

  • \(C\left( {A,B} \right) \ne 1-C(B,A),\) and

  • \(A\) is better than \(B\) if \(C\left( {A,B} \right) \) is found to be higher than \(C\left( {B,A} \right) \) over many trials.

The second measure is the domination metric \(Dom\) which is defined as:

$$\begin{aligned} Dom\left( {A,B} \right) =\frac{d(A,B)}{d\left( {A,B} \right) +d(B,A)} \end{aligned}$$
(9)

where \(d\left( {X,Y} \right) =\sum _{x\in X} \left| {\{y\in Y|x\succ y\}} \right| \), is known as the domination factor. Mutually non-dominating pairs are ignored in calculating the dominance factor \(d(A,B)\). If each solution in \(A\) dominates every solution in \(B\), then \(Dom\left( {A,B} \right) =1\). Thus, \(Dom\left( {A,B} \right) =1-Dom(B,A)\). Table 3 presents these two metrics when comparing the quality of non-dominated set obtained by MOEA/D (denoted as \(A)\) against those of NSGA-II (denoted as \(B)\). Also, here best results are given in bold.

Table 3 Comparison of coverage set and domination metric of \({\textit{MOEA}}/D\) versus \({\textit{NSGA}}\hbox {-}{\textit{II}}\) for \(10\) WSNs in each test case

From the results, we can consider that on overall, NSGA-II performs better than MOEA/D, except for two test instances (i.e., test instance 1 and 4) where there is a slight difference between the solutions quality attained by the two algorithms.

For the purpose of illustration, Figs. 3 and 4 depict the non-dominated solution sets provided by the both algorithms for two WSNs belong to the two extreme test instances experimented in our simulations (i.e. test instance \(1\) and \(15\), respectively). \(x,y-coordinates\) in the figures correspond to number of set covers (DRSC) and total coverage reliability. In the first test instance, the average number of upper bound of sensor covers \(\overline{ub}\) is 5.00. Projection in Fig. 3 shows the cardinality and quality of both MOEA/D’s and NSGA-II’s Pareto Fronts. The number of non-dominated solutions being obtained by both algorithms is similar (i.e., \(\left| A \right| =\left| B \right| =5)\). Also, we can see that both \(A\) and \(B\) have nearly equal quality.

Fig. 3
figure 3

Pareto solutions of \({\textit{MOEA}}/D\) (black ball) and \({\textit{NSGA-II}}\) (\(+\)), for one network with number of sensors \(=\) 50 and \(R_s =100\)

Fig. 4
figure 4

Pareto solutions of \({\textit{MOEA}}/D\) (black ball) and \({\textit{NSGA-II}}\) (+), for one network with number of sensors = 90 and \(R_s =300\)

However, the dissimilarity in the quality and number of solutions in \(A\) and \(B\) becomes more apparent while increasing \(\overline{ub}\) to its extreme in the last test instance (i.e., 42.10). For this test instance, Fig. 4 clarifies that NSGA-II discovers more non-dominated solutions (\(\left| B \right| =24\) against \(\left| A \right| =12)\). Also, we can see that more than \(70\,\% \) of solutions in \(B\) are with better quality than those in \(A\). On the other hand, about \(25\,\% \) of MOEA/D solutions are with better quality than those obtianed by NSGA-II.

Figures 5 and 6 also compare the quality and diversity (or width) of Pareto Front solutions of both NSGA-II and MOEA/D in terms of the extremes solutions found there (i.e., minimum and maximum values). Here, \(9\) simulations examine \(90\) different WSNs (each simulation with \(10\) networks) with the same sensing radius setting (\(R_s =500)\), but with different number of sensor nodes (\(\{50,55,\ldots ,90\})\). Thus, each simulation carries out a couple of extremes averaged over 10 different networks with certain number of sensors and fixed value of \(R_s \). Figure 5 depicts the extremes in terms of number of sensor covers, while Fig. 6 depicts the corresponding extremes’ coverage reliability (covers reliability). For complete illustration, we also present the best results delivered by GAMDSC in both figures. Moreover, for more clarity, the overall statistics of NSGA-II, MOEA/D and GAMDSC are presented in Table 4.

Fig. 5
figure 5

Pareto extreme solutions (in terms of number of sensor covers) of \({\textit{MOEA}}/D\) (\(+\)) and \({\textit{NSGA-II}}\) (filled circle). GAMDSC’s best solutions (empty circle) are also included in the figure. The simulation is experimented under 90 networks with \(R_s =500\) and number of sensors \(=\{50,55,\ldots ,90\}\)

Fig. 6
figure 6

Pareto extreme solutions (in terms of covers reliability) of \({\textit{MOEA}}/D\) (\(+\)) and \({\textit{NSGA-II}}\) (filled circle). GAMDSC’s best solutions (empty circle) are also included in the figure. The simulation is experimented under 90 networks with \(R_s =500\) and number of sensors \(=\{50,55,\ldots ,90\}\)

Table 4 Overall statistics of Pareto extremes of MOEA/D and NSGA-II and the best solutions of GAMDSC given in Figs. 5 and 6

According to the tradeoff between the two optimization objectives (i.e., number of sensor covers and covers reliability), an extreme solution (in Fig. 5) which is dedicated on maximizing the number of sensor covers is connected to the extreme solution (in Fig. 6) that is dedicated to the minimum coverage reliability and vice versa.

From the results of both figures, we can see that NSGA-II provides a wider Pareto Front than MOEA/D and with better quality. The detailed results of Fig. 5 clarifies that the average distance between the extremes of NSGA-II is about \(28.26\) sensor covers, while for the extremes of MOEA/D is \(22.68\) sensor covers. Also, the results reveals that the maximum number of sensor covers provided by MOEA/D ranges from \(34.8\) to \(61.\)8 with an average of \(49.12\). On the other hand, NSGA-II provides more number of sensor covers ranges from \(35.2\) to \(63.9\) with average of \(50.34\). Both MOEA/D and NSGA-II gain more sensor covers than GAMDSC. In Fig. 5, we can see that GAMDSC can provide from \(33.8\) to \(60.6\) with average of \(47.91\) sensor covers.

Similarly, Fig. 6 shows that NSGA-II provides wider Pareto Front than MOEA/D. Also, in Fig. 6 we can see that all algorithms have nearly equal reliability performance represented by \(min\) extremes of MOEA/D and NSGA-II and by the best results of GAMDSC. Knowing that these reliability results are associated with the \(max\) extremes (in terms of number of sensor covers) of MOEA/D and NSGA-II and with the best results of GAMDSC in Fig. 5, we can then realize that NSGA-II facilitate more qualitative sensor covers. In contrast, GAMDSC is incurred by the lowest number of sensor covers. This reflects the positive impact of the proposed heuristic perturbation operator over the combined collaboration of crossover, mutation, and the scattering operators of GAMDSC algorithm.

4.3 Comparison of Worst Case Time Complexity

We also compare the worst-case computational complexity of the single-objective GAMDSC algorithm and the multi-objective MOEA/D and NSGA-II algorithms for solving Disjoint Reliable Sensor Covers Problem (DRSC). Due to optimizing multiple objectives instead of only one objective, multi-objective evolutionary algorithms, in general, are more expensive than single-objective algorithms. Without loss of generality, the computational complexity of single-objective genetic algorithms, such as GAMDSC is \(O(max_i \times K)\) where \(K\) is the size of solutions to be evolved, via certain evolution operators, for \(max_i \) iterations. On the other hand, the general computational complexity of an MOEA/D is \(O(max_i \times 2\times K\times K_{neighbor} )\) (where \(2\) means number of objective functions and \(K_{neighbor} <K\) means size of neighboring solutions for each individual solution) [35]. The computational complexity of NSGA-II is \(O(max_i \times 2\times K^{2})\) [32, 33].

Now, let us consider the computation time needed for the most critical parts of the above algorithms when applied for solving DRSC problem. It is stated in [15] that in each generation of GAMDSC, the fitness function is the most critical part which represents the computation time bottleneck. The evaluation of the fitness function is linearly related to the number of sensors and the number of targets. Thus, the worst-case time complexity of GAMDSC is:

$$\begin{aligned} O\left( {{\textit{GAMDSC}}} \right) =max_i \times K\times \left| n \right| \times \left| m \right| \end{aligned}$$
(10)

where \(\left| n \right| \) is the number of targets and \(\left| m \right| \) is the number of sensors. On the other hand, for MOEA/D and NSGA-II, the proposed heuristic perturbation operator (Algorithm 1) costs the most computation time and in the worst-case time it linearly related to the maximum number of sensors and targets, and upper bound of sensor covers \(ub\). Thus, the worst-case time complexity of MOEA/D and NSGA-II are, respectively, as:

$$\begin{aligned} O\left( {{\textit{MOEA}}/D} \right)&= max_i \times 2\times K\times K_{neighbor} \times \left| n \right| \times \left| m \right| \times \left| {ub} \right| \end{aligned}$$
(11)
$$\begin{aligned} O\left( {{\textit{NSGA-II}}} \right)&= max_i \times 2\times K^{2}\times \left| n \right| \times \left| m \right| \times \left| {ub} \right| \end{aligned}$$
(12)

5 Conclusions

Maximum DSC is a well-known WSN design which is NP-hard. Unlike all of the existing work that rely on the traditional single-objective definition, this paper introduces a multi-objective Disjoint Reliable Sensor Covers Problem (DRSC). The formulated problem is then tackled by two well-known and competitor multi-objective evolutionary algorithms, namely MOEA/D and NSGA-II. Instead of the traditional use of two perturbation operators (i.e., crossover and mutation), one heuristic operator is proposed that involves the incorporation of problem-specific knowledge in its mechanism. The proposed heuristic operator is proved to be of positive effect on the overall performance of the implemented MOEAs. The results over 150 different WSN topologies signifies the importance of utilizing the two contradictory criteria of getting large number of disjoint sensor covers and more coverage reliability and optimizing them simultaneously rather than optimizing only number of disjoint sensor covers. Good overall performance of the multi-objective algorithms for the modeled DRSC highlights that it may be motivated to consider another WSN’s design criterion while solving disjoint sensor covers problem. As an example, we can include communication constraints (e.g., sensor-sink connectivity) to the multi-objective definition of disjoint sensor covers. Also, by satisfying the sensor energy constraints, a sensor can participate in multiple sensor covers and the definition of DRSC could be reformulated to satisfy a multi-objective non-disjoint sensor covers (nDRSC) problem.