1 Introduction

Practically, reliability enhancement may be involved with other mutually conflicting objectives such as system's cost, weight, volume, etc. Multi-objective optimization problems (MOOPs) are an evitable part of the reliability-based system design. Conventional optimization methods assume that all the design data involved in the system design are precisely known, the constraints delimit a well-defined set of feasible decisions, and the objectives are well defined and easy to formulate. However, incompleteness and unreliability of input information are typical for many practical problems of multi-objective optimization decision-making of reliability. Park [1] suggested that a fuzzy non-linear optimization technique is a superior way to analyze the system reliability. The fuzzy approach [2] deals with the kind of uncertainty that arises due to imprecision associated with the complexity of the system as well as the vagueness of human judgment. Dhingra and Moskowitz [3] used fuzzy set theory effectively in engineering design problems where uncertainty or ambiguity arises about the preciseness of permissible parameters, degree of credibility, and correctness of statements and judgments. Bellman and Zadeh [4] initially proposed a fuzzy optimization technique, which combines fuzzy goals and fuzzy decision space by using aggregate operators. Dhingra [5], Rao and Dhingra [6] solved fuzzy multi-objective optimization problems in four and five stages series system subject to several design constraints. Huang [7] developed a fuzzy multi-objective decision-making method in the series system. Ravi et al. [8] proposed various kinds of aggregate operators to look into the impacts on their optimal system design in complex systems. Huang et al. [9] proposed a coordination method in solving fuzzy multi-objective optimization of system reliability. Mahapatra and Roy [10] proposed a fuzzy multi-objective optimization method in the decision-making of multi-objective reliability optimization. Garg and Sharma [11] also report that the probabilistic approach is unsuitable to reliability analysis due to its random nature and model the multi-objective reliability-redundancy allocation problem in a fuzzy environment.

In order to solve the MOOPs, generally weighted sum approach is used [12]. This method finds one Pareto solution at a time by assigning a weight to each objective function. Such a method is not fit for producing a well distributive solution set which is considered an ideal approach to an MOOP. Apart from this, such methods require tedious work to find multiple solutions, adequately showing limitations in the search space like plateaus, or ridges in the fitness (search) landscape, or discontinuous. A set of well-distributive trade-off optimal solutions is an important demand for ideal multi-objective optimization. This procedure is more practical, methodical and less subjective [12]. Keeping these views in mind, a number of MOEAs such as MOGA-Multi-Objective Genetic Algorithm [13], NPGA-Niched Pareto Genetic Algorithm [14], NSGA-Non-dominated Sorting Genetic Algorithm [15], SPEA-Strength Pareto Evolutionary Algorithm [16], PAES-Pareto Archived Evolution Strategy [17], PESA-Pareto Envelop-based Selection Algorithm [18], MOMGA-Multiobjective Optimization with Messy Genetic Algorithm [19], PESA-II [20], SPEA2 [21], NSGA-II [22], MOEA/D-Multi-objective Evolutionary Algorithm based on Decomposition [23], AGE-Approximation Guided Evolutionary-II [24], NSGA-III [25] etc., are suggested to tackle the different types of the MOOPs. MOEAs and their solution approaches can be viewed in Deb [12], Coello et al. [26], and Das and Panigrahi [27]. A more updated survey on MOEAs can be viewed in Zhang and Xing [28].

NSGA-II is a second-generation MOEA, effectively used by Salazar et al. [29] in constrained multi-objective reliability and redundancy problems to identify a Pareto-optimal set. Later, Wang et al. [30] solved the multi-objective reliability redundancy allocation problem (MORRAP) using NSGA-II and compared their results with single-objective approaches. Kishore et al. [31, 32] suggested fuzzy multi-objective reliability optimization using NSGA-II. Safari [33] proposed a variant of NSGA-II in solving an MORRAP. Vitorino et al. [34] tackled the reliability problem in multi-objective power distribution system using NSGA-II. Sharifi et al. [35] solved MORRAP for a series–parallel problem and k-out-of-n subsystems with 3 objectives using NSGA-II. Kumar et al. [36,37,38,39] analyzed the reliability optimization problem using NSGA-II. In recent work, Muhuri et al. [40] addressed the problem of MORRAP using NSGA-II with an interval type-2 fuzzy set.

In this paper, an approach is developed to find the optimal system design from the POF. The optimization goals are set in the following fashion:

  • The distance of the resulting POF to the true POF should be minimized.

  • The obtained Pareto-optimal solutions should be uniformly distributed.

  • The spread of the obtained POF should be maximized, i.e., each objective function should have a wide range of values.

To achieve the above goals, hybrid NSGA-II is proposed for fuzzy multi-objective reliability optimization. The optimal system design is found in a purely multi-objective manner where no aggregate operator is used and a well distributive solution set is obtained. Finally, a fuzzy ranking method [41] is used to obtain the best compromise solution from the solution set. A comparative experiment has been performed to show its efficacy with the existing approach.

The rest of the paper is organized as follows. Section 2 gives some preliminaries. Section 3 gives a brief description of NSGA-II. Section 4 gives the proposed approach in step by step. Section 5 gives the problem statement of the over-speed protection system. Section 6 gives the results and discussion and Sect. 7 gives conclusions.

2 Preliminaries

In this section, some basic facts of this paper are given.

2.1 Formulation of an MOOP

In general, an MOOP is defined as follows:

$$\begin{aligned} & {\text{Minimize}}\;F\left( X \right) = \left[ {f_{1} \left( X \right), f_{2} \left( X \right), \ldots ,f_{k} \left( X \right)} \right]^{T} \\ & {\text{subject to}}\;g_{i} \left( X \right) = 0, \quad i = 1, 2, \ldots , m_{e} ; \\ & g_{i} \left( X \right) \ge 0, \quad i = m_{e} + 1, \quad m_{e} + 2, \ldots , m; \\ & x_{j}^{l} \le x_{j} \le x_{j}^{u} ,\quad j = 1, 2, \ldots ,n \\ \end{aligned}$$
(1)

where \(k \ge 2\) is the number of objectives; m is the total number of constraints; \(m_{e}\) is the number of equality constraints; \(X = \left[ {x_{1} , x_{2} , \ldots , x_{n} } \right]^{T}\) is n-dimensional decision vector from the feasible region or decision space \(\varOmega\) defined by

$$\varOmega = \left\{ {X \in {\mathbb{R}}^{n}\Big | \begin{array}{*{20}c} {g_{i} \left( X \right) \ge 0,\quad g_{i} \left( X \right) = 0, \quad x_{j}^{l} \le x_{j} \le x_{j}^{u} ,} \\ { j = 1, 2, \ldots , n} \\ \end{array} } \right\}$$

The image of the feasible region is denoted by \(Z \subset {\mathbb{R}}^{k}\) and it is called a feasible objective region or objective space which is defined by \(Z = \left\{ {F\left( X \right) \in {\mathbb{R}}^{k} {\mid }X \in \varOmega } \right\}\). The elements of \(Z\) are called objective vectors or criterion vector denoted by \(z = F\left( X \right) = \left[ {f_{1} \left( X \right), f_{2} \left( X \right), \ldots ,f_{k} \left( X \right)} \right]^{T}\); \(x_{j}^{l}\) and \(x_{j}^{u}\) are the lower and upper bounds of the decision variable \(x_{j}\) respectively. For every point \(X\) in the decision space Ω, there exists a point \(F\left( X \right)\) in the objective space \(Z\). It is a mapping between the n-dimensional solution vector and the k-dimensional objective vector.

2.2 Basic definitions

The concept of optimality in an MOOP depends on Pareto optimality. Therefore, the following definitions can be defined in terms of Pareto terminology [26].

Definition 1

(Pareto dominance) A vector \(U = \left[ {u_{1} , u_{2} , \ldots ,u_{k} } \right]^{T}\) is said to dominate another vector \(V = \left[ {v_{1} , v_{2} , \ldots ,v_{k} } \right]^{T}\) (denoted by \(U{ \preccurlyeq }V\)) if and only if \(U\) is partially less than \(V\), i.e., ∀ \(r \in \left\{ {1, \ldots ,k} \right\}\), \(u_{r} \le v_{r}\) and ∃ at least one \(\dot{r} \in \left\{ {1, \ldots ,k} \right\} {\text{s}}.{\text{t}}. u_{{\dot{r}}} < v_{{\dot{r}}}\).

Definition 2

(Pareto optimality) A solution vector \(X \in \varOmega\) is said to be Pareto-optimal with respect to \(\varOmega\) if and only if there is no \(X^{\prime } \in \varOmega\) for which \(V = \left[ {f_{1} \left( {X^{'} } \right), f_{2} \left( {X^{'} } \right), \ldots ,f_{k} \left( {X^{'} } \right)} \right]^{T}\) dominates \(U = \left[ {f_{1} \left( X \right), f_{2} \left( X \right), \ldots ,f_{k} \left( X \right)} \right]^{T}\). The phrase “Pareto-optimal” is taken to mean with respect to the entire decision variable space unless otherwise specified.

Definition 3

(Pareto-optimal set) For a given MOOP, \(F\left( X \right)\), the Pareto-optimal set, \(\wp^{*}\), is defined as:

$$\wp^{*} : = \left\{ {X \in \varOmega {\mid }\neg \exists X^{\prime } \in \varOmega \; {\text{s}}.{\text{t}}.\; F\left( {X^{\prime } } \right){ \preccurlyeq }\,F\left( X \right)} \right\}$$

Definition 4

(Pareto-optimal front) For a given MOOP, \(F\left( X \right)\), and the Pareto-optimal set, \(\wp^{*}\), the Pareto-optimal front \(\wp {\mathcal{F}}^{*}\) is defined as:

$$\wp {\mathcal{F}}^{*} : = \left\{ {\left[ {f_{1} \left( X \right), f_{2} \left( X \right), \ldots ,f_{k} \left( X \right)} \right]^{T} {\mid }X \in \wp^{*} } \right\}$$

Definition 5

(Ideal objective vector) [26] If a decision vector \(X^{*} = \left[ {x_{1}^{*} ,x_{2}^{*} , \ldots ,x_{n}^{*} } \right]^{T} \in \varOmega\) is such that \(f_{i} \left( {X^{*} } \right) = \min_{X \in \varOmega } f_{i} \left( X \right)\); \(i \in \{ 1, 2, \ldots ,k\}\) then the vector \(F\left( {X^{*} } \right) = \left[ {f_{1} \left( {X^{*} } \right),f_{2} \left( {X^{*} } \right), \ldots ,f_{k} \left( {X^{*} } \right)} \right]^{T} \in Z\) is called an ideal objective vector for an MOOP given by (1) and \(X^{*} = \left[ {x_{1}^{*} ,x_{2}^{*} , \ldots ,x_{n}^{*} } \right]^{T} \in \varOmega\) is called an ideal vector.

Definition 6

(Fuzzy set) [2] Let \(X\) be a collection of objects generically denoted by \(x\). A fuzzy set \(\tilde{A}\) in \(X\) is a set of ordered pair defined in the form as

$$\tilde{A} = \left\{ {\left( {x, \mu_{{\tilde{A}}} \left( x \right)} \right){\mid }x \in X} \right\}$$

where \(\mu_{{\tilde{A}}}\): \(X \to [0, 1]\) is called the membership function and its value is called the grade of membership of \(x\) in \(\tilde{A}\).

Definition 7

[Linguistic hedge (or modifier)] [37] A linguistic hedge (or a linguistic modifier) is an operation that modifies the meaning of the term. Suppose \(\tilde{H}\) is a fuzzy set in \(X\), then the modifier m generates the composite term \(\tilde{M} = m(\tilde{H})\). Modifiers are frequently used in mathematical models as follows.

Concentration It decreases the membership grades of all the members of \(\tilde{H}\) by spreading in the curve. It is defined as:

$$\mu_{{con\left( {\tilde{H}} \right)}} \left( x \right) = \left( {\mu_{{\left( {\tilde{H}} \right)}} \left( x \right)} \right)^{2} \quad {\text{for all}}\quad x \in X$$

Dilation Similarly, it increases the membership grades of all members by spreading out the curve. It is defined as:

$$\mu_{{dil\left( {\tilde{H}} \right)}} \left( x \right) = \left( {\mu_{{\left( {\tilde{H}} \right)}} \left( x \right)} \right)^{1/2} \quad {\text{for all}}\quad x \in X$$

Therefore, in general, strong and weak modifiers are given as: \(m_{\delta } \left( {\mu_{{\left( {\tilde{H}} \right)}} \left( x \right)} \right) = \left( {\mu_{{\left( {\tilde{H}} \right)}} \left( x \right)} \right)^{\delta }\) = a strong modifier or concentrator, if \(\delta > 1\) and a weak modifier or dilator if \(\delta < 1.\)

The following linguistic hedges are associated with above mathematical operators: very \(\tilde{H}\)= con(\(\tilde{H}\)); more or less \(\tilde{H}\)= dil(\(\tilde{H}\)); Indeed \(\tilde{H}\) = Int(\(\tilde{H}\)); plus \(\tilde{H}\)=\(\tilde{H}^{1.25}\); slightly \(\tilde{H}\)= int[plus \(\tilde{H}\) and not (very \(\tilde{H}\)].

3 Elitist non-dominated sorting genetic algorithm (NSGA-II)

Srinivas and Deb [15] proposed the NSGA which is an early dominance-based MOEA. The purpose of developing NSGA is to find better solutions according to their non-domination levels. NSGA uses the naive and slow [12] sorting approach to distribute a population into different non-domination levels, and a sharing function method to maintain the diversity of the population. However, it has high computational complexity \(O(kN^{3} )\), where k is the number of objectives and N is the population size in the non-dominated sorting. So, NSGA is computationally expensive for large population sizes. Moreover, NSGA is a non-elitist approach that affects its convergence rates compared to other MOEAs and it also requires a sharing parameter to calculate the sharing fitness, which ensures the diversity of the population.

Deb et al. [22] proposed NSGA-II to overcome the drawbacks of NSGA. Specifically, NSGA-II presents a fast-non-dominated sorting approach with the worst-case computational complexity \(O(k(2N)^{2} )\). This approach searches iteratively non-dominated solutions into different fronts.

First, for each solution i in the population, the algorithm calculates two entities:

  1. 1.

    \(n_{i}\), the number of solutions dominating i,

  2. 2.

    \(S_{i}\), a set of solutions dominated by i.

The solutions for which \(n_{i} = 0\) belong to the first front. Second, for each member j in the set \(S_{i}\), the value of \(n_{j}\) is reduced by one. If any \(n_{j}\) is reduced to zero during this stage, the corresponding member j is put in the second front. The above process is continued with each member in the second front to identify the third front and so on. Furthermore, NSGA-II applies the concept of crowding-distance with the worst-case computational complexity \(O(k\left( {2N} \right)\log \left( {2N} \right))\). The introduction of crowding-distance replaces the fitness sharing approach that requires a sharing parameter to be set by the user.

The crowding-distance value (\(CD_{i} )\) (See Fig. 1) of the ith solution is calculated as follows:

$$d_{ip} = \frac{{f_{p}^{i + 1} - f_{p}^{i - 1} }}{{f_{p}^{max} - f_{p}^{min} }}$$
(2)
$$CD_{i} = \mathop \sum \limits_{p = 1}^{k} d_{ip}$$
(3)

where \(f_{p}^{i + 1}\) and \(f_{p}^{i - 1}\) denote the pth objective function of the \(\left( {i + 1} \right)\) th and \(\left( {i - 1} \right)\) th individual (solution) respectively, and \(f_{p}^{max}\) and \(f_{p}^{min}\) represent the maximum and minimum values of the pth objective function.

Fig. 1
figure 1

Fitness evaluation and individual crowding distance estimation

According to Deb et al. [22] “A higher value of crowding-distance gives the lesser crowded region and vice versa”. After applying the fast-non-dominated sorting procedure, the crowding-distance picks those solutions which locate in the less-crowded region or possesses a higher value of crowding distance. This procedure is extended up to the entire POF for maintaining the diversity in the solutions set. Finally, NSGA-II applies an elitist strategy with the worst-case computational complexity of \(O(2N\log \left( {2N} \right))\). The elitist strategy [42, 43] improves the convergence of an MOEA and avoids the loss of optimal solutions after getting it.

For handling the constraints, the binary tournament selection method is applied. A search space (decision space) is categorized into two regions—feasible and infeasible. According to Deb et al. [22], “a solution \(X\) is called a constrained-dominate to a solution Y if

  • X is feasible and Y is infeasible.

  • X and Y are infeasible, but X contains a smaller overall constraint violation.

  • X and Y are feasible, but X dominates Y”.

In Fig. 2, the evaluation cycle of the NSGA-II is shown. First, an offspring \(Q_{t}\) of size N is obtained by using genetic operators such as selection, recombination, and mutation. A combined population \(R_{t}\) of size 2N is then formed which consists of the current population \(P_{t}\) and the offspring population \(Q_{t}\). By using fast non-dominated sorting, \(R_{t}\) is divided into different fronts \(PF_{1} , PF_{2} , \ldots ,PF_{n}\). Let the number of solutions in each front \(PF_{i}\) be \(N_{i}\). Next, we choose members for the new population \(P_{t + 1}\) from the front \(PF_{1}\) to \(PF_{t - 1}\), noting that \(N_{1} + N_{2} + \cdots + N_{t} > N\) and \(N_{1} + N_{2} + \cdots + N_{t - 1} \le N\). Afterwards, to get the exactly N population members in \(P_{t + 1}\), we sort the solutions in front \(PF_{t}\) using the crowding distance sorting procedure and choose the best solutions to fill an empty slot in the new population \(P_{t + 1}\). This process is continued until the termination condition is satisfied. The flow diagram of NSGA-II is shown in Fig. 3.

Fig. 2
figure 2

An evaluation cycle of the NSGA-II algorithm

Fig. 3
figure 3

Flow diagram of the NSGA-II

4 Proposed methodology

This section describes the proposed methodology in step by step as follows.

Step 1 Determine the best and worst values for each objective function.

Each objective function is solved as a single-objective optimization problem (SOOP) one by one to get its boundaries as follows:

$$\begin{aligned} & {\text{Minimize}}\;f_{p} (X);\quad p = 1, 2, \ldots ,k \\ & {\text{subject to}}\;g_{i} \left( X \right) = 0, \quad i = 1, 2, \ldots , m_{e} ; \\ & g_{i} \left( X \right) \ge 0, \quad i = m_{e} + 1, m_{e} + 2, \ldots , m; \\ & x_{j}^{l} \le x_{j} \le x_{j}^{u} ,\quad j = 1, 2, \ldots ,n \\ \end{aligned}$$
(4)

Let us consider \(X_{i}^{*}\), \(i = 1, 2, \ldots ,k\) be the optimal design vector for each objective function \(f_{i}\), then construct a matrix as

$$M = \left[ {\begin{array}{*{20}c} {f_{1} (X_{1}^{*} )} & {\quad f_{2} (X_{1}^{*} )} & {\quad \ldots } & {\quad f_{k} (X_{1}^{*} )} \\ {f_{1} (X_{2}^{*} )} & {\quad f_{2} (X_{2}^{*} )} & {\quad \ldots } & {\quad f_{k} (X_{2}^{*} )} \\ \vdots & {\quad \vdots } & {\quad \ddots } & {\quad \vdots } \\ {f_{1} (X_{k}^{*} )} & {\quad f_{2} (X_{k}^{*} )} & {\quad \ldots } & {\quad f_{k} (X_{k}^{*} )} \\ \end{array} } \right]$$
(5)

where the diagonal entries of matrix M are found to be the minima in their respective columns. The best \(m_{p}\) and worst \(M_{p}\) values of the pth objective function \(f_{p}\) are identified as:

$$m_{p} = f_{p} \left( {X^{*} } \right) = \mathop {\hbox{min} }\limits_{p} f_{p} \left( {X_{p}^{*} } \right)$$
(6)
$$M_{p} = f_{p} \left( {X^{*} } \right) = \mathop {\hbox{max} }\limits_{p} f_{p} \left( {X_{p}^{*} } \right)$$
(7)

Step 2 Construct the membership function.

Here, the sigmoidal shape is chosen for the membership function in the given problem because “When the decision-maker (DM) is worse off with respect to a goal for the objective function, the DM tends to have a higher marginal rate of satisfaction with respect to that objective function. A convex shape captures that behavior in the membership function. On the other hand, when the DM is better off with respect to a goal, the DM tends to have a smaller marginal rate of satisfaction. Such behavior is modeled using the concave portion of the membership function” [5]. It is defined as:

$$\mu_{H} \left( {f_{p} } \right) = \left\{ {\begin{array}{*{20}l} {1,} \hfill & { \quad if\,f_{p} \le m_{p} } \hfill \\ {\frac{1}{2}tanh\left( {\left( {\frac{{M_{p} + m_{p} }}{2} - f_{p} } \right)\alpha_{p} } \right) + \frac{1}{2},} \hfill & {\quad if\, m_{p} \le f_{p} \le M_{p} ;\quad \alpha_{p} = 6/\left( {M_{p} - m_{p} } \right)} \hfill \\ {0,} \hfill & { \quad if\, f_{p} \ge M_{p} ,} \hfill \\ \end{array} } \right.$$
(8)

Step 3 Reformulate the MOOP as a fuzzy MOOP.

Each membership function is expected to achieve the maximum satisfaction level. Therefore, a fuzzy MOOP is suggested to maximize all the membership functions so that it could get the maximum satisfaction level simultaneously [39]. From the extreme values of \(f_{p}\) obtained by (6) and (7), a fuzzy MOOP can be given as:

$${\text{Maximize}}\;\mu \left( F \right) = \left[ {\left( {\mu_{1} \left( {f_{1} } \right),\mu_{2} \left( {f_{2} } \right), \ldots ,\mu_{k} \left( {f_{k} } \right)} \right)} \right]^{T}$$

or, Minimize

$$\begin{aligned} & \mu \left( F \right) = \left[ {\left( { - \mu_{1} \left( {f_{1} } \right), - \mu_{2} \left( {f_{2} } \right), \ldots , - \mu_{k} \left( {f_{k} } \right)} \right)} \right]^{T} \\ & {\text{subject to}}\;g_{i} \left( X \right) = 0, i = 1, 2, \ldots , m_{e} ; \\ & g_{i} \left( X \right) \ge 0, i = m_{e} + 1, m_{e} + 2, \ldots , m; \\ & x_{j}^{l} \le x_{j} \le x_{j}^{u} ,\quad j = 1, 2, \ldots ,n \\ \end{aligned}$$
(9)

Theorem 1

The Pareto-optimal solution of the fuzzy MOOP (9) satisfies the MOOP (1).

Proof

Let \(X^{*}\) be a Pareto-optimal solution of the fuzzy MOOP (9). Then, it is defined by definition as follows:

  • \({\nexists }\) \(X \in \varOmega\) s.t. \(- [\mu_{i} \left( f_{i} \right)] \left( {X} \right)\le - [\mu_{i} \left( f_{i} \right)] \left( {X^{*} } \right)\)\(i = 1, 2, \ldots ,k\), and \(- [\mu_{j} \left( f_{j} \right)] \left( {X} \right)< - [\mu_{j} \left( f_{j} \right)] \left( {X^{*} } \right)\) for at least one \(j \in \left\{ {1, 2, \ldots ,k} \right\}\), \(j \ne i\).

  • \({\nexists }\) \(X \in \varOmega\) s.t. \(- h_{i} \left[ {f_{i} (X)} \right] \le - h_{i} \left[ {f_{i} (X^{*} )} \right]\)\(i = 1, 2, \ldots ,k\), and \(- h_{j} \left[ {f_{j} (X)} \right] < - h_{j} \left[ {f_{j} (X^{*} )} \right]\) for at least one \(j \in \left\{ {1, 2, \ldots ,k} \right\}\), \(j \ne i\).

  • \({\nexists }\) \(X \in \varOmega\) s.t. \(f_{i} (X) \le f_{i} (X^{*} )\)\(i = 1, 2, \ldots ,k\), and \(f_{j} (X) < f_{j} (X^{*} )\) for at least one \(j \in \left\{ {1, 2, \ldots ,k} \right\}\), \(j \ne i\); since h is monotonically decreasing function.

  • \({\nexists }\) \(X \in \varOmega\) s.t. \(- f_{i} \left( X \right) \ge - f_{i} (X^{*} )\)\(i = 1, 2, \ldots ,k\), and \(- f_{j} (X) > - f_{j} (X^{*} )\) for at least one \(j \in \left\{ {1, 2, \ldots ,k} \right\}\), \(j \ne i\); since \(h\) is monotonically increasing function.

  • \({\nexists }\) \(X \in \varOmega\) s.t. \(f_{i} (X) \le f_{i} (X^{*} )\)\(i = 1, 2, \ldots ,k\), and \(f_{j} (X) < f_{j} (X^{*} )\) for at least one \(j \in \left\{ {1, 2, \ldots ,k} \right\}\), \(j \ne i\); since \(h\) is either monotonically decreasing or increasing function.

  • \(X^{*}\) is a Pareto-optimal solution of the MOOP given by (1).□

Step 4 Biased Pareto-optimal solutions.

The proposed approach does not find a single compromise solution; instead, it finds a biased distribution of solutions. If a user wants to have a bias towards a particular objective, the modified membership function produces more solutions towards the preferred region in the search space. Finding more dense solutions in the region of preference is a much better approach then predefining a weighted sum of objective and then finding only one the optimum solution. Here, the linguistic modifier is proposed to show biasedness towards the objective.

Step 5 Set the parameters of NSGA-II and apply to the fuzzy MOOP (9) with modification of linguistic modifier.

The NSGA-II parameters such as number of objectives, number of design constraints, number of design variables, population size, the maximum number of generations, crossover probability, mutation probability, distribution indices [12] for SBX crossover and Polynomial mutation are set accordingly to the given problem. Rigorous experimentation and tuning of the parameters are required to achieve a well-distributive Pareto-optimal solution set.

Step 6 Find the Pseudo-weight vector.

The purpose of local search can be fulfilled by finding a pseudo weight to each Pareto point. It is calculated as follows:

$$w_{i} = \frac{{\mu_{i} \left( {f_{i} } \right)}}{{\mathop \sum \nolimits_{j = 1}^{k} \mu_{j} \left( {f_{j} } \right)}} ,\quad i = 1, 2, \ldots ,k$$
(10)

\(w = \left[ {w_{1} ,w_{2} , \ldots ,w_{k} } \right]^{T}\) represents the Pseudo-weight vector dictating roughly the priority of different objective functions for the solution \(X\), \(\sum\nolimits_{i = 1}^{k} {w_{i} } = 1\), \(w_{i} \ge 0\).

Step 7 Start the local search.

After getting the pseudo-weight to each Pareto-point, the local search strategy is used to update the Pareto-optimal solution. It helps in enhancing the convergence rate to each Pareto point. The local search strategy is given as follows.

The crisp formulation of fuzzy MOOP (9) is given by

$$\begin{aligned} & {\text{Maximize}}\; \left( {1{\wedge }\mu_{1} \left( {f_{1} } \right)/w_{1} } \right){\wedge }\left( {1{\wedge }\mu_{2} \left( {f_{2} } \right)/w_{2} } \right){\wedge } \ldots {\wedge }\left( {1{\wedge }\mu_{k} \left( {f_{k} } \right)/w_{k} } \right) \\ & {\text{subject to}}\;g_{i} \left( X \right) = 0, i = 1, 2, \ldots , m_{e} ; \\ & g_{i} \left( X \right) \ge 0, i = m_{e} + 1, m_{e} + 2, \ldots , m; \\ & x_{j}^{l} \le x_{j} \le x_{j}^{u} ,\quad j = 1, 2, \ldots ,n \\ \end{aligned}$$
(11)

where \(\mu \left( F \right) = \left[ {\left( {\mu_{1} \left( {f_{1} } \right),\mu_{2} \left( {f_{2} } \right), \ldots ,\mu_{k} \left( {f_{k} } \right)} \right)} \right]^{T}\) is a vector of k-fuzzy regions of satisfaction corresponding to k-objective vector, \({\wedge }\) denotes the aggregate operator known as intersection and it is defined as the minimum operator i.e., \(\left( {1{\wedge }\mu_{i} \left( {f_{i} } \right)/w_{i} } \right)\) gives the minimum of two values 1 and \(\mu_{i} \left( {f_{i} } \right)/w_{i}\) that value lies in [0, 1].

Step 8 Apply the clustering technique.

In order to maintain a good spread in the solutions set, a clustering technique is used. It also reduces the size of the solutions set that is useful for a practical point of view [12]. Let \(P_{t + 1}\) be the set of non-dominated solutions of size N.

The clustering algorithm (see Fig. 4) is given as follows:

Fig. 4
figure 4

Flow diagram of the clustering technique

  • Step 1 Initially, each solution belongs to a distinct cluster or \(C_{i} = \left\{ i \right\}\) so that \(C = \left\{ {C_{1} ,C_{2} , \ldots ,C_{N} } \right\}\).

  • Step 2 If \(N \le N^{\prime }\) (required no. of solutions), go to Step 5. Otherwise, go to Step 3.

  • Step 3 For each pair of clusters (there are \(\left( {\begin{array}{*{20}c} {\left| C \right|} \\ 2 \\ \end{array} } \right)\) of them), calculate the cluster-distance by using the following formula:

    $$D_{12} = \left( {1/\left| {C_{1} } \right|\left| {C_{2} } \right|} \right) \mathop \sum \limits_{{i \in C_{1} ,j \in C_{2} }} D\left( {i, j} \right)$$
    (12)

    where \(\left| {C_{1} } \right|\) and \(\left| {C_{2} } \right|\) number of solutions in the cluster set \(C_{1}\) and \(C_{2}\) respectively, and \(D\left( {i, j} \right)\) is the Euclidean distance between two solutions i and j. Find the pair \(\left( {i_{1} ,i_{2} } \right)\) which corresponds to the minimum cluster-distance.

  • Step 4 Merge the clusters \(C_{{i_{1} }}\) and \(C_{{i_{2} }}\) together. This deducts the size of C by one. Go to Step 2.

  • Step 5 Select only one solution from each cluster and remove others. The solution having the minimum average distance from other solutions in the cluster represents a solution to that cluster.

Step 9 Choose the best compromise solution.

The fuzzy ranking method [41] is used to rank the solutions as per their satisfaction levels. In this way, a solution achieves the highest degree of satisfaction is considered as the best compromise solution. It is given as follows.

$$\mu_{best} = \mathop {\hbox{max} }\limits_{P} \left[ {min\left\{ {\mu_{1} \left( {f_{1} } \right),\mu_{2} \left( {f_{2} } \right), \ldots ,\mu_{k} \left( {f_{k} } \right)} \right\}} \right]$$
(13)

where P is the number of Pareto-optimal solutions.

5 Problem statement

Let us consider a four-stage over-speed protection system for a gas turbine [5]. Here, the electrical and mechanical systems continuously provide detection. If an over-speed occurs, then 4-control valves (V1–V4) get closed and the fuel supply gets interrupted. Each component has a constant failure rate in the system. A schematic representation of the over-speed protection system is shown in Fig. 5. This system is a series–parallel system configuration having several parallels and identical components arrayed at each stage. It is shown in Fig. 6. The objective is to determine the optimal design variables i.e., \(r_{i}\) (component reliability) and \(n_{i}\) (number of the redundant components) at stage i such that:

Fig. 5
figure 5

A schematic diagram of a four-stage over-speed protection system

Fig. 6
figure 6

Series-parallel system configuration

  1. 1.

    the system reliability (\(R_{S}\)) is maximized,

  2. 2.

    the system cost (\(C_{S}\)) is minimized,

  3. 3.

    the system weight (\(W_{S}\)) is minimized.

In addition, several design constraints such as the minimum requirement for the system reliability \(R_{lim}\), overall system cost of the system \(C_{lim}\), the total permissible volume of the system \(V_{lim}\), and maximum allowable weight of the system \(W_{lim}\) are considered.

The mathematical formulation of the over-speed protection system is given as:

$${\text{Maximize}}\;f_{1} \equiv R_{S} \left( {r, n} \right) = \mathop \prod \limits_{i = 1}^{4} \left[ {1 - \left( {1 - r_{i} } \right)^{{n_{i} }} } \right]$$
(14)
$${\text{Minimize}}\;f_{2} \equiv C_{S} \left( {r,n} \right) = \mathop \sum \limits_{i = 1}^{4} c_{i} \left( {r_{i} } \right)\left[ {n_{i} + exp\left( {n_{i} /4} \right)} \right]$$
(15)
$${\text{Minimize}}\;f_{3} \equiv W_{S} \left( {r,n} \right) = \mathop \sum \limits_{i = 1}^{4} w_{i} n_{i} { \exp }\left( {n_{i} /4} \right)$$
(16)
$${\text{subject to}}\;g_{1} \equiv \mathop \sum \limits_{i = 1}^{4} v_{i} n_{i}^{2} \le V_{lim}$$
(17)
$$g_{2} \equiv \mathop \sum \limits_{i = 1}^{4} w_{i} n_{i} { \exp }\left( {n_{i} /4} \right) \le W_{lim}$$
(18)
$$g_{3} \equiv \mathop \prod \limits_{i = 1}^{4} \left[ {1 - (1 - r_{i} )^{{n_{i} }} } \right] \ge R_{lim}$$
(19)
$$g_{4} \equiv \mathop \sum \limits_{i = 1}^{4} \gamma_{i} \left[ { - T/\ln \left( {r_{i} } \right)} \right]^{{\delta_{i} }} \left[ {n_{i} + exp\left( {n_{i} /4} \right)} \right] \le C_{lim}$$
(20)
$$0.5 \le r_{i} \le 1 - 10^{ - 6} , 1 \le n_{i} \le 10, i = 1, 2, 3, 4;$$
(21)

“Each component of the system has a constant failure rate \(\uplambda_{i}\) that follows an exponential distribution” [5]. The reliability of each component is given as:

$$r_{i} \left( T \right) = \mathop \int \nolimits_{T}^{\infty }\uplambda_{i} e^{{ -\uplambda_{i} T}} dT = e^{{ -\uplambda_{i} T}}$$
(22)

It is also mentioned that “\(c_{i} \left( {r_{i} } \right)\) is assumed to be an increasing function of \(r_{i}\) (conversely, a decreasing function of the component failure rate)” [5] in the form:

$$c_{i} \left( {r_{i} } \right) = \gamma_{i} /\uplambda_{i}^{{\delta_{i} }}$$
(23)

From (22) and (23), we have

$$c_{i} \left( {r_{i} } \right) = \gamma_{i} \left[ { - T/\ln \left( {r_{i} } \right)} \right]^{{\delta_{i} }}$$
(24)

where the parameters \(\delta_{i}\) and \(\gamma_{i}\) are the physical feature (shaping and scaling factor) of the cost-reliability curve of each component in the ith subsystem; T is the active operational time; \(v_{i}\) is the volume of a component at stage i; \(w_{i}\) is the weight of a component at stage i; the factor \(exp\left( {n_{i} /4} \right)\) is responsible for the additional cost due to the interconnection between the parallel components. The design data of the given problem is listed in Table 1. The given problem is an MORRAP which contains a series–parallel system configuration. This system contains several parallels and identical components arrayed at each stage. Generally, redundancy is applied to increase the system reliability but this technique gives more complexity in terms of cost, weight, and volume to the system design. So, it is suitable to adopt the multi-objective programming model-based problem. Here, two-fold design variables are required to determine the optimal design of the system. One is the reliability of each component and the other is to select a number of redundant components at each stage. Moreover, several design constraints are considered in this model. Such type of problem is a mixed-integer, non-linear programming problem as well as NP-hard [44] for which only approximate solutions have been proposed [45, 46]. Heuristic approaches to such types of problems can be viewed in [47,48,49,50].

Table 1 Design data for the given problem

6 Results and discussion

After applying the proposed approach, this section describes the numerical results. Initially, the lower and upper limits on each objective are evaluated with the given constraints using MATLAB optimization tool-box function namely “fmincon” [51]. Here, integer variables \(n_{i}\) are initially treated as real variables but during the evaluation of the objective functions, the real values are transformed into the nearest integer values. The results of the SOOPs are given in Table 2 and compared with heuristic methods such as Genetic Algorithm (GA) [52], Particle Swarm Optimization (PSO) [53], and Dhingra’s approach [5]. It gives the best and worst solutions to each of the objective functions. These values are used as boundaries in the membership function of the objective.

Table 2 The optimal solutions for each objective function

All experiments are performed in MATLAB (R2017a) on Intel(R) Core (TM) i3-2370 M CPU @ 2.40 GHz with 4 GB RAM. The parameter settings for the MOEAs presented in this study are as follows.

  1. 1.

    NSGA-II: Population size N = 80; crossover probability pc = 0.9; mutation probability pm = 1/n, where n is the number of design variables; the distribution indices [22] for crossover and mutation operators as \(\eta_{c}\) = 10 and \(\eta_{m}\) = 100; maximum number of generation tmax = 100.

  2. 2.

    SPEA2 [21]: Population size N = 80; pc = 0.9, pm = 1/n; and \(\eta_{c}\) = 10, \(\eta_{m}\) = 100, archive size = 80; tmax = 100.

  3. 3.

    PESA-II [20]: Hyper-grid size = 10 × 10; pc = 0.9, pm = 1/n; and \(\eta_{c}\) = 10, \(\eta_{m}\) = 100, archive size = 80; tmax = 100.

  4. 4.

    MOPSO [54]: Inertia weight w  = 0.8; acceleration constants c1 = c2 = 2; archive size = 80; tmax = 100.

In order to solve SOOP, the parameter settings for GA are taken as selection type = binary tournament, pop_size = 30, crossover rate = 0.9, mutation rate = 0.01, max_gen = 100; in PSO swarm size = 30, cognition learning factors as c1 = c2 = 2, maximum and minimum inertia weights 0.9 and 0.4 respectively. To avoid the stochastic discrepancy, 25 independent runs are made to these algorithms. The given problem is solved by Dhingra [5] using a fuzzy optimization technique (Zimmerman’s approach) [2], goal programming [12], goal attainment [25] methods. The intention was to show how the fuzzy optimization technique can be effective to solve such type of a problem. The sigmoidal shape is chosen for each membership function. At the same time, the preference is given to reliability more than the system cost and the system weight of the system. It is mentioned by Dhingra [5] that the reliability is twice as important as cost and weight. Linguistic hedge “very” is used instead of assigning the weight. The shape of the membership function of the objectives is as shown in Fig. 7a and b. Apart from this, crisp multi-objective reliability optimization is shown by goal programming, goal attainment methods by assigning the weighs to the objectives as w1 = 0.5, w2 = 0.25, w3 = 0.25. Huang’s approach [7] is also applied, where the fuzzy optimization technique combines the degree of satisfaction to each of the objectives using the “min” operator. In this approach, the weights are given as w1 = 1, w2 = 0.5, w3 = 0.5 following the non-convexity. In fuzzy optimization, the assigned weight to the objective function may affect the result especially the case of convexity and non-convexity search space. It is advantageous to adopt the ideal approach by taking the whole POF instead of one solution. That is why the proposed approach does not use any kind of aggregation and finds the solution set in one simulation run. The biasedness of the POF is shown by the linguistic hedge “very” to the reliability while system cost and its weight are taken with no hedge. This behavior is used in the proposed methodology to make the results effective in terms of better POF. This preference is the same as Dhingra [5]. The proposed approach gives the solution set in the desired region. This is an ideal multi-objective optimization procedure to search for the optimal system design [12]. In Fig. 8a, firstly, multiple fronts are shown in the membership values (degree of satisfactions) by applying NSGA-II to fuzzy MOOP. In Fig. 8b, all the fronts are compared with respect to the highest degree of satisfaction. In Fig. 9a, the degree of satisfactions lead to find the solutions in the objective space. From here, the search for the optimal front starts using local search which is later maintained the diversity in the solution set with the help of the clustering technique. In Fig. 9b, the optimal front of the hybrid NSGA-II is compared with first front of the NSGA-II. The obtained optimal front is comparatively shown with other MOEAs in Fig. 10a and b. In Fig. 11a and b, the optimal front of the proposed approach is compared with MOPSO and the preference-based approach (goal programming, goal attainment, Huang’s approach, Zimmerman’s approach) [12] respectively. In Table 3, a well diverse solution set is shown by the proposed approach. In Table 4, the list of the optimal system design obtained by the various approaches is given. Table 5 gives a comparative analysis of the hypervolume metric [55] of the POFs obtained by the various approaches. This performance is based on 10 independent runs of each multi-objective algorithm.

Fig. 7
figure 7

Membership function for a monotonically increasing function i.e. system reliability; b monotonically decreasing functions i.e. system cost and weight

Fig. 8
figure 8

a Multiple fronts are shown in the membership values; b satisfaction level achieved by multiple fronts

Fig. 9
figure 9

a Multiple fronts are shown in the objective space; b The optimal front obtained by the proposed approach is compared with the first front of NSGA-II

Fig. 10
figure 10

The optimal front by the proposed approach is compared with other MOEAs a PESA-II; b SPEA2

Fig. 11
figure 11

The optimal front by the proposed approach is compared with a MOPSO; b classical (preference-based) approach

Table 3 The trade-off optimal solutions or Pareto-optimal solutions obtained by the proposed approach
Table 4 The Pareto-optimal results by the classical (preference-based) and MOEAs approaches
Table 5 Comparison of performance calculated by the hypervolume metric

7 Conclusions

In this piece of work, a hybrid NSGA-II based decision-making is proposed for fuzzy multi-objective optimization of system reliability. The results are shown by taking three mutually conflicting objectives such as system reliability, system cost and system weight subject to several design constraints. The conclusions of the proposed work can be drawn as follows.

  • From Table 2, a design with high system reliability has a high total cost and a large weight. While attempting to minimize the system cost, it results in heavy design with low reliability. When the system weight is minimized, the resulting system has a high cost and low reliability.

  • Single-objective optimization is unable to give compromise solution among these mutually conflicting objectives.

  • From Table 3, a reasonable number of “compromise” or “trade-off” optimal solutions are obtained which can be useful in practical practices.

  • The proposed technique is seen as an ideal multi-objective optimization procedure, where a set of well-distributive solutions is obtained.

  • From Table 4, the hybrid technique finds the optimal system design which satisfies a comparatively maximum degree of satisfaction compared to other MOEAs.

  • From Table 5, a statistical analysis of the hypervolume metric performance is shown which concludes that the proposed approach gives a better distribution of solution set with minimum deviation compared to the others.

  • The preference-based approach does not follow the ideal multi-objective optimization procedure. In fact, it behaves like a random walk in the search space.

  • However, in our experiment, MOPSO achieves the maximum degree of satisfaction (see Table 4), but slightly worse than the proposed approach with respect to the hypervolume metric (see Table 5).

  • The proposed approach can be effective in the decision-making of a fuzzy multi-objective reliability-based system design problem.