1 Introduction

Analog circuit synthesis process is typically characterized by selecting a set of design parameters and topology. With the rapid progress in VLSI technology, manual design of an analog circuit is often confined to the usage of design automation for a fixed topology in finding a solution space, satisfying a set of design constraints. Due to high degree of nonlinearity and interdependence among design parameters, extensive computation is required to evaluate various design parameters during circuit design process. Although it is beneficial to reduce the computation cost associated with generating design performance estimates by modeling the circuit design problem in the form of a linear programming problem and solving it using efficient optimization techniques, relevant emphasis is needed to analyze tradeoffs among various design specifications.

Many optimization techniques have been presented in literature to address the problem associated with analog circuit synthesis. One approach is to analyze analog circuit design using geometric programming [1], which derives a set of specific equations by considering design constraints before applying it to the design process. The look-up table approach [2] is used to approximate circuit equations and replace device circuit models for accurate circuit simulation during optimization. With increase in number of design parameters and device dimension, the size of the table grows exponentially and it becomes difficult to generate and store such large tables efficiently. Further, there are methods, such as goal attainment method, which makes use of weights to control the tradeoffs among design specifications [3]. Further, nature inspired optimization techniques, such as genetic algorithm [4, 5], particle swarm optimization [6], differential evolution [7] etc. solve cost functions involving design specifications, which is translated from modeling the behavior of analog circuit design subject to specific set of design constraints. Such nature inspired algorithms function by emulating certain natural phenomenon to obtain global optimum solution and demand the design of circuit equations to be more realistic by considering suitable design parameters.

Genetic algorithm (GA) has become one of the popular optimization techniques in recent years after introduced by Holland [8]. Mostly, GA has been employed as a stochastic procedure to obtain global optimum solution for different combinatorial optimization problems such as scheduling problems [9], traveling salesman problem [10, 11] and in machine learning [12]. Further, GA has been applied to solve both single objective, multi-objective and many objective problems in literature [13, 14]. Nondominated sorting genetic algorithm (both NSGA-II and NSGA-III) has been presented to handle multi- and many-objective problems including the ability to find best possible solution with acceptable convergence. Although other approaches satisfy themselves by compensating for suboptimal performance in certain cases, NSGA-II has an appealing feature of overcoming difficulties in traditional multi-objective circuit optimization.

The application of NSGA-II has not been seriously addressed in multi-objective circuit optimization as per our knowledge. Therefore, in this paper a hierarchical version of NSGA-II (hNSGA-II) is presented to perform analog circuit optimization using hierarchy in generation of new population during polynomial mutation operation. At first, the multidimensional search space is updated with new parent population using hierarchical polynomial mutation strategy. Later nondominated sorting is used to obtain local optimal solutions among neighborhood in a single hierarchical mutation growth. After this phase, the local individuals are checked for distinct fitness values for diversity preservation. Once the search for unique fitness values is completed, the local best individuals are merged with parent population similar to NSGA-II framework. Further, the proposed hNSGA-II framework facilitates the optimization process to achieve required Pareto front [15] by carrying out repeated calculations, and evaluates circuit design equations comfortably. This paper also develops a theoretical framework for hNSGA-II based on the asymptotic states of a homogeneous Markov chain model and proves the existence of stationary distribution with constant nonzero hierarchical mutation probability value (strictly positive). Further, a strong ergodicity bound is established to ensure the convergence to limit distribution asymptotically. For showcasing the effectiveness of proposed hierarchical scheme in polynomial mutation operation, six different mutation operators are incorporated in hNSGA-II framework and compared for convergence and diversity of Pareto optimal solutions. Further, to evaluate the performance of proposed hNSGA-II framework, several experiments are performed on twelve multi-objective test functions and two design examples. The effectiveness of hNSGA-II is shown by comparing it with standard optimization algorithms.

The rest of the paper is organized as follows. Overview of multi-objective optimization problem and NSGA-II tool are presented in Sect. 2. The proposed hNSGA-II framework is showcased in Sect. 3. A detailed performance analysis of proposed framework is shown by running extensive simulations on standard benchmarks and by comparing the results with other conventional optimization techniques in Sect. 4. The effectiveness of proposed framework is shown by optimizing design specifications for two design examples in Sect. 5. Section 6 concludes the manuscript.

2 Background

2.1 Multi-objective optimization

Multi-objective optimization (MOP) is a method centered on various conditions to satisfy behavior of contradictory objectives or cost functions simultaneously. Contrary to single objective optimization, the solution obtained by MOP is not confined to a single point solution, but a set of balanced solutions known as Pareto optimal solutions, which when plotted forms a Pareto front (PF) in the multi-objective search space [15]. The novelty of MOP using any optimization technique is to evaluate a set of solutions close to PF without loss of convergence and diversity properties. These properties ensure to deal with non-dominated solutions [13], while carrying out continuous design space exploration. Given an n-dimensional Euclidean space \(\mathbf{R}^n\) and objective function vector \(\mathbf{F}(\mathbf{x})\), the multi-objective optimization attempts to find the solution \(\mathbf{F}(\mathbf{x}')\) satisfying,

$$\begin{aligned} \mathbf{F}(\mathbf{x}')&= \min \limits _{ \mathbf{x}\in X} \mathbf{F}(\mathbf{x})\nonumber \\&= \min \limits _{ \mathbf{x}\in X}\ (f_1(\mathbf{x}),f_2(\mathbf{x}), \ldots , f_m(\mathbf{x})) \end{aligned}$$
(1)

where \(\mathbf{F} : \mathbf{R}^n \rightarrow \mathbf{Q}^m\) consists of m real-valued objective functions and \(\mathbf{Q}^m\) is the objective space [15]. Since the objective functions in (1) often contradict with each other, it is impossible to find a single solution in \(\mathbf{R}^n\) that minimizes all objectives simultaneously. Therefore, multi-objective optimization finds the best tradeoffs among objective functions in terms of Pareto optimality, which is guided by a dominance relation between solutions. Let \(a,b\in \mathbf{Q}^m\), a dominates b if and only if \(a_i \le b_i\) \(\forall i \in \{1,\ldots ,m\}\) and \(a_k < b_k\) for at least one index \(k \in \{1,\ldots ,m\}\) [15]. A point \(x' \in \mathbf{Q}^m\) is said to be Pareto optimal [15] if there is no point in \(\mathbf{R}^n\) such that \(\mathbf{F}(\mathbf{x})\) dominates \(\mathbf{F}(\mathbf{x}')\), where \(\mathbf{F}(\mathbf{x})\) is a vector of solutions called Pareto optimal vector if it contains only Pareto optimal solutions. The set of all Pareto optimal objective vectors forms PF.

2.2 Nondominated sorting genetic algorithm (NSGA-II)

NSGA-II employs nondominated sorting approach for fitness assignment among individual solutions to solve multi-objective optimization problems (MOPs) [15]. The algorithm starts with a random generation of initial population (range of design alternatives for design variables) of N numbers. The current parent population (\(P_t\)) at tth generation is used to produce offspring population (\(Q_t\)) with population size N by using polynomial mutation and simulated binary crossover operator [13]. The two populations \(P_t\) and \(Q_t\) are merged together to form a mixed population \(R_t\) (of population size 2N). The best N numbers are selected from \(R_t\) pool using nondominated sorting approach based on dominance principle. Nondominated solutions are selected from \(R_t\) to fill for a population size of N in next generation. The remaining solutions in next generation pool are selected by employing a crowding distance scheme [13], i.e., the solutions having largest crowding distance values are chosen for next generation to maintain a desired diversity.

3 Hierarchical mutation based genetic algorithm

A framework for hierarchical version of NSGA-II (hNSGA-II) is presented in this section. The proposed framework is based on the logic of conventional NSGA-II. Original NSGA-II makes use of a fast nondominated sorting approach with computational complexity \(O(kN^2)\) [13], where k is the number of objectives and N is the population size. Along with this elitist-preserving approach, NSGA-II uses crowding distance scheme to estimate the density of solutions for any specific PF. However, the formulation of this scheme becomes unstable when two or more solutions share a common fitness. Although NSGA-II keeps track of convergence by giving emphasis on nondominated solutions, there is scope of improvement in this category by employing suitable mutation or crossover strategies. The objective of our proposed approach is to improve the convergence property by incorporating hierarchy in the existing polynomial mutation scheme of NSGA-II. A generic hierarchical approach is presented for generation of child population during mutation operation, i.e., each child individual is formed in local neighborhood of its siblings along with parent population. The hierarchical approach for mutation operation is illustrated in Fig. 1(a). The child population is generated from corresponding parent population in a hierarchical structure, which forms local blocks of manageable size. Let us consider the generation of N local blocks and subsequent formation of one global block combining N local blocks and previous population. Each of the N local blocks is modeled as a tree like structure (having height h) with child population forming different branches. The first prime individual in a local block is called a prime parent. Prime parent generates a set of child individuals in first level (\(h=1\)) and the individuals in first level generate child individuals for second level (\(h=2\)) through mutation operation. Hence, the individuals in first level act as parents for individuals of second level (sub-parents) and so on. In such manner, child population in each local block is generated using polynomial mutation operator and the process continues till maximum height of the tree (\(h_{\max }\)) is achieved. The number of child individuals of prime parent and each sub-parent is decided by the degree of branching d, and a total of \(\beta\) individuals are generated in each local block including the prime parent during each generation. The generated local best solutions in nth local block having \(d=4\) moving towards PF in each level, is shown in Fig. 1(b), where \(n\in N\). Further, all child individuals in a local block along with prime parent are sorted and ranked using nondominated sorting approach. The individuals having highest ranks in each local block are selected as local best individuals. The local best individuals are pulled to global block to generate the offspring population. The offspring population is merged with parent population to generate a population having 2N individuals. These individuals are again sorted and ranked using nondominated sorting approach to select the best individuals for next generation. However, to preserve the diversity during selection, it is necessary to select distinct local best individuals before being pulled to global block without affecting the convergence property. Instead of performing the ranking through traditional crowding distance approach, we propose to perform an indirect exclusive selection by performing hierarchical mutation operation iteratively. For an exclusive selection, at first N local best individuals are sorted in descending order of crowding distance. All local best individuals are checked for unique fitness and the individuals having unique fitness values (\(\varepsilon\)) are retained in the local best pool. The remaining \(N-\varepsilon\) local best individuals are discarded and hierarchical mutation operation is carried out in corresponding \(N-\varepsilon\) local blocks to generate new local best individuals. The process is continued till N local best individuals with distinct fitness values are generated. As the population in each local block is more than one (\(\beta > 1\)), where \(\beta \le \sum _i^h d^i\), the probability of generating local best individuals having distinct fitnesses is more as compare to traditional crowding distance scheme. In this regard, it accounts for the instability of crowding distance scheme in choosing common fitness values. While performing an indirect exclusive selection using hierarchical mutation operation, the number of individuals in each local block also play an important role in convergence of solutions. A detailed theoretical framework of hNSGA-II along with convergence property is described in following subsections.

Fig. 1
figure 1

a General model of hierarchical mutation strategy, b hierarchical growth for \(h = 3\) and \(d = 4\) showing growth towards true PF

figure e
figure f

3.1 Computational complexity

As hNSGA-II is based on the framework of conventional NSGA-II, the algorithm starts by initializing parent population \(P_1\) of first generation. The hierarchical mutation procedure described in Algorithm 1, starts with initializing local block of height h and random generation of degree d for generating a temporary offspring population \(C_t\). Once the number of offsprings in each level d is defined, the algorithm follows polynomial mutation operator to generate offspring population in each local block. This requires \(O(hd^j)\), where \(j\in [1,h]\) number of computations in each block and a total of \(O(hd^h(MN))\) number of computations (as j can go upto h) for generating the local offspring population \(C_t\) with M-dimensional objectives. The nondominated sorting in line 13 of Algorithm 1 having N population size requires \(O(MN^2)\) computations [13]. As the search for unique fitness value among elite local individuals is recursive, it requires \(O(\varepsilon (MN^2))\) computations with M-dimensional objectives. Once the local elite individuals are generated with unique fitness values, they are pulled to the global block, which forms the offspring population (\(Q_t\)) of tth generation. This \(Q_t\) offspring population is combined with \(P_t\) to generate \(R_t\) having population size of 2N. The nondominated sorting described in line 5 of Algorithm 2 having 2N population size requires \(O(M(2N)^2)\) computations [13]. Crowding distance assignment in line 9 requires O(k(2N)log(2N)) computations in the worst-case and quicksort involves O(2Nlog(2N)) computations, which speaks for an overall complexity of \(O(M(\beta N + \varepsilon N^2))\) for hNSGA-II procedure, where \(1 \le \beta \le (\sum _i^h d^i)\). With increase in the height of each block, the runtime of hNSGA-II increases with an advantage of reaching convergence at minimum generations. One of the reasons for this functional advantage is the selection of an elite offspring from a pool of randomly generated individuals spanning alternative search spaces. Further, this ensures the evolution process to be promising as hierarchy in mutation operation facilitates the introduction of new offspring individual from random locations within the feasible space to preserve diversity in each generation.

3.2 Theoretical framework

This section presents a theoretical framework for hNSGA-II based on asymptotic states of a homogeneous Markov chain algorithm. As hNSGA-II is based on NSGA-II, it uses the same basic operators such as selection, crossover and mutation to create next generation from current population. The generation of successive populations can be viewed as a stochastic process with finite state space, and each operation carried out by the operators can be modeled as corresponding stochastic matrices. Further, the probabilistic changes of real parameters within a population using genetic operators can be formulated in the form of transition probability matrix \(P_r\) [16], which can be represented as the product of stochastic matrices,

$$\begin{aligned} P_r = P_s \cdot P_m \cdot P_c, \end{aligned}$$
(2)

where \(P_s\), \(P_m\) and \(P_c\) denote intermediate transitions caused by selection, mutation and crossover operation, respectively. Let us consider each genetic operation to be performed in succession. It is assumed that proportional selection is performed with probability \(P_s\) on an initial population vector \(\bar{x} \in S\) having set of all possible population states S. The conditional probability \(P_{s}(\bar{y} |\bar{x})\) of selecting a solution vector \(\bar{y} \in S\) from initial population state vector \(\bar{x}\in S\) can be represented as [16],

$$\begin{aligned} P_s(\bar{y}|\bar{x}) &= \frac{ \Pi _{k\in S}\ f(y_k)}{(\Sigma _{k \in S}\ f(y_k))^n},\ \ \text{if}\ \bar{y} \subset \bar{x},\ \ \ (\bar{y}, \bar{x}) \in S \nonumber \\ &= 0 \ \ \text{otherwise} \end{aligned}$$
(3)

Mutation operation is performed to maintain the genetic diversity in successive generations by making changes to the states of decision parameters (variables) of current generation. As mutation operation is carried out independently on each individual decision variable by setting a user-defined mutation probability \(p_m\) to lower values (i.e. \(p_m \in \{0,1\}\)), the conditional probability \(P_m (\bar{y}| \bar{x})\) that the new population \(\bar{y} \in S\) resembles the initial population \(\bar{x} \in S\) after altering real values of variables at different decimal positions (mutation operation) can be aggregated to [16],

$$\begin{aligned} P_m(\bar{y}|\bar{x}) = p_m^{\mathrm{poly}(x_k,y_k)} (1-p_m)^{\mathrm{poly}(x_k,y_k)}, \end{aligned}$$
(4)

where \(\mathrm{poly}(x_k,y_k)\) is the polynomial mutation operatorFootnote 1 [17] to generate child individual \(y_k\) from parent \(x_k\). Typically, crossover operation is performed to generate child solution from two or more parent solutions. Here, simulated binary crossover (SBX) operator is used for crossover operation in hNSGA-II because of the classic property to keep difference between child individuals proportionate to their corresponding parent individuals. SBX coefficient provides the flexibility to control generation of child population in search space, i.e. child individuals having high SBX coefficients are more probable to provide solutions close to optimum. Although SBX operator is used with crossover probability \(P_c \in \left[ 0,1\right]\), convergence analysis can be performed even without the knowledge of elements of transition crossover matrix as the choice of any specific operator does not affect the analysis.

The formal description of our proposed approach relies heavily on the hierarchical scheme of genetic evolution. The behavior of this scheme can be interpreted as a tree like homogeneous time bivariate Markov chain \(\{ (K_t, G_t), t\ge 0 \}\), where \(K_t\) represents all individuals in the local block having height \(h=2\) and degree \(d=4\), and \(G_t \in \mathbf{N}, G_t>0\) as shown in Fig. 2. Further, the prime parent is denoted by \(\alpha\) and all other individuals in the local block are denoted by strings of integers in \([1, \beta ]\), where \(\beta \le \sum _i^h d^i\). Moreover, the behavior of the hierarchical mutation scheme implies few restrictions on the Markov chain for proper functionality. All transitions are considered to be of zero probability with the following two exceptions.

  1. 1.

    At each step the process can make one transition to itself \(( K_{t+1} = K_t)\) (no change in parent population after mutation operation) or,

  2. 2.

    Transition to one of its child \(( K_{t+1} = K_{t+s},\) for some \(1\le s\le d)\) (change in parent population after mutation operation).

Fig. 2
figure 2

Markov chain model of hierarchical strategy

Local block of the hNSGA-II framework represents the evolution of child population from parent population using mutation operation. The transition probabilities of a parent to its children and to itself in one level, may be different even though both transitions are from a single individual. In view of this, Markov chain in each local block can be illustrated as the sum of probabilities of all possible sample paths in forward direction starting from prime individual and ending at child individuals of final level. Therefore, the transition probability matrix of mutation can be rewritten as a combination of intermediate transitions described in (4) of each probable states. The transition probability matrix of mutation at level h having degree d can be represented as,

$$\begin{aligned} P_m^{(h)}(\bar{y}|\bar{x}) = \Sigma _{l \in h} \left( \Sigma _{j \in d^l}\ p_m^{\mathrm{poly}^l_j(x_k,y_k)} (1-p_m)^{\mathrm{poly}^l_j(x_k,y_k)}\right) \end{aligned}$$
(5)

It can be pointed out that the hierarchical mutation operator holds the key for robust convergence of hNSGA-II because it exhibits the decision whether the algorithm achieves limiting distribution (strong ergodicity) [18] or not. With the increase in the height h of each local block, the number of randomly generated individuals increase allowing the hierarchical mutation operator to explore additional search points in the feasible region. This proposal of augmented upper bound in mutation operation facilitates sufficient conditions for strong ergodicity.

3.3 Convergence analysis

Considering the search properties of hNSGA-II, various conclusions can be made relating to the sufficient conditions for hNSGA-II converging to a stationary distribution.

Proposition 1

Any state probability distribution \(\xi \in X\) of a homogeneous Markov chain is a stationary distribution in X if it satisfies the equality, \(\xi ^T = \xi ^T P_r\), where \(P_r\) is the transition probability matrix and X is the state space containing all possible Markov states [16].

Theorem 1

For \(p_m >0\) and a strictly positive defined crossover and selection probability \(P_c, P_s\) respectively, h NSGA-II converges in limit towards a stationary distribution \(\xi\) , where \(\xi\) is strictly positive, i.e. \(\xi >0\), i.e.,

$$\begin{aligned} \lim _{r\rightarrow \infty } P_r = \begin{bmatrix} \xi \\ \xi \\ \vdots \\ \xi \end{bmatrix} \end{aligned}$$
(6)

Proof

Let (\(x_t, x_{t+1}\)) be the given arbitrary states at any time point (\(t, t+1\)) respectively before hierarchical mutation. The transition from \(x_t\) to \(x_{t+1}\) in one time step with appropriate probability \(P_m\) can be given by (5). Let us assume the corresponding relative fitness be \(f_{(t,t+1)}\). Depending upon the number of states in each block, there are two possible cases.

Case 1 For \(x_{t+1}=0\), there exists exactly one state with a relative fitness \(f_{(x_t,x_t)} >0\) for population of fixed size N as the state will transit to itself every time. It can be concluded that,

$$\begin{aligned} P_m^h (\bar{x_t}|\bar{x_t}) = p_m^{\mathrm{poly}^l_j(x_t,x_t)} (1-p_m)^{\mathrm{poly}^l_j(x_t,x_t)} > 0 \quad \left( l,j =0 \right) \end{aligned}$$
(7)

Case 2 For \(x_{t+1}>0\),

$$\begin{aligned} P_m^h (\bar{x_{t+1}}|\bar{x_t}) = \Sigma _{l \in h} \left( \Sigma _{j \in d^l}\ p_m^{\mathrm{poly}^l_j(x_{t+1},x_t)} (1-p_m)^{\mathrm{poly}^l_j(x_{t+1},x_t)}\right)> 0 \quad \left( l,j >0\right) \end{aligned}$$
(8)

Therefore, for arbitrary states \((x_t, x_{t+1}) \in X\), there exists strictly positive numbers in \(P_m^h\). As both crossover and selection transitions are strictly positive, the entry in \(P_r\) consists of a product of strictly positive numbers.□

Fig. 3
figure 3

Variation of limiting distribution entropy with local block population \(\beta\)

As the stationary distribution is strictly positive, all states are considered to be associated with finite probability. This means that a limiting behavior may be possible and to showcase the possibility, it is important to evaluate the entropy with respect to increase in population size inside each local block during mutation operation. The usefulness of hierarchical mutation scheme is shown in Fig. 3 by presenting the variations in limiting distribution entropy [16] with respect to \(\beta\). The results illustrate the approximate computed stationary distribution entropy on a test problem (ZDT1) at different selected values of \(\beta\). The values of \(\beta\) are evaluated by considering \(d=4\) and \(1\le h\le 6\). It can be observed from Fig. 3 that the entropy reduces to zero with increase in \(\beta\). Further, it can be suggested that with increase in \(\beta\), it may be possible to approach limiting distribution behavior (i.e., a globally optimal solution with probability one) by making the probability associated with nonoptimal states as small as required. Since mutation operation is carried out for both polynomial and binary coded genetic evolution in such a way to enforce the relative fitness, \(f>0\) [19], it may be possible to parameterize hNSGA-II to reduce bias towards solutions by analyzing ergodic bound of hierarchical mutation probability. The presence of uncertainty in transitions during hierarchical mutation to choose between next state and current state can cause finite number of revisits to already traversed state showing condition of transient states (\(P_m^{(h)}(\bar{y}|\bar{x}) < \infty\)) (Theorem 2). Further, it may account for few states to be recurrent (\(P_m^{(h)}(\bar{y}|\bar{x})= \infty\)) if the state transits to itself in all transitions. In such cases, variations in population size parameter \(\beta\) inside local block of hNSGA-II has important implications for recurrence and transience conditions.

Proposition 2

In any state space X, a state \(x_t\) is recurrent if and only if \(\sum _m^{\infty }P_m^h (\bar{x_{t}}|\bar{x_t}) = \infty\), transient otherwise [19].

Theorem 2

All states in X of a Markov chain are either recurrent or all states in X are transient. Thus, if there is a transition from \(x_t\) to \(x_{t+1}\) and state \(x_t\) is assumed to be recurrent, then \(x_{t+1}\) is also recurrent. Equivalently if a transition occurs from state \(x_t\) to \(x_{t+1}\) and state \(x_t\) is transient, then state \(x_{t+1}\) is also transient. In particular, it can be said that for an irreducible Markov chain, either all states are recurrent or all states are transient [19].

Proof

If there is a transition between \(x_t\) and \(x_{t+1}\) and \(x_t\) is recurrent, then \(x_{t+1}\) must be recurrent because every time \(x_t\) is revisited after some interval with a success probability, \(x_{t+1}\) will also be revisited with a finite time interval. As \(x_t\) is recurrent, it will be visited a number of times with a probable success rate making \(x_{t+1}\) to be recurrent. Eventually, this can be viewed to be a sequence of random walk repeating itself at regular intervals.

3.4 Conditions for ergodicity

Mutation operation plays a prominent role in robust convergence of a simple genetic algorithm. With inclusion of hierarchical mutation operation, the number of individuals participating in genetic evolution process is comparatively more than NSGA-II, which in turn increases the search space of hNSGA-II and ensures a rigid schedule bound for ergodicity. In this section, we present a schedule bound for hierarchical mutation in hNSGA-II that will establish strong ergodicity. There are a number of conditions to showcase the existence of ergodicity in the proposed Markov chain. One of the necessary and sufficient conditions for ergodicity is the existence of a non-trivial measure [20], \(\mu\) satisfying (9).

$$\begin{aligned} \mu _m^{h}(\bar{y}) \ge \int _X \mu _m^h(d\bar{x}) P_m^h (\bar{y}|\bar{x}) \end{aligned}$$
(9)

According to general state space context, it is quite difficult to verify the criterion in (9). Therefore, a Doeblin condition [21] can be established along with the condition in (9) satisfying (10) having a probability measure (\(\mu\)) and a fixed integer \(\delta , 0<\delta \le \beta\), that can account for a schedule bound for ergodicity of the proposed Markov chain model, i.e.,

$$\begin{aligned} P_m^h (\bar{y}|\bar{x}) \le \delta , \quad \forall \bar{x} \in X \end{aligned}$$
(10)

Theorem 3

Following both conditions in (9) and (10), a sufficient condition for the proposed Markov chain to be ergodic is the existence of a non-trivial measurable function g such that [20],

$$\begin{aligned} \int _X P_m^h (d\bar{y}|\bar{x}) g_m^h(\bar{y}) \le g_m^{h}(\bar{x})-1, \quad \forall \bar{x} \notin X \end{aligned}$$
(11)

and for a fixed integer \(\delta > 0\),

$$\begin{aligned} \int _X P_m^h (d\bar{y}|\bar{x}) g_m^h(\bar{y}) = \omega (\bar{x}) \le \delta < \infty , \quad \forall \bar{x} \in X \end{aligned}$$
(12)

Proof

Considering g to be a non-measurable function as described in Theorem 3 and let there is a transition from state \(\bar{x}\) to state \(\bar{y}\) using mutation probability, \(P_m^h\) having height h of a each local block, then at \((t+1)\)th time point,

$$\begin{aligned} g_{m(t+1)}^{h}(\bar{x}) = \int _X P_{m(t)}^h (d\bar{y}|\bar{x}) g_m^h(\bar{y}),\quad t\ge 1 \end{aligned}$$

If there exists an intermediate state space \(\kappa , \kappa ^c \in X\), such that \(X = \kappa \cup \kappa ^c\) then

$$\begin{aligned} g_{m(t+2)}^{h}(\bar{x})&= \int _X P_{m(t)}^h (d\bar{z}|\bar{x}) \left[ \int _X P_{m(t)}^h (d\bar{y}|\bar{z}) g_m^h(\bar{y})\right] \nonumber \\&\le \int _{\kappa } P_{m(t)}^h (d\bar{z}|\bar{x}) \omega (\bar{z}) + \int _{\kappa ^c} P_{m(t)}^h (d\bar{z}|\bar{x})[ g_m^h(\bar{z})-1] \nonumber \\&\le \int _{\kappa } P_{m(t)}^h (d\bar{z}|\bar{x}) [\omega (\bar{z}) +1] \\ &\quad + \int _{\kappa ^c} P_{m(t)}^h (d\bar{z}|\bar{x}) g_m^h(\bar{z})- P_{m(t)}^h (X|\bar{x}) \\&\le (\delta + 1) P_{m(t)}^h (\kappa |\bar{x}) + g_{m(t+1)}^{h}(\bar{x}) -1 \end{aligned}$$
(13)

Iterating (13) and dividing by t, we have

$$\begin{aligned} \frac{1}{t} g_{m(t+2)}^{h}(\bar{x})&\le (\delta + 1) \left[ \frac{1}{t}\Sigma _{i=1}^t P_{m(i)}^h (\kappa |\bar{x}) \right] + \frac{1}{t} g_{m(2)}^{h}(\bar{x}) - \frac{1}{t} \Sigma _{i=1}^t 1 \nonumber \\&\le (\delta + 1) \left[ \frac{1}{t}\Sigma _{i=1}^t P_{m(i)}^h (\kappa |\bar{x}) \right] + \frac{1}{t} g_{m(2)}^{h}(\bar{x}) - 1 \end{aligned}$$
(14)

As per [20], for \(\forall x\),

$$\begin{aligned} \lim _{t\rightarrow \infty } \frac{1}{t}\Sigma _{i=1}^t P_{m(i)}^h (\kappa |\bar{x}) = \pi (\kappa |\bar{x}) \end{aligned}$$
(15)

exists and \(\pi (\kappa |\bar{x})\) is finite when \(\kappa \in X\). With increase in time point as \(t\rightarrow \infty\) and by the non-negativity of \(g_{m(t)}^h\) and finiteness of \(g_{m(2)}^h\) in (14), we have

$$\begin{aligned} \pi (\kappa |\bar{x})\ge \frac{1}{\delta +1} \ge \frac{1}{\beta } > 0 \end{aligned}$$
(16)

where for a finite population size \(\beta\), there exists a schedule bound for ergodicity of proposed Markov chain model.

4 Performance analysis

4.1 Parametric study

In this subsection, we perform a parametric study for population size parameter \(\beta\) of hierarchical polynomial mutation operator. As the performance of proposed hierarchical scheme can be characterized for different population sizes of each local block (\(\beta\)), the performance can change with change in \(\beta\). For a constant degree d, height h can be increased to postulate variations in both convergence and diversity of Pareto optimal solutions at different values of \(\beta\) for a number of generations. To measure the performance in terms of convergence towards optimal Pareto front and diversity of solutions along Pareto front, three different quality metrics are evaluated, i.e. Generational Distance (GD), Inverted Generational Distance (IGD) and Hypervolume (HV). GD metric [22] is used to measure minimum Euclidean distance between generated final individuals and sample points on true Pareto front. Smaller value of GD indicates a better convergence towards true Pareto front. On the other hand, both IGD metric [22] and HV metric [23] are a measure of both convergence and diversity of solutions along true Pareto front on a single scale. If the IGD value is close to zero, it means the solutions have better convergence and diversity, and HV is a fair measure of the maximum area covered by the nondominated solutions with respect to a reference point. In view of this, several experiments are carried out on ZDT1 test function for a number of generations and it can be observed from Fig. 4 that with increase in \(\beta\), solution (optimal PF points) attempts to attain smaller (best) GD (Fig. 4a) and IGD (Fig. 4b) values and higher HV (Fig. 4c) values at fewer generations. One of the reasons for these best values at fewer generations for higher \(\beta\) value can be due to the increase in number of offsprings in each local block with increase in \(\beta\). It can be seen from Fig. 4(c) that with increase in number of generations, HV values also increase. However, the HV values remain unchanged after 500 number of generations (for \(\beta = 256\) and 700 generations for other \(\beta\) values). As intermediate traces of PF exist at fewer generations, the population has to be evolved over sufficient number of generations (500 generations for \(\beta =256\)) to account for the convergence and diversity metric to saturate during function evaluations. With more population size, the diversity can be preserved by exploring extra search space by combining all new offsprings of each local block. Another reason for best values at fewer generations can be the reduction in selection pressure by generating more offsprings through proposed hierarchical mutation strategy. With decrease in selection pressure, the convergence issues are addressed to a great extent by mitigating the instability in crowding distance scheme. However, \(\beta\) cannot be increased to any arbitrary value. As \(\beta\) accounts for the variations in limiting distribution entropy of mutation operation, it should be kept within finite limits (\(0<\beta <N\)) for better performance of hNSGA-II (as described in Sect. 3.4).

Fig. 4
figure 4

Performance of hNSGA-II on ZDT1 test function at different generations for various population sizes of local block (\(\beta\)), a mean GD values, b mean IGD values, c mean HV values

4.2 Comparison with other mutation strategies

As mutation operator operates on an individual independent of other population members, it plays an important role in making overall search efficient [24]. Therefore, it is necessary to investigate the effectiveness of proposed hierarchical mutation operator by comparing the performance with other mutation operators. Hierarchical scheme for six different mutation operators (Adaptive Levy mutation [25], Cauchy mutation [26, 27], uniform mutation [28], nonuniform mutation[28], Gaussian mutation [29] and Adaptive-mean mutation [26]) are implemented and also incorporated to study the search efficiency of hNSGA-II framework. This study is extensive in evaluating the performance of proposed hierarchical polynomial mutation operator. Further, a total 30 decision variables (\(p= 30\)) are considered during evaluation and each variable is bounded within \(x_i \in [0,1],\) where \(1\le i\le 30\). Crossover probability is kept at 0.9 and mutation probability is set to 1 / p for all mutation schemes. Both crossover and mutation distribution indices are set to 20 during the entire process. As various mutation operators are employed within hNSGA-II framework, different parameters are addressed each time a mutation operator is exercised. During Cauchy mutation, a Cauchy density function [27], \(f_{cauchy} = \frac{1}{\pi }\frac{t}{t^2 + x^2}\), where scale parameter t is set to 1 during operation. Adaptive Levy mutation scheme is employed by using parameter \(\alpha\) fixed at 1.7 for all experiments. Box-Muller transform is adopted to generate Gaussian distribution from uniformly distributed numbers during Gaussian mutation operation. Population size of each local block, \(\beta\) is kept at 256 for achieving hierarchy (as described in Sect. 4.1) during all operations. All mutation operators exercised using hierarchical scheme are compared based on convergence and diversity of solutions (mean GD values, mean IGD values and mean HV values), while 200 population individuals are allowed to evolve over 100 generations on ZDT1 test function. For each mutation operation, 30 independent runs are performed and the results are shown in Fig. 5. It can be observed from Fig. 5(a, b) that proposed hierarchical polynomial mutation operator performs better than other mutation operators as corresponding GD and IGD values are minimum. Further, it is observed from Fig. 5(c) that the proposed hierarchical polynomial mutation operator has larger HV value as compared to other mutation operators. As the proposed hierarchical scheme is performed on top of polynomial mutation operator, the inherent probability distribution of creating an offspring population is similar to polynomial mutation operator, which is instrumental in the convergence of Pareto optimal solutions. Moreover, the implicit generation of distinct fitness values in each generation favors the preservation of diversity among Pareto optimal solutions.

Fig. 5
figure 5

Performance analysis of several mutation strategies on ZDT1 test function showcasing mean, a GD values, b IGD values, c HV values at different generations (100–1000)

4.3 Performance of hNSGA-II on multi-objective test functions

In this subsection, different sets of multi-objective test functions are solved using hierarchical polynomial mutation based hNSGA-II framework. These test functions have been used as standard benchmarks in literature. To demonstrate the applicability of hNSGA-II, a total of twelve test problems with two different groups of benchmarks are considered (both constrained and unconstrained test problems). With the aim of giving a complete overview of performance of hNSGA-II, several standard optimization algorithms (NSGA-II-DE [30], MOEA/D-DE [30], NSGA-II [13], \(\epsilon\)-MOEA [17], NSGA-IIa [31] and ssNSGA-II [31]) are considered for evaluation on all unconstrained benchmarks (ZDT1, ZDT2, ZDT3, ZDT4, ZDT6, Schaffer, Fonseca, and Kursawe test functions [13]). For the purpose of comparison, we have used the authors implementation of all algorithms. However, the implementation framework of NSGA-II-DE and MOEA/D-DE does not have any provision of constraint handling management. Therefore, during showcase of results, we are forced to leave the spaces blank during the analysis of constrained benchmarks (Binh2, ConstrEx, Ozyczka2 and Tanaka test functions [13]). During evaluation of all algorithms, a population size of 200 is considered to evolve over 500 generations. For all NSGA-II and MOEA variants (i.e., NSGA-II, NSGA-IIa, ssNSGA-II, NSGA-II-DE, MOEA/D-DE and \(\epsilon\)-MOEA) and proposed hNSGA-II approach, the crossover probability is kept at 0.9. Mutation probability is set to 1 / p for all algorithms, where p is the number of decision variables. The crossover and mutation distribution indices are set to 20 as commonly used in literature. The population size parameter (\(\beta\)) of each local block is set to 256 to have a reasonable amount of offspring population to preserve sufficient diversity across the obtained PF. Table 1 represents the mean and standard deviations of GD values for hNSGA-II along with other standard optimization algorithms based on 30 independent runs. It can be seen from Table 1 that GD values are better for seven out of twelve benchmarks (ZDT1, ZDT2, ZDT3, Fonseca, Kursawe, Ozyczka2 and Tanaka [13]), while the performance is inferior for both ZDT4, ZDT6, Schaffer, Binh2 and ConstrEx benchmarks. This indicates that the resulting Pareto fronts from hNSGA-II are closer to the optimal Pareto fronts as compared to those computed by other standard optimization algorithms. From Table 2, it can be seen that hNSGA-II obtains lowest (best) values for the IGD metric in seven out of twelve benchmarks (ZDT1, ZDT2, ZDT3, Schaffer, Fonseca, Kursawe and Tanaka). This points out that the obtained Pareto fronts of multi-objective test problems are closer to true Pareto fronts and the non-dominated solutions are better spread for hNSGA-II than other algorithms. Again to showcase the performance of hNSGA-II as a measure of both convergence and diversity in a concise manner, HV metric is evaluated for all benchmarks (as described in Sect. 4.1). It can be seen from Table 3 that hNSGA-II obtains best (highest) HV values in six out of twelve benchmarks. In view of the results, hNSGA-II shows better performance than other algorithms (seven out of twelve, seven out of twelve, and six out of twelve for GD, IGD and HV metric respectively). The usage of hierarchical scheme in polynomial mutation to generate an elite offspring appears to be fruitful in order to achieve better convergence and diversity of solutions along true PF. However, the reason for better performance can be attributed to the generation of individuals having unique fitness values during mutation operation of hNSGA-II.

Table 1 The GD statistics (mean and SD values) based on 30 independent runs of hNSGA-II for different multi-objective test problems keeping \(\beta =256\)
Table 2 The IGD statistics (mean and SD values) based on 30 independent runs of hNSGA-II for different multi-objective test problems keeping \(\beta =256\)
Table 3 The HV statistics (mean and SD values) based on 30 independent runs of hNSGA-II for different multi-objective test problems keeping \(\beta =256\)

5 Application to analog/RF circuit optimization

At the lowest level of design hierarchy, the process of circuit synthesis comes down to the step of circuit optimization that makes use of sizing principle and biasing performances of all devices in the circuit to meet certain design constraints. Once an appropriate topology is selected during design process, certain performance specifications are represented as objective functions by formulating a set of physical equations that relate the device characteristics to design parameters. However, it is difficult to solve these equations explicitly under a set of design constraints. Therefore, it is necessary to employ numerical optimization techniques to implicitly analyze these equations while optimizing the circuit under certain design constraints.

The design of analog and RF circuits has become complex with the advent of deep submicron technology. It has imposed a discontinuity on the seer practice of knowledge-based sizing [32] which has been a popular choice of design methodology since the beginning of semiconductor design and manufacturing. However, with increase in integration density and complexity, the use of simple models designed using this methodology will lead to false design. Therefore, new approaches [33, 34] are presented in literature to handle the design complexity of integrated circuit chip. These approaches appear to translate the problem of circuit design into a function minimization or maximization problem, which can be solved through numerical techniques [33]. Mostly, it starts with the development of an efficient performance evaluator framework, which evaluates the minimization (maximization) function [33]. This function is created based on the behavior of circuit topology [33, 35, 36] and they rely on the analytical equation formulation. The efficiency of the evaluator lies in the efficient design of circuit equations. Although a number of deterministic algorithms have been presented to perform circuit design, the assessment of initial design point requires prior knowledge or additional design equation formulation. This favours the usage of indeterministic approaches, like use of nature-inspired algorithms in circuit design. A number of efforts on nature-inspired algorithms have been made in literature to capture the design space boundaries of analog and RF circuits. Genetic algorithm (GA) is employed as one of the popular optimization routines for optimization of analog and RF circuits in both industry and academia [4, 37]. Several variants of GAs are also being employed to design CMOS operational amplifier circuits with constrained topologies. Genetic programming is used as an analog synthesis tool in several filter designs and circuit synthesis problems [4]. With the advent of new and complex circuits, it is necessary to come up with new design strategies using efficient techniques. Therefore, many nature-inspired algorithms like, particle swarm optimization (PSO) [6], simulated annealing (SA) [36], ant colony optimization (ACO) [38], etc. are also suggested as one of the design space boundary exploration techniques for circuit optimization. These techniques are employed to design efficient recursive digital filters with infinite impulse response [39] and for finite impulse response (FIR) filters [40, 41]. Besides these research efforts, several commercial design tools have also been introduced for circuit sizing in the past few years, such as NeoCircuit [42], Genius’s ADA [43], Barcelona Design [44], etc. However, these techniques are employed to optimize a single cost function instead of an optimally located space spanned by a number of cost functions (multi-objective optimization). As circuit modeling is often affected by uncertainties and errors at many levels, it is necessary to analyze the trade-offs among performance specifications to avoid any measurement errors and presence of inconsistent solutions during design. To overcome these problems, several multi-objective techniques are presented in literature with emphasis on nature-inspired algorithms. A hybrid combination of SA and PSO is also suggested in [45] to find solutions to cope with design specifications at circuit-level and exploration of trade-offs of different analog circuits. Among all presented multi-objective techniques, NSGA-II [13] is employed as a leading optimizer in various analog and RF IC synthesis tools [46, 47].

In this section, the proposed hNSGA-II method is employed to optimize the performance of circuit. All circuit design parameters are given as input bounds to optimize two contradicting design costs which are realized by evaluating a Pareto front. Here, an analog circuit (two-stage operational transconductance amplifier) and an RF circuit (low noise amplifier) are taken as case studies to showcase the performance and applicability of proposed hNSGA-II method by optimizing different design costs subject to the set of design constraints.

5.1 Case study 1: two-stage operational amplifier

Operational amplifiers are depicted as one of the fundamental components of analog circuit due to flexible controllability. A case study of a two stage operational amplifier has been considered as shown in Fig. 6 to showcase the effectiveness of proposed hNSGA-II method. The design problem is formulated as a multi-objective optimization problem to analyze tradeoff between maximum gain and minimum power consumption. The closed-form expressions are approximated according to the design reported in [48]. Subsequently, the details on design variables are listed in Table 4 and all design constraints are summarized in Table 5. The supply voltage is kept constant at 1.8 V for suitable operation of transistors. Before starting the process of optimization, channel lengths of 0.18 \(\upmu \text{m}\) are selected and inversion coefficient is kept at below unity (\(\simeq 0.8\)) for M1 and M2 transistors.

Fig. 6
figure 6

Two-stage operational amplifier circuit [48]

However, other transistors are allowed to operate at the strong inversion side to minimize their \(V_{DS}^{\mathrm{sat}}\) or drain thermal-noise current inputs. This suggests that the gain of the amplifier can be maximized at an expense of power consumption. During optimization of two-stage operational amplifier, 200 populations are allowed to evolve over 500 generations till all design constraints are met for efficient realization of proposed approach. Final values of all constraints after optimization of gain and power of the amplifier are listed in Table 5. Figure 7 shows the distribution of final nondominated solutions obtained by several optimization techniques including proposed hNSGA-II approach for achieving maximum gain and minimum power consumption of two-stage operational amplifier. It can be concluded from the distribution of solutions that NSGA-II with hierarchical mutation strategy can produce better approximations of final nondominated solutions over a finite search space than other optimization techniques. Further, it can be seen from Fig. 7(e) that using hNSGA-II approach, voltage gain maximizes sub-linearly for a fixed biasing at an expense of core 256 mW power consumption under several design constraints (see Table 5). This sub-linear plot of Pareto front (PF) can be distinguished by two regions. In the first region, where gain is less than \(67.5\,\text{dB}\), gain increases with an increase in power consumption at a slower pace. In second region, power consumption increases significantly with a small increase in gain. However, it is not desirable to achieve such higher gain at an expense of significant power consumption.

Table 4 Two-stage operational amplifier variables and ranges
Table 5 Two-stage operational amplifier constraints and specifications
Fig. 7
figure 7

Plot of nondominated fronts for gain (max) and power consumption (min) of two-stage operational transconductance amplifier by a NSGA-II, b \(\epsilon\)-MOEA, c NSGA-IIa, d ssNSGA-II, e hNSGA-II

5.2 Case study 2: low noise amplifier (LNA)

The design of LNA is a relatively complex process as there exists tradeoffs among different performance specifications, such as noise figure (NF), transducer power gain, power consumption, area etc. Here, a case study of source degenerate cascode LNA is taken as shown in Fig. 8 to demonstrate the applicability of proposed hNSGA-II method. LNA is designed in \(0.18~\upmu \text{m}\) CMOS technology with 1.8 V power supply centered around 2.4 GHz frequency using Cadence Virtuoso [49]. Although the design of LNA is subjected to keep a minimum NF (\(\le 2.5\)), a power consumption of 8 mW is maintained with a moderately high quality factor (\(Q\in [3,5]\)) and a drain current of 1 mA excluding the bias reference circuit. Further, it is crucial to maintain a low NF without sacrificing much gain during the design process. Otherwise noise gets transmitted to other blocks of receiver system. In view of this, design problem is transformed into a multi-objective constrained optimization problem with an objective to minimize NF and to maximize gainFootnote 2 (S21 in dB) of LNA. The closed-form expressions of NF and other design constraints are based on the design as reported in [50]. Table 6 summarizes the ranges of variables specified during optimization of LNA. The details of different specifications and constraints during optimization process, and the results obtained after optimization are shown in Table 7.

Fig. 8
figure 8

Low noise amplifier circuit [50]

Table 6 LNA variables and ranges
Table 7 LNA constraints and specifications

In order to have a qualitative study of LNA design, simulation results of S-parameters and NF at different frequencies are plotted and shown in Fig. 9. It shows that NF and gain are minimum and maximum respectively at 2.4 GHz center frequency. To start the process of optimization, initial values are assigned to design parameters from first-cut design of LNA. A total of 200 populations are evolved over 500 generations to ensure an optimal PF during LNA optimization. Figure 10 plots the distributions of final nondominated solutions found in several optimization techniques including proposed hNSGA-II for LNA optimization. It can be seen from Fig. 10 that the distribution of nondominated solutions in the PF is quiet nonuniform, i.e., solutions are crowded in a corner of the PF rather being distributed in the objective space or the distribution of solutions is not continuous along the PF, for each optimization technique except hNSGA-II approach. Figure 10(e) represents an optimal PF for minimum NF and maximum gain (S21 parameter) of LNA generated using hNSGA-II approach. It can be seen from this plot that with increase in gain below 15 dB, NF increases slowly and after that NF starts to increase significantly at high gain values. LNAs having high NF deteriorates the performance of a receiver system significantly by transmitting input signals coupled with large noise to other blocks. Therefore, the feasible region of operation is to operate at a minimum noise figure of 0.67 dB at a moderate gain of 13.01 dB with a power consumption of less than 8 mW.

Fig. 9
figure 9

Different S-parameters and noise figure of LNA

Fig. 10
figure 10

Plot of nondominated fronts for gain (max) and noise figure (min) of low noise amplifier by a NSGA-II, b \(\epsilon\)-MOEA, c NSGA-IIa, d ssNSGA-II, e hNSGA-II

5.3 Performance evaluation

To measure the performance in terms of convergence towards optimal Pareto front and diversity or spread of solutions along Pareto front, Hypervolume (HV) metric [23] is evaluated for all peer algorithms including proposed hNSGA-II algorithm. Both median and interquartile range (IQR) values of HV metric obtained after 25 independent runs are listed in Table 8 (IQR is reported within braces).

Table 8 Median and IQR of the HV metric for analog/RF circuit optimization problems

As HV is a fair measure of the maximum area covered by the nondominated solutions, Pareto fronts having higher HV values represent better quality of optimized solutions in terms of both convergence and diversity of solutions. It can be observed from Table 8 that the proposed hNSGA-II algorithm achieves superior HV values in both test circuits with statistical confidence (‘\(+\)’ symbol denotes a confidence level of \(95\%\) [51]).

6 Conclusion

This paper presents a multi-objective framework based on hNSGA-II and its application in optimization of analog and RF circuits. A new hierarchical scheme in polynomial mutation operation is employed in this framework to produce random generation of improved individuals to enhance the possibility of independent local search in each generation. The stochastic behavior of proposed framework is demonstrated by developing a Markov chain model with bounded ergodicity. Superiority of hierarchical scheme in polynomial mutation operation is shown by comparing the performance with six other mutation operators. Applicability of proposed hNSGA-II method is shown by carrying out extensive simulations on several multi-objective test problems, and the effectiveness of this method is verified by generating optimal Pareto fronts for a two-stage operational amplifier and a cascode low noise amplifier with inductive source degeneration. Although simulation results support improvement in the performance of hNSGA-II, there is still scope of improvement in both convergence and diversity preservation of Pareto optimal solutions by preserving unique fitness values in each generation. A comprehensive study of hierarchical scheme with real and binary mutation operators can be performed in future to extend the application of proposed method to analyze multi-objective optimization problems having large scale decision variables. Further, the hierarchical scheme can be subjected to parallelization by using efficient parallel strategies over multicore and manycore platforms to reduce the number of function evaluations during optimization process.