Keywords

12.1 Introduction

Adaptive equalization [1] plays an important role in the high-speed digital transmission to remove and recover the problem of inter-symbol interference (ISI). The adaptive algorithms [2] such as steepest descent, least mean square (LMS), recursive least square (RLS), affine projection algorithm (APA), and their variants [3] reported in literature have the chance of getting trapped in local minima [4,5,6] while optimizing the equalizer weights. The performance of these algorithms further degraded in nonlinear channel conditions [6]. To overcome these problems, different derivative-free optimization algorithms are proposed, whereas PSO is one among them. For solving optimization algorithms, PSO is proven as an efficient method and was applied successfully in the area of adaptive equalization [6]. PSO stays as one of the best algorithms for channel equalization in the recent years [5,6,7]. And also it provides minimum mean square error (MSE) compared to genetic algorithms used in the channel equalization [6].

From the first introduction of PSO [8], several variants [9,10,11,12,13,14,15,16] are proposed. The inertia weight parameter ‘w’ is the first modification found in literature which plays a major role in convergence and improves the simulation time. Initially, Shi and Eberhart introduced inertia weight [10]. In their work, a range of constant w values are used and found that PSO shows a weak exploration for large w values, i.e., w > 1.2, and it tends to trap in local optima with w < 0.8. When w is in the range [0.8, 1.2], PSO shows the global optimum in least average number of iterations. A random value is selected to track the optima in [11], as given in Eq. (1.1).

$$ w = 0.5 + \frac{{{\text{rand}}()}}{2} $$
(12.1)

where rand() is a random value between [0, 1].

Time-varying inertia weight is a common approach used in PSO, which determines the inertia weight based on the current and total iteration. Time-decreasing inertia weight methods are used in literature to improve the convergence rate. A linearly decreasing inertia weight has been used for adaptive equalization in [6], based on the update law:

$$ w_{n} = \left[ {\left( {w_{i} - w_{f} } \right)\left( {m - n} \right)/\left( {m - 1} \right)} \right] + w_{f} $$
(12.2)

where w i is the initial weight, w f the is maximum weight, ‘m’ is the maximum iteration value, and ‘n’ is the current iteration index. A nonlinear decreasing inertia weight is proposed by Chatterjee and Siarry [12] based on the equation:

$$ w_{n} = \left[ {\left( {w_{i} - w_{f} } \right)\left( {m - n} \right)^{np} /\left( {m - 1} \right)^{np} } \right] + w_{f} $$
(12.3)

where np is the nonlinear modulation index. With different values of np, inertia weight gives different variations from w i to w f .

Feng et al. [13, 14] proposed another modification as given in Eq. (1.4).

$$ w_{n} = \left[ {\left( {w_{i} - w_{f} } \right)\left( {m - n} \right)/m} \right] + w_{f} * z $$
(12.4)

where z = 4z(1 − z). Initially, z is the random value between (0, 1).

Lei et al. [15] choose a fuzzy complement function as inertia weight defined by Eq. (1.5).

$$ w = \frac{1 - \beta }{1 - s\beta } $$
(12.5)

where β is n/m and s is greater than −1.

The other nonlinear approach [16] is given as:

$$ w = \left( {\frac{2}{n}} \right)^{{^{0.3} }} $$
(12.6)

Zheng et al. [17] proposed an increasing inertia weight. Similarly, Jiao et al. [18] proposed another nonlinear increasing inertia weight as in Eq. (1.7),

$$ w = w_{\text{initial}} \times u^{n} $$
(12.7)

Here w initial is usually between [0, 1] and u is [1.0001, 1.005]. As per [18], u is set as 1.0002.

This paper analyzes the above existing modifications and proposes novel time-varying inertia weight methods for adaptive equalization with minimum mean square error. The following section describes the methodology used in adaptive equalization. Section 12.3 explains the structure and training of the equalizer filter using PSO. Section 12.4 discusses the results.

12.2 Methodology

Figure 12.1 depicts a basic block diagram used in adaptive equalization [1].

Fig. 12.1
figure 1

Block diagram of digital communication system

The equalizer input is the convolution sum of the random input Bernoulli sequence {x(n)} = ±1 (with zero mean and variance 1) and channel model. The channel output is added with the random additive white Gaussian noise (AWGN). The noise sequence has zero mean and variance 0.001.

Initially, the channel model is assumed as a linear channel [2] with three paths as given in Eq. (12.8).

$$ \begin{array}{*{20}l} {h_{n} = \frac{1}{2}\left[ {1 + \cos \left( {\frac{2\pi }{W}\left( {n - 2} \right)} \right)} \right],} \hfill & {n = 1,2,3} \hfill \\ {h_{n} = 0} \hfill & {\text{otherwise}} \hfill \\ \end{array} $$
(12.8)

The factor W controls the amount of distortion. The effect of amplitude distortion is analyzed and given in Table 12.3. The effect of nonlinearities generated in the transmitter is modeled as three different nonlinear equations in (12.9), (12.10), and (12.11). The nonlinearity is introduced by relating output y(n) and input x(n) as in [19],

$$ {\text{Channel}}\;1{:}\,y(n) = \tanh (x(n)) $$
(12.9)
$$ {\text{Channel}}\;2{:}\,y(n) = x(n) + 0.2x^{2} (n) - 0.1x^{3} (n) $$
(12.10)
$$ {\text{Channel}}\;3{:}\,y(n) = x(n) + 0.2x^{2} (n) - 0.1x^{3} (n) + 0.5\cos (\pi x(n)) $$
(12.11)

The nonlinear effect is introduced in the input x(n) and it is denoted as y(n). The channel output is represented as

$$ r(n) = d_{\text{distorted}} (n) + v_{n} $$

Here d distorted(n) is the distorted version of the desired signal. The distortion is introduced by applying the input signal through any one of the nonlinear equations (Eqs. (12.9), (12.10), and (12.11)) and then convolved with the linear channel in Eq. (12.8).

$$ d_{\text{distorted}} (n) = y(n) * h_{n} $$

The signal can be distorted by bandwidth limitation, multipath effect, and the nonlinearities introduced in the transmitter. v n is the noise component modeled as white Gaussian noise with variance \( \sigma_{n}^{2} \).

The error e(n) is calculated as

$$ e\left( n \right) = d\left( n \right) - r\left( n \right) $$

The adaptive algorithm updates the equalizer weights iteratively to minimize e 2(n). Since e 2(n) is positive and gives the instantaneous power, selected as cost(fitness) function.

12.3 System model

The system model [1] used for equalization is the simple linear transversal equalizer as in Fig. 12.2 and decision feedback equalizer shown in Fig. 12.3.

Fig. 12.2
figure 2

Linear transversal equalizer structure

Fig. 12.3
figure 3

Decision feedback equalizer structure

12.3.1 Linear Transversal Equalizer (LTE)

In this type of structure, the present and old received samples r(tkT) are appropriately weighted by the coefficients c q and added to generate the output. The weights are trained to optimum value using adaptive algorithms. The output Z k becomes

$$ Z_{k} = \sum\limits_{q = 0}^{N - 1} {c_{q} } r(t - qT) $$
(12.12)

12.3.2 Decision Feedback Equalizer (DFE)

This type of nonlinear equalizer uses a forward and feedback filter. The forward and feedback filter outputs are summed to find the output of the equalizer. The forward filter output is sent back via the feedback filter. The ISI is canceled by deducting past symbol values from the equalizer output. The output of DFE is calculated as in Eq. (12.13)

$$ Z_{k} = \sum\limits_{q = 0}^{N - 1} {c_{q} } r(t - qT) + \sum\limits_{i = 1}^{m} {b_{i} } X_{k - m} $$
(12.13)

12.3.3 Basic PSO

PSO [8] starts by initializing particles in the random search space and then considers the social and cognitive behavior of the particles. The candidate solutions called ‘particle’ move around the n-dimensional search space with a velocity, which is adjusted based on its own experience and its neighbors’ experiences. The own experience of a particle is denoted as P best; all the particles’ experience is denoted as G best. A new G best is generated in each update process, making other particles to fly toward G best. The new velocity and position are calculated as:

$$ V_{i} \left( {t + 1} \right) = w * V_{i} \left( t \right) + ac_{1} * {\text{rand}}_{1} * \left( {P_{\text{best}} - X_{i} } \right) + ac_{2} * {\text{rand}}_{2} * \left( {G_{\text{best}} - X_{i} } \right) $$
(12.14)
$$ c_{i} \left( {t + 1} \right) = c_{i} \left( t \right) + V_{i} \left( {t + 1} \right) $$
(12.15)

where ac 1 and ac 2 are positive constants, called as cognitive and social acceleration coefficients, respectively. The two random functions rand1 and rand2 are in the interval [0, 1]. V i (t) and c i (t) are velocity and position of particle i, respectively, in tth iteration. Location of the best solution (best tap weights) for particle i is P best and G best represents the best solution among all particles. In Eq. (12.14), w is the inertia weight, which controls the local search and global search.

12.3.4 Training by PSO

The PSO-based equalizer [6] is optimized the tap weights based on the following steps:

For LTE:

  • T number of tap weights is assigned for equalizer.

  • ws’ samples of data are passed from channel output (distorted signal) to equalizer which generates ws numbers of estimated samples.

  • Error is estimated by comparing delayed version of each input sample with equalizer output

  • The mean square error function of each particle P is

    $$ {\text{MSE}}(P) = \frac{{\sum\nolimits_{i = 1}^{K} {e_{i}^{2} } }}{ws} $$
  • Fitness value MSE (P) is minimized using PSO-based optimization.

  • If the MSE of a particle is less than its previous value, then term it as current local best value and its corresponding weight values as P best

  • The minimum of MSE of all particles in every iteration is taken as global best value.

  • If the global best value is better than the previous one, select the corresponding tap weights to G best.

  • Calculated the change in position (Tap weights) of each particle using Eq. (12.14)

  • Moved each particle (Tap weights) c k in Eq. (12.12) to new position by Eq. (12.14).

  • Repeated the above steps for the number of iterations specified or stopped when the algorithm converges to an optimum value with least MSE value.

For DFE:

  • The coefficients are initialized randomly for forward and feedback filter.

  • In the first iteration, only forward filter is active and after calculating the error, the output of the forward filter is fed back through feedback filter.

  • The mean square error is calculated by subtracting the output of forward and feedback filters.

  • The forward and feedback filter coefficients c k and b i in Eq. (12.13) are updated based on Eqs. (12.14) and (12.15).

12.3.5 Proposed Strategies

In PSO, higher values of inertia weight enhance global search while smaller values improve local search. Generally, the inertia weight decreases linearly from 1 to 0 over the entire iterations. The inertia value is high at first, which allows all particles to move freely in the search space during the initial steps, while it decreases over time in the later steps. The later search requires a low inertia value to gradually shift the search from global search to local search. The speed of convergence and minimum MSE is improved when the inertia weight is suddenly changed from high to low after a particular iteration. Hence, the inertia weight is updated in three different ways as given in Eqs. (12.16), (12.17), and (12.18).

$$ iw_{1} = \eta $$
(12.16)
$$ iw_{2} = \left\{ {\begin{array}{*{20}l} \eta \hfill & {{\text{if}}\,n \le {\text{N}}} \hfill \\ {\eta / 2} \hfill & {{\text{if}}\,n > } \hfill \\ \end{array} } \right. $$
(12.17)
$$ iw_{3} = \left\{ {\begin{array}{*{20}l} {\sqrt \eta } \hfill & {{\text{if}}\,n \le {\text{N}}} \hfill \\ {\sqrt {\eta / 2} } \hfill & {{\text{if}}\,n > {\text{N}}} \hfill \\ \end{array} } \right. $$
(12.18)

Figure 12.4 shows that, the inertia weight iw 1 decreases linearly from 1 to 0, whereas in the case of iw 2 and iw 3 the inertia weight decreases linearly until N iteration after that there is a sudden reduction in inertia weight. The common factor used in all time-varying inertia weight algorithms is η = (mn)/m, where m denotes maximum iteration and n denotes current iteration. This factor η changes linearly from 1 to 0. If Eq. (12.14) is modified with a decreasing control function, it gives an effective time-varying inertia weight strategy as shown in Eqs. (12.17) and (12.18). The term N in Eqs. (12.17) and (12.18) is the intermediate iteration value used to reduce the inertia weight suddenly after Nth iteration. This reduction produces optimum performance compared to existing inertia weight-modified methods in terms of convergence speed and MSE.

Fig. 12.4
figure 4

Proposed inertia weight strategies

12.4 Simulation Results

The simulations are performed in MATLAB 7.7 version. The parameters are set as amplitude distortion W = 2.9 which gives an eigenvalue spread 6, particle size P = 40, input window size ws = 20, the acceleration coefficients ac 1 = 1 and ac 2 = 1, tap weights T = 7, and the initial and final inertia weights w i  = 0.9 and w f  = 0.3, respectively. The term np used in Eq. (12.3) is taken as 0.7 and w initial in Eq. (12.7) is selected as 0.5. The parameters considered above are minimum values to minimize the computational complexity in the algorithm. If the parameter values such as P, ws, ac 1 and ac 2 are increased, it can improve the performance and the same is analyzed in Sect. 4.1.

The PSO variants considered are PSO-LDIW [9], PSO-Chatterjee [17], PSO-Feng [18, 19], PSO-Lei, PSO-Zheng and PSO-Jiao. The proposed techniques, PSO with time-varying inertia weights, PSO-TVW, PSO-TVW1, PSO-TVW2, and PSO-TVW3 based on Eqs. (12.14), (12.15), and (12.16), are compared with the above variants in Fig. 12.5. Among all it is seen that the PSO variant PSO-LDIW and PSO-Chatterjee gives the best MSE value with convergence in 80th iteration. PSO variant PSO-LDIW is compared with proposed techniques in Table 12.1. Figure 12.6 depicts the convergence and MSE analysis of the proposed algorithms for LTE structure. The proposed PSO-TVW3-based algorithm converges fast with minimum MSE value.

Fig. 12.5
figure 5

Performance of proposed and other time-varying strategies in linear channel for LTE

Table 12.1 Comparison of convergence rate and MSE with different channel models
Fig. 12.6
figure 6

Proposed PSO enhancements in linear channel for LTE

Table 12.1 gives the comparison and effect of different channels using least mean square (LMS) algorithm, PSO-LDIW [9], PSO-TVW2, and PSO-TVW3 for LTE and DFE structures. From Table 12.1 and Figs. 12.5 to 12.6, it is shown that the proposed modifications outperform the other existing modifications based on convergence and MSE. The PSO-TVW3 algorithm shows the best performance in all channel conditions. The LTE and DFE structures give approximately same MSE value, but differ in convergence rate which is shown in Table 12.1.

12.4.1 Sensitivity Analysis

The simulation-based sensitivity analysis is carried out to select optimum parameters in the algorithm. The parameter’s values and choices have high impact on the efficiency of the method, and few others have less or no effect. The analysis is done with respect to six key parameters, namely the intermediate iteration value N, data window size ws, acceleration constants ac 1 and ac 2, the population size P, number of tap weights T, and distortion factor W. The effect of the basic PSO parameters swarm size or number of particles, window size, number of tap weights and acceleration coefficients are analyzed in [6]. The same is analyzed for PSO-TVW3 and is given in Table 12.2.

Table 12.2 Effect of PSO parameters on PSO-TVW3

On average, an increase in the number of particles will always provide faster convergence. In contrast, the computational complexity can increase linearly with increase in population size. In Table 12.2, population size of 40 gives better convergence. So a problem-dependent minimum population size is good for better performance.

Setting the acceleration coefficients to a minimum value slows down the convergence speed. The local search and global search are best when the summation of acceleration coefficients becomes ac 1  + ac 2  < 4 in adaptive equalization. The acceleration coefficients greater than 1 also seem to give the best performance. For equal value of acceleration constants, the algorithm reaches its minimum in least number of iterations. The MSE calculated on each iteration is the average of the MSE over the window; a large window size increases the complexity per iteration and time consumption. In Table 12.2, window size does not make any greater changes in the MSE value. If the window size is small, the complexity can be reduced.

The tap weights are problem dependent. As given in Table 12.2, the increase in tap weights above a certain limit does not make much difference in MSE value, but it may increase the complexity. Figure 12.7 shows the analysis for different intermediate iteration N for PSO-TVW2. Table 12.3 compares the convergence rate and MSE for PSO-TVW2 and PSO-TVW3 with reference to N. An increase in the value of N increases the number of iterations required for convergence. Decreasing N value degrades the MSE performance. N value between 30 and 40 exhibits minimum MSE with faster convergence.

Fig. 12.7
figure 7

Effect of different intermediate iteration value N on PSO-TVW2

Table 12.3 Comparison of convergence rate with different intermediate iteration value N

Table 12.4 explains the effect of amplitude distortion parameter W in linear channel. The MSE is computed with different amplitude distortion that leads to different eigenvalue spread. Increase in amplitude distortion degrades the MSE performance. The performance degradation is not severe in proposed PSO-based algorithms compared to existing algorithms.

Table 12.4 Effect of amplitude distortion W on PSO-TVW2 for LTE and DFE

12.5 Conclusion

In this work, two new strategies are proposed to update in PSO algorithm for adaptive equalization. The results are discussed in linear and nonlinear channels for LTE and DFE structures. The proposed time-varying PSO algorithms, PSO-TVW2 and PSO-TVW3, show better performance than the existing algorithms in linear and nonlinear channels. Also it shows performance better for LTE and DFE structures. The PSO-TVW3 algorithm outperforms other modifications based on convergence and MSE. The sensitivity analysis is done to find the optimum parameter values. The performance is degraded for severe amplitude distortions in linear channel. The DFE structure delays the convergence but it performs well in severe amplitude distortion conditions.