1 Introduction

Over the past three decades, flexible-joint manipulator has attracted a great deal of interests owing to its obvious superiority of small actuators, high precision and low energy consumption [1,2,3,4,5,6,7]. In contrast to the rigid manipulator, flexible-joint manipulator possess flexibility, high security and low rate of damages [8,9,10]. In 1989, Spong derived a reduced-order model of the manipulator which was the fist simplified dynamic model and they presented the first adaptive control law for flexible-joint robots [11, 12]. In the modeling and control, the flexible-joint manipulator presents serious problems because of the inherent highly coupling, nonlinearity and model uncertainty. Therefore, it improves the difficulty of the controller design which has led to a great deal of research using advanced control theory to design more appropriate controllers [2, 13,14,15].

At present, backstepping approach is one of the most commonly used design methods for solving nonlinear systems. In [16], a new adaptive backstepping controller was proposed to the adaptive tracking problem in uncertain flexible-joint manipulator system. [17] presented an adaptive sliding control method using a backstepping-like design for single-link flexible-joint robot. However, it is difficult to obtain complete or partial machine parameters in many practical applications, and motion in robot is a complicated nonlinear process that is hard to model as a linear-in-the-parameter process. The function approximation technique has the great advantages to deal with this issue, which does not require the system dynamics to be exactly known [18,19,20]. Adaptive backstepping controllers combined with several universal approximators have been successfully presented to control the uncertain manipulator system, such as recurrent neural networks [21], self-recurrent wavelet neural network [22], neural networks [23,24,25] and fuzzy system [26, 27]. With the advent of the fuzzy set theory proposed by Zadeh (1965), the fuzzy system was proven to be an effective approach for investigating a bank of complex nonlinear control design problems [28,29,30]. The type-1 fuzzy model offers a general framework for nonlinear system analysis and controller synthesis [31]. However, due to crisp antecedents and consequents of the rule base, type-1 fuzzy system (T1FS) cannot efficiently handle the uncertainties. Interval type-2 fuzzy system (IT2FS) was widely applied to many practical applications and can obtain better performance for highly nonlinear systems with various uncertainties than T1FS [32,33,34,35]. An improved social spider optimization algorithm was proposed to adjust the predecessor parameters of general type-2 fuzzy system (GT2FS) in [36], but one of the important limitations is that the computational cost of the proposed GT2FS is increased in the high-dimensional problems. Due to the simplicity and efficiency of the IT2FS, it is very valuable to use the IT2FS to handle highly nonlinearity of manipulator system in this paper.

IT2FS can improve the system’s ability to deal with uncertainties and approximate uncertain unknown functions effectively. [37] solved the globally stable adaptive backstepping control based on IT2FS for a class of nonlinear systems. In [38], backstepping and IT2FS were combined in a unified controller for induction machine. An adaptive backstepping robust control approach based on IT2FS was proposed for uncertain multi-inputs multi-outputs chaotic system in [39]. IT2FNN has potential to improve approximation accuracy, but there are very few researches to design adaptive controller for flexible-joint manipulators using IT2FNN approximator. In [40, 41], sliding mode control methods with the IT2FS approximator for flexible-joint manipulator were proposed. In the current rare attempts in developing IT2FNN approximator, high time consumption of the iterative K–M algorithm is still a problem that cannot be ignored. In the previous papers on IT2FS approximator, the iterative K–M algorithm is used to rearrange the rule’s consequent weights in ascending order and find the left and right crossover points [42]. However, high computational complexity and high time consumption of the iterative K–M algorithm in type-reduction of IT2FS make it very hard to be used in practical applications [43]. Therefore, improving the type-reduction algorithm has always been the focus of researchers. The adaptive control factors \(q_{l}\) and \(q_{r}\) were proposed in [44], instead of finding the left and right crossover points in the iterative K–M algorithms. Then, in 2017, Bibi et al. [45] replaced K–M algorithms by the adaptive modulation factor \(\alpha \) between the upper output \(y_{r}\) and the lower output \(y_{l}\) in the IT2FNN. The adaptive modulation factor improves the practicability of the algorithm. In this paper, the adaptive modulation factor \(\alpha \) is applied to IT2FNN approximator and effectively overcomes computational complexity and high time consumption. To the best of the authors knowledge, improved IT2FNN approximator has not yet been investigated for flexible-joint manipulator, thereby leaving room for continued promotion. This study is the first attempt to apply adaptive modulation factor into IT2FNN approximator to design adaptive backstepping controller for flexible-joint manipulator with uncertain dynamics. Thus, the tracking performance of the system can be improved by the proposed method.

Based on the above discussion, the motivation of this study is triggered. An adaptive backstepping control method based on IT2FNN approximator for flexible-joint manipulator is proposed. Using Lyapunov stability theory, all the signals in the closed-loop system are guaranteed to be ultimately bounded. Compared to the existing works, the proposed approach does not require the unknown parameters to be linear parameterizable and the tracking error can be reduced to arbitrarily small values. The main contributions of this paper are as follows: (1) This paper is the first attempt in developing the adaptive modulation factor \(\alpha \) into IT2FNN approximator to control flexible-joint manipulators with mismatched uncertainties. (2) We devise the adaptive law of adaptive modulation factor \(\alpha \) and the adaptive parameters from Lyapunov stability analysis. The adaptive modulation factor can be iteratively updated, and thus the ability of adaptive ponderation between the upper output \(y_{r}\) and the lower output \(y_{l}\) can be improved. (3) The proposed controller can guarantee not only the stability of manipulator system but also the boundedness of all the signals in the closed-loop system. (4) Compared with the T1FNN and the NN approximator, simulation results demonstrate that the proposed scheme has better steady-state performance, less fluctuation and higher approximation accuracy.

The rest of paper is organized as follows. The problems formulation is presented in Sect. 2. Section 3 describes the IT2FNN approximator. In Sect. 4, we derive the proposed controller and verify stability of the closed loop using Lyapunov approach. Section 5 presents the feasibility and the effectiveness of proposed controller for flexible-joint manipulator by comparing with the others control methods. Finally, conclusions are drawn in Sect. 6.

2 Problem formulation

This section is referenced from [17]. A schematic model of a single-link flexible-joint manipulator is shown in Fig. 1. We assume that its joint can only be deformed when rotating in a vertical plane in the direction of joint rotation. The operating mechanism of the flexible-joint manipulator is that the motor shaft and the rigid link are, respectively, driven by the motor and spring to rate. Assuming that the viscous damping is ignored and the states are measurable, its dynamic equation is given by

$$\begin{aligned} {\left\{ \begin{array}{ll} I\ddot{q_{1}}+MgL\sin q_{1}+K(q_{1}-q_{2})=0\\ J\ddot{q_{2}}+K(q_{2}-q_{1})=u\\ \end{array}\right. } \end{aligned}$$
(1)

where \(q_{1}\in R^{n}\) and \(q_{2}\in R^{n}\) are the angular displacements of flexible-joint link and motor, K is the spring stiffness of joints, \(u\in R^{n}\) is the external input, which is the torque delivered by the motor, I and J are, respectively, the moment of inertia of flexible-joint link and the motors, M is the mass of flexible-joint link, and L is the length between the center of gravity of the manipulator and flexible-joints.

Fig. 1
figure 1

Schematic of flexible-joint manipulator model

We define \(x_{1}=q_{1}\), \(x_{2}=\dot{q_{1}}\), \(x_{3}=q_{2}, x_{4}=\dot{q_{2}}\) (1) can be rewritten as the following state-spaced representation

$$\begin{aligned} {\left\{ \begin{array}{ll} \dot{x_{1}}=x_{2}\\ \dot{x_{2}}=-\frac{1}{I}(MgL\sin (x_{1})+K(x_{1}-x_{3}))\\ \dot{x_{3}}=x_{4}\\ \dot{x_{4}}=\frac{1}{J}(u+K(x_{1}-x_{3}))\\ \end{array}\right. } \end{aligned}$$
(2)

where \(x_{i} \in R^{n},i=1,2,3,4\) are state variables and \(y=x_{1}\) is the link angular displacement. Considering a single-link flexible-joint manipulator with mismatched uncertainties, the above model cannot be available. Since the robot is basically a link powered by the electric motor through a twisted spring, we can represent it as a cascade of two subsystems: link dynamics and the motor dynamics. The control input is in the subsystem describing the motor dynamics, and its output is coupled to another subsystem with the spring and link dynamics. Therefore, we can write (1) as a simplified system equation

$$\begin{aligned} {\left\{ \begin{array}{ll} \dot{x_{1}}=x_{2}\\ \dot{x_{2}}=x_{3}+g(x)\\ \dot{x_{3}}=x_{4}\\ \dot{x_{4}}=f(x)+mu\\ \end{array}\right. } \end{aligned}$$
(3)

It is obvious that \(g(x)=-x_{3}-MgL\sin (x_{1})/I-K(x_{1}-x_{3})/I\), \(f(x)=K(x_{1}-x_{3})/J\), \(m=1/J\). We assume that g(x), f(x) and m are unknown constants where the lower bound of m is known and satisfies \(m \ge \underline{m}\) and \(\underline{m}>0\). An adaptive backstepping controller with fuzzy approximation is designed to desired trajectory tracking. IT2FNN is utilized to approximate unknown nonlinear functions.

3 IT2FNN approximator

This section introduces an IT2FNN approximator, which can obtain a very accurate and robust approximation. The architecture of IT2FNN is shown in Fig. 2. IT2FNN has obvious advantages of handling uncertainties and approximating unknown nonlinear functions by using lower and upper membership functions. IT2FNN can be thought as consisting of two parts: one part contains some IF-THEN rules, and the second part is the fuzzy inference engine.

Fig. 2
figure 2

Architecture of IT2FNN

Each rule in the IT2FNN approximator is presented in the following form:

$$\begin{aligned} \mathbb {R}^k: \hbox {if}\ \begin{array}{*{20}{c}} {{x_1}\ \begin{array}{*{20}{c}} {is} \end{array}\begin{array}{*{20}{c}} {\hat{F}_1^{k}\ \begin{array}{*{20}{c}} \hbox {and} \end{array}\begin{array}{*{20}{c}} \ldots \end{array}} \end{array}} \end{array}\begin{array}{*{20}{c}} {\mathrm{{and}}\begin{array}{*{20}{c}} {\ {x_n}\ \begin{array}{*{20}{c}} \hbox {is} \end{array}} \end{array}} \end{array}\begin{array}{*{20}{c}} {\hat{F}_n^{k}} \end{array}\nonumber \\ \hbox {then}\begin{array}{*{20}{c}} {\ y\ \begin{array}{*{20}{c}} {\hbox {is}\begin{array}{*{20}{c}} {{\theta _{k}}} \end{array}} \end{array}} \end{array} \quad k=1,\ldots ,N \end{aligned}$$
(4)

where \(x_{1},x_{2},\ldots ,x_{n}\) are the input variables and y is the output variable. N is the total number of fuzzy rules. \(\hat{F}_{i}^{k}\), \(i=1,2,\ldots ,n\), \(k=1,2,\ldots ,N\), are interval type-2 fuzzy antecedent. \(\theta _{k}=[\underline{\theta }_{k},\overline{\theta }_{k}] \) represents the lower and the upper singleton consequent type-2 fuzzy sets. Mathematical function of each operator are described as follows.

For an input vector \(x=[x_{1},x_{2},\ldots ,x_{n}]\), using the singleton fuzzifier and product t-norm, the lower and upper bounds of the firing the kth rule \(\varphi ^{k}\) strength can be computed as follows:

$$\begin{aligned} \varphi ^{k}=[\underline{f}^k, \overline{f}^k], \quad k=1,2,\ldots ,N \end{aligned}$$
(5)

where:

$$\begin{aligned} {\left\{ \begin{array}{ll} \underline{f}^k=\mu _{\underline{F}_{1}^{k}}(x_{1})*\cdots *\mu _{\underline{F}_{i}^{k}}(x_{i})\\ \overline{f}^k=\mu _{\overline{F}_{1}^{k}}(x_{1})*\cdots *\mu _{\overline{F}_{i}^{k}}(x_{i})\\ \end{array}\right. } \end{aligned}$$
(6)

in which \(\mu _{\underline{F}_{i}^{k}}(x_{i})\) and \(\mu _{\overline{F}_{i}^{k}}(x_{i})\), respectively, are the lower membership function and the upper membership function. Then, the type-reduction converts the interval type-2 fuzzy sets to an interval set. Finally, defuzzifier maps the interval set into a crisp output.

There are many methods of designing the type-reduction of interval type-2 fuzzy sets. The most common method is the center-of-sets type-reduction, which can be expressed by:

$$\begin{aligned} Y_\mathrm{COS}(x)=\frac{\sum \nolimits _{k=1}^{N}\varphi ^{k}(x_{i})\theta _{k}}{\sum \nolimits _{k=1}^{N}\varphi ^{k}(x_{i})}=[y_{l},y_{r}] \end{aligned}$$
(7)

where \(y_{l}\) and \(y_{r}\) are computed as follows:

$$\begin{aligned} y_{l}=\frac{\sum \nolimits _{k=1}^{N}\underline{f}^{k} \underline{\theta }_{k}}{\sum \nolimits _{k=1}^{N}\underline{f}^{k}}=\sum \nolimits _{k=1}^{N}\underline{\xi }^{k}\underline{\theta }_{k} =\underline{\xi }(x)\underline{\theta }^{T} \end{aligned}$$
(8)

and

$$\begin{aligned} y_{r}=\frac{\sum \nolimits _{k=1}^{N}\overline{f}^{k} \overline{\theta }_{k}}{\sum \nolimits _{k=1}^{N}\overline{f}^{k}}=\sum \nolimits _{k=1}^{N}\overline{\xi }^{k}\overline{\theta }_{k} =\overline{\xi }(x)\overline{\theta }^{T} \end{aligned}$$
(9)

where \(\underline{\theta }=[\underline{\theta }_1,\underline{\theta }_2,\ldots ,\underline{\theta }_N]\) and \(\overline{\theta }=[\overline{\theta }_1,\overline{\theta }_2,\ldots ,\overline{\theta }_N]\) are the adjustable parameters, \(\underline{\xi }(x)=[\underline{\xi }^1,\underline{\xi }^2,\ldots ,\underline{\xi }^k]\) and \(\overline{\xi }(x)=[\overline{\xi }^1,\overline{\xi }^2,\ldots ,\overline{\xi }^k]\) are the fuzzy basis functions that are computed as follows:

$$\begin{aligned} \underline{\xi }^{k}=\frac{\underline{f}^{k}}{\sum \nolimits _{p=1}^{N}\underline{f}^{p}}, \overline{\xi }^{k}=\frac{\overline{f}^{k}}{\sum \nolimits _{p=1}^{N}\overline{f}^{p}} \end{aligned}$$
(10)

K–M and EIASC iterative algorithms can determine some crossover points, which combine the lower output \(y_{l}\) and the upper output \(y_{r}\). But it consumes a lot of time in iterative process, especially for real-time applications. An adaptive factor \(\alpha \) is proposed to obtain an adaptive modulation between \(y_{l}\) and \(y_{r}\), which can overcome those drawbacks of such iterative algorithm, including high time-consuming and low accuracy-calculating [45].

The defuzzified output \(Y(x,\underline{\theta },\overline{\theta })\) is computed as follows:

$$\begin{aligned} Y(x,\underline{\theta },\overline{\theta })=\alpha y_{r}+(1-\alpha ) y_{l} \end{aligned}$$
(11)

Substituting (8) and (9) into (11), we get

$$\begin{aligned} Y(x,\underline{\theta },\overline{\theta })=\alpha \overline{\xi }(x)\overline{\theta }^{T}+(1-\alpha ) \underline{\xi }(x)\underline{\theta }^{T} \end{aligned}$$
(12)

4 Controller design

In this section, an adaptive backstepping controller with the IT2FNN approximator is proposed for a flexible-joint manipulator with mismatched uncertainties.

4.1 Backstepping controller

In the process of backstepping, at each recursive step, virtual controllers \(x_{id}, i=2,\ldots ,m\), are designed to force errors \(e_{i-1}=x_{i-1}-x_{(i-1)d}\) as small as possible. The final virtual controller \(x_{id}\) is the part of actual controller u. An actual controller for u is used to minimize the error between \(x_{m}\) and \(x_{md}\) to be as much as possible. The design of the controller is divided into several steps.

Step 1 Define \(e_{1}=x_{1}-x_{1d}\) and let \(x_{1d}=y_{d}\). We get

$$\begin{aligned} \dot{e}_{1}=\dot{x}_{1}-\dot{x}_{1d}=x_{2}-\dot{x}_{1d} \end{aligned}$$
(13)

Define \(e_{2}=x_{2}-x_{2d}\) and a virtual controller \(x_{2d}\).

$$\begin{aligned} x_{2d}=\dot{x}_{1d}-k_{1}e_{1} \end{aligned}$$
(14)

where \(k_{1}\) is a positive constant.

Then, (13) can be rewritten as:

$$\begin{aligned} \dot{e}_{1}=e_{2}+x_{2d}-\dot{x}_{1d}=-k_{1}e_{1}+e_{2} \end{aligned}$$
(15)

Consider the following Lyapunov function candidate:

$$\begin{aligned} V_{1}=\frac{1}{2}{e_{1}}^{2} \end{aligned}$$
(16)

The time derivative of \(V_{1}\) is

$$\begin{aligned} \dot{V}_{1}=-k_{1}{e_{1}}^{2}+e_{1}e_{2} \end{aligned}$$
(17)

If \(e_{2}=0\), then \(\dot{V}_{1}\le 0\).

Step 2 Taking the time derivative of \(e_{2}=x_{2}-x_{2d}\), then

$$\begin{aligned} \dot{e}_{2}=\dot{x}_{2}-\dot{x}_{2d}=x_{3}+g-\dot{x}_{2d} \end{aligned}$$
(18)

Define \(e_{3}=x_{3}-x_{3d}\) and a virtual controller \(x_{3d}\).

$$\begin{aligned} x_{3d}=-\hat{g}+\dot{x}_{2d}-k_{2}e_{2}-e_{1} \end{aligned}$$
(19)

where \(k_{2}\) is a positive constant and \(\hat{g}\) is the estimated value of g.

From (14), the time derivative of \(x_{2d}\) is:

$$\begin{aligned} \dot{x}_{2d}=\ddot{x}_{1d}-k_{1}\dot{e}_{1}=\ddot{x}_{1d}-k_{1}(x_{2}-\dot{x}_{1d}) \end{aligned}$$
(20)

From (18) and (19), the derivative of \(e_{2}\) can be obtained as

$$\begin{aligned} \dot{e}_{2}=e_{3}+x_{3d}-\dot{x}_{2d}+g=g-\hat{g}-k_{2}e_{2}+e_{3}-e_{1} \end{aligned}$$
(21)

Consider the following Lyapunov function candidate:

$$\begin{aligned} V_{2}=V_{1}+\frac{1}{2}{e_{2}}^{2} \end{aligned}$$
(22)

The time derivative of \(V_{2}\) can be obtained as

$$\begin{aligned} \dot{V}_{2}=-k_{1}{e_{1}}^{2}-k_{2}{e_{2}}^{2}+(g-\hat{g})e_{2}+e_{2}e_{3} \end{aligned}$$
(23)

If \(e_{3}=0\) and \(\hat{g}=g\), then \(\dot{V}_{2}\le 0\).

Step 3 Taking the time derivative of \(e_{3}=x_{3}-x_{3d}\), then

$$\begin{aligned} \dot{e}_{3}=\dot{x}_{3}-\dot{x}_{3d}=x_{4}-\dot{x}_{3d} \end{aligned}$$
(24)

From (18), (19), (20) and (22), the time derivative of \(x_{3d}\) is:

$$\begin{aligned} \dot{x}_{3d}=&-\dot{\hat{g}}+\ddot{x}_{2d}-k_{2}\dot{e}_{2}-\dot{e}_{1}\nonumber \\ =&-\dot{\hat{g}}+\dddot{x}_{1d}-k_{1}(x_{3}+g-\ddot{x}_{1d}) \nonumber \\&-k_{2}(x_{3}+g-\dot{x}_{2d})-x_{2}+\dot{x}_{1d} \end{aligned}$$
(25)

We divide \(\dot{x}_{3d}\) into two parts. \(\dot{x}_{3d}'\) is the known part without model information, and \(\bar{\dot{x}}_{3d}\) is the unknown part with model information (25), which can be rewritten as

$$\begin{aligned} \dot{x}_{3d}=\dot{x}_{3d}'-\bar{\dot{x}}_{3d} \end{aligned}$$
(26)

where

$$\begin{aligned} \dot{x}_{3d}'= & {} \dddot{x}_{1d}-k_{1}(x_{3}-\ddot{x}_{1d})-k_{2}(x_{3}-\dot{x}_{2d})-x_{2}+\dot{x}_{1d} \nonumber \\\end{aligned}$$
(27)
$$\begin{aligned} \bar{\dot{x}}_{3d}= & {} \dot{\hat{g}}+k_{1}g+k_{2}g \end{aligned}$$
(28)

Define \(e_{4}=x_{4}-x_{4d}\), \(\bar{\dot{x}}_{3d}=d\) and a virtual controller \(x_{4d}\). Choose a positive constant \(k_{3}\), we have

$$\begin{aligned} x_{4d}=\dot{x}_{3d}'-\hat{d}-k_{3}e_{3}-e_{2} \end{aligned}$$
(29)

Substituting (26)–(29) into (24), we can get

$$\begin{aligned} \dot{e}_{3}=&x_{4}-\dot{x}_{3d}=x_{4}-\dot{x}_{3d}'+\bar{\dot{x}}_{3d}\nonumber \\ =&-k_{3}e_{3}-e_{2}+e_{4}+(d-\hat{d}) \end{aligned}$$
(30)

Consider the following Lyapunov function candidate:

$$\begin{aligned} V_{3}=V_{2}+\frac{1}{2}{e_{3}}^{2} \end{aligned}$$
(31)

The time derivative of \(V_{3}\) can be obtained as

$$\begin{aligned} \dot{V}_{3}=&-k_{1}{e_{1}}^{2}-k_{2}{e_{2}}^{2}-k_{3}{e_{3}}^{2}+(g-\hat{g})e_{2}\nonumber \\&+(d-\hat{d})e_{3}+e_{3}e_{4} \end{aligned}$$
(32)

If \(e_{4}=0\), \(\hat{g}=g\) and \(\hat{d}=d\), then \(\dot{V}_{3}\le 0\).

Step 4 To realize stability of control system, we take an actual controller into this step. The time derivative of \(e_{4}=x_{4}-x_{4d}\) is

$$\begin{aligned} \dot{e}_{4}=\dot{x}_{4}-\dot{x}_{4d}=f+mu-\dot{x}_{4d} \end{aligned}$$
(33)

From (24), (26), (27) and (29), the time derivative of \(x_{4d}\) is:

$$\begin{aligned} \dot{x}_{4d}&=-\dot{\hat{d}}+\ddot{x}_{3d}'-k_{3}\dot{e}_{3}-\dot{e}_{2}\nonumber \\&\quad =\ddddot{x}_{1d}-k_{1}(x_{4}-\dddot{x}_{1d})-k_{2}(x_{4}-\dddot{x}_{1d}\nonumber \\&\qquad +k_{1}(x_{3}-\ddot{x}_{1d}))+\ddot{x}_{1d}-x_{3}-k_{3}(x_{4}-\dot{x}_{3d}')\nonumber \\&\qquad -(x_{3}-\dot{x}_{2d})-k_{1}k_{2}g\nonumber \\&\qquad -g-\dot{\hat{d}}-k_{3}\bar{\dot{x}}_{3d}-g \end{aligned}$$
(34)

We divide \(\dot{x}_{4d}\) into two parts. \(\dot{x}_{4d}'\) is the known part without model information, and \(\bar{\dot{x}}_{4d}\) is the unknown part with model information (34), which can be rewritten as

$$\begin{aligned} \dot{x}_{4d}=\dot{x}_{4d}^{\prime }+\bar{\dot{x}}_{4d} \end{aligned}$$
(35)

where

$$\begin{aligned} \dot{x}_{4d}^{\prime }= & {} \ddddot{x}_{1d}-k_{2}(x_{4}-\dddot{x}_{1d}\nonumber \\&+\,k_{1}(x_{3}-\ddot{x}_{1d}))+\ddot{x}_{1d}-x_{3}-k_{3}(x_{4}-\dot{x}_{3d}')\nonumber \\&-\,(x_{3}-\dot{x}_{2d})-k_{1}(x_{4}-\dddot{x}_{1d}) \end{aligned}$$
(36)
$$\begin{aligned} \bar{\dot{x}}_{4d}= & {} -k_{1}k_{2}g-g-\dot{\hat{d}}-k_{3}\bar{\dot{x}}_{3d}-g \end{aligned}$$
(37)

Define \(h=f-\bar{\dot{x}}_{4d}\), and (33) can be rewritten as

$$\begin{aligned} \dot{e}_{4}&=h+\bar{\dot{x}}_{4d}+mu-\dot{x}_{4d}\nonumber \\&=h-\dot{x}_{4d}'+(m-\hat{m})u+\hat{m}u \end{aligned}$$
(38)

where \(\hat{m}\) is the estimated value of m.

Choosing the following control law

$$\begin{aligned} u=\frac{1}{\hat{m}}(-\hat{h}+\dot{x}_{4d}'-k_{4}e_{4}-e_{3}) \end{aligned}$$
(39)

where \(\hat{h}\) is the estimated value of h and \(k_{4}\) is a positive constant.

Substituting (39) into (38), we can get

$$\begin{aligned} \dot{e}_{4}=(h-\hat{h})-(m-\hat{m})u-k_{4}e_{4}-e_{3} \end{aligned}$$
(40)

Consider the following Lyapunov function candidate:

$$\begin{aligned} V_{4}=V_{3}+\frac{1}{2}{e_{4}}^{2} \end{aligned}$$
(41)

The time derivative of \(V_{4}\) can be obtained as

$$\begin{aligned} \dot{V}_{4}=&-k_{1}{e_{1}}^{2}-k_{2}{e_{2}}^{2}-k_{3}{e_{3}}^{2}\nonumber \\&-k_{4}{e_{4}}^{2}+(m-\hat{m})ue_{4}\nonumber \\&+(g-\hat{g})e_{2}+(d-\hat{d})e_{3}+(h-\hat{h})e_{4} \end{aligned}$$
(42)

If \(\hat{m}=m\), \(\hat{g}=g\), \(\hat{d}=d\) and \(\hat{h}=h\), then \(\dot{V}_{4}\le 0\).

4.2 Adaptive fuzzy controller

We use the proposed approximator in this section to approximate unknown nonlinear functions g(x), d(x) and h(x), where the approximations are \(\hat{g}(x)\), \(\hat{h}(x)\) and \(\hat{d}(x)\).

Taking the proposed adaptive factor into g(x), d(x) and h(x), we can get

$$\begin{aligned} g(x)=&(1-\alpha _{g})\underline{\xi }_{g}(x){\underline{\theta }_{g}^{*}}^{T}+\alpha _{g} \overline{\xi }_{g}(x){\overline{\theta }_{g}^{*}}^{T}\nonumber \\&+(1-\alpha _{g})\underline{\varepsilon }_{g}(x)+\alpha _{g}\overline{\varepsilon }_{g}(x) \end{aligned}$$
(43)
$$\begin{aligned} d(x)=&(1-\alpha _{d})\underline{\xi }_{d}(x){\underline{\theta }_{d}^{*}}^{T}+\alpha _{d} \overline{\xi }_{d}(x){\overline{\theta }_{d}^{*}}^{T}\nonumber \\&+(1-\alpha _{d})\underline{\varepsilon }_{d}(x)+\alpha _{d}\overline{\varepsilon }_{d}(x) \end{aligned}$$
(44)
$$\begin{aligned} h(x)=&(1-\alpha _{h})\underline{\xi }_{h}(x){\underline{\theta }_{h}^{*}}^{T}+\alpha _{h} \overline{\xi }_{h}(x){\overline{\theta }_{h}^{*}}^{T}\nonumber \\&+(1-\alpha _{h})\underline{\varepsilon }_{h}(x)+\alpha _{h}\overline{\varepsilon }_{h}(x) \end{aligned}$$
(45)

where \(\underline{\varepsilon }_{g}(x)\) and \(\overline{\varepsilon }_{g}(x)\), \(\underline{\varepsilon }_{d}(x)\) and \(\overline{\varepsilon }_{d}(x)\), \(\underline{\varepsilon }_{h}(x)\) and \(\overline{\varepsilon }_{h}(x)\) are the approximation errors; \(\underline{\xi }_{g}\) and \(\overline{\xi }_{g}\), \(\underline{\xi }_{d}\) and \(\overline{\xi }_{d}\), \(\underline{\xi }_{h}\) and \(\overline{\xi }_{h}\), are, respectively, the lower and the upper membership functions; \({\underline{\theta }_{g}^{*}}^{T}\) and \({\overline{\theta }_{g}^{*}}^{T}\), \({\underline{\theta }_{d}^{*}}^{T}\) and \({\overline{\theta }_{d}^{*}}^{T}\), \({\underline{\theta }_{h}^{*}}^{T}\) and \({\overline{\theta }_{h}^{*}}^{T}\) are the optimal lower and upper approximation parameters of g(x), d(x) and h(x). \(\alpha _{g}\), \(\alpha _{d}\), \(\alpha _{h}\) are the adaptive factor.

According to the proposed approximator, the nonlinear functions \(\hat{g}(x,\hat{\underline{\theta }}_{g},\hat{\overline{\theta }}_{g})\), \(\hat{d}(x,\hat{\underline{\theta }}_{d},\hat{\overline{\theta }}_{d})\), \(\hat{h}(x,\hat{\underline{\theta }}_{h},\hat{\overline{\theta }}_{h})\) can be expressed as

$$\begin{aligned} \hat{g}(x,\hat{\underline{\theta }}_{g},\hat{\overline{\theta }}_{g})= & {} (1-\hat{\alpha _{g}})\underline{\xi }_{g}(x) {\hat{\underline{\theta }}_{g}}^{T}+\hat{\alpha _{g}}\overline{\xi }_{g}(x){\hat{\overline{\theta }}_{g}}^{T} \end{aligned}$$
(46)
$$\begin{aligned} \hat{d}(x,\hat{\underline{\theta }}_{d},\hat{\overline{\theta }}_{d})= & {} (1-\hat{\alpha _{d}})\underline{\xi }_{d}(x) {\hat{\underline{\theta }}_{d}}^{T}+\hat{\alpha _{d}}\overline{\xi }_{d}(x){\hat{\overline{\theta }}_{d}}^{T} \end{aligned}$$
(47)
$$\begin{aligned} \hat{h}(x,\hat{\underline{\theta }}_{h},\hat{\overline{\theta }}_{h})= & {} (1-\hat{\alpha _{h}})\underline{\xi }_{h}(x) {\hat{\underline{\theta }}_{h}}^{T}+\hat{\alpha _{h}}\overline{\xi }_{h}(x){\hat{\overline{\theta }}_{h}}^{T} \end{aligned}$$
(48)

From (43) to (48), we have

$$\begin{aligned} \tilde{g}(x)&=g(x)-\hat{g}(x,\hat{\underline{\theta }}_{g},\hat{\overline{\theta }}_{g})\nonumber \\&=(1-\hat{\alpha }_{g})\underline{\xi }_{g}(x) {\tilde{\underline{\theta }}_{g}}^{T}+\hat{\alpha }_{g}\overline{\xi }_{g}(x){\tilde{\overline{\theta }}_{g}}^{T}\nonumber \\&+(\overline{\xi }_{g}(x){\hat{\overline{\theta }}_{g}}^{T}-\underline{\xi }_{g}(x){\hat{\underline{\theta }}_{g}}^{T})\tilde{\alpha }_{g}\nonumber \\&+(\overline{\xi }_{g}(x){\tilde{\overline{\theta }}_{g}}^{T}-\underline{\xi }_{g}(x){\tilde{\underline{\theta }}_{g}}^{T})\tilde{\alpha }_{g}\nonumber \\&+(1-\alpha _{g})\underline{\varepsilon }_{g}(x)+\alpha _{g}\overline{\varepsilon }_{g}(x) \end{aligned}$$
(49)
$$\begin{aligned} \tilde{d}(x)&=d(x)-\hat{d}(x,\hat{\underline{\theta }}_{d},\hat{\overline{\theta }}_{d})\nonumber \\&=(1-\hat{\alpha }_{d})\underline{\xi }_{d}(x) {\tilde{\underline{\theta }}_{d}}^{T}+\hat{\alpha }_{d}\overline{\xi }_{d}(x){\tilde{\overline{\theta }}_{d}}^{T}\nonumber \\&+\left( \overline{\xi }_{d}(x){\hat{\overline{\theta }}_{d}}^{T}-\underline{\xi }_{d}(x){\hat{\underline{\theta }}_{d}}^{T}\right) \tilde{\alpha }_{d}\nonumber \\&+\left( \overline{\xi }_{d}(x){\tilde{\overline{\theta }}_{d}}^{T}-\underline{\xi }_{d}(x){\tilde{\underline{\theta }}_{d}}^{T}\right) \tilde{\alpha }_{d}\nonumber \\&+(1-\alpha _{d})\underline{\varepsilon }_{d}(x)+\alpha _{d}\overline{\varepsilon }_{d}(x) \end{aligned}$$
(50)
$$\begin{aligned} \tilde{h}(x)&=h(x)-\hat{h}(x,\hat{\underline{\theta }}_{h},\hat{\overline{\theta }}_{h})\nonumber \\&=(1-\hat{\alpha }_{h})\underline{\xi }_{h}(x) {\tilde{\underline{\theta }}_{h}}^{T}+\hat{\alpha }_{h}\overline{\xi }_{h}(x){\tilde{\overline{\theta }}_{h}}^{T}\nonumber \\&+\left( \overline{\xi }_{h}(x){\hat{\overline{\theta }}_{h}}^{T}-\underline{\xi }_{h}(x){\hat{\underline{\theta }}_{h}}^{T}\right) \tilde{\alpha }_{h}\nonumber \\&+\left( \overline{\xi }_{h}(x){\tilde{\overline{\theta }}_{h}}^{T}-\underline{\xi }_{h}(x){\tilde{\underline{\theta }}_{h}}^{T}\right) \tilde{\alpha }_{h}\nonumber \\&+(1-\alpha _{h})\underline{\varepsilon }_{h}(x)+\alpha _{h}\overline{\varepsilon }_{h}(x) \end{aligned}$$
(51)

where \(\tilde{\underline{\theta }}_{g}=\underline{\theta }_{g}^{*}-\hat{\underline{\theta }}_{g}\), \(\tilde{\overline{\theta }}_{g}=\overline{\theta }_{g}^{*}-\hat{\overline{\theta }}_{g}\), \(\tilde{\underline{\theta }}_{d}=\underline{\theta }_{d}^{*}-\hat{\underline{\theta }}_{d}\), \(\tilde{\overline{\theta }}_{d}=\overline{\theta }_{d}^{*}-\hat{\overline{\theta }}_{d}\), \(\tilde{\underline{\theta }}_{h}=\underline{\theta }_{h}^{*}-\hat{\underline{\theta }}_{h}\), \(\tilde{\overline{\theta }}_{h}=\overline{\theta }_{h}^{*}-\hat{\overline{\theta }}_{h}\), \(\tilde{\alpha }_{g}=\alpha _{g}-\hat{\alpha }_{g}\), \(\tilde{\alpha }_{d}=\alpha _{d}-\hat{\alpha }_{d}\) and \(\tilde{\alpha }_{h}=\alpha _{h}-\hat{\alpha }_{h}\).

The adaptive law of m is chosen as nonlinear functions, which can be expressed as

$$\begin{aligned} \dot{\hat{m}}= {\left\{ \begin{array}{ll} \gamma _{m}e_{4}u,&{} \quad e_{4}u> 0\\ \gamma _{m}e_{4}u,&{} \quad e_{4}u\le 0,\hat{m}>\underline{m}\\ \gamma _{m},&{} \quad e_{4}u\le 0,\hat{m}\le \underline{m}\\ \end{array}\right. } \end{aligned}$$
(52)

where the initial value \(\hat{m}(0) \ge \underline{m}\). If the estimated value of \(\hat{m}\) is too small, then the control signal u will be too large. Thus \(\hat{m}\) has a wide range of changes resulting in \(\hat{m}=0\). In order to prevent this situation, we choose the initial value \(\hat{m}(0)=500\), and then \(\hat{m}\) can always be a large value.

The adaptive law of adaptive parameters are derived from Lyapunov stability analysis and chosen as

$$\begin{aligned} \dot{\hat{\theta }}_{g}= & {} {\left\{ \begin{array}{ll} \dot{\hat{\overline{\theta }}}_{g}=\overline{\gamma }_{g}e_{2}\hat{\alpha }_{g}\overline{\xi }_{g}(x) -2\overline{\lambda }_{g}\hat{\overline{\theta }}_{g}\\ \dot{\hat{\underline{\theta }}}_{g}=\underline{\gamma }_{g}e_{2}(1-\hat{\alpha }_{g})\underline{\xi }_{g}(x) -2\underline{\lambda }_{g}\hat{\underline{\theta }}_{g}\\ \end{array}\right. } \end{aligned}$$
(53)
$$\begin{aligned} \dot{\hat{\theta }}_{d}= & {} {\left\{ \begin{array}{ll} \dot{\hat{\overline{\theta }}}_{d}=\overline{\gamma }_{d}e_{3}\hat{\alpha }_{d}\overline{\xi }_{d}(x) -2\overline{\lambda }_{d}\hat{\overline{\theta }}_{d}\\ \dot{\hat{\underline{\theta }}}_{d}=\underline{\gamma }_{d}e_{3}(1-\hat{\alpha }_{d})\underline{\xi }_{d}(x) -2\underline{\lambda }_{d}\hat{\underline{\theta }}_{d}\\ \end{array}\right. } \end{aligned}$$
(54)
$$\begin{aligned} \dot{\hat{\theta }}_{h}= & {} {\left\{ \begin{array}{ll} \dot{\hat{\overline{\theta }}}_{h}=\overline{\gamma }_{h}e_{4}\hat{\alpha }_{h}\overline{\xi }_{h}(x) -2\overline{\lambda }_{h}\hat{\overline{\theta }}_{h}\\ \dot{\hat{\underline{\theta }}}_{h}=\underline{\gamma }_{h}e_{4}(1-\hat{\alpha }_{d})\underline{\xi }_{h}(x) -2\underline{\lambda }_{h}\hat{\underline{\theta }}_{h}\\ \end{array}\right. } \end{aligned}$$
(55)
$$\begin{aligned} \dot{\hat{\alpha }}_{g}= & {} \gamma _{\alpha _{g}}e_{2}\left( \overline{\xi }_{g}(x){\hat{\overline{\theta }}_{g}}^{T}- \underline{\xi }_{g}(x){\hat{\underline{\theta }}_{g}}^{T}\right) -2\lambda _{\alpha _{g}}\hat{\alpha }_{g} \end{aligned}$$
(56)
$$\begin{aligned} \dot{\hat{\alpha }}_{d}= & {} \gamma _{\alpha _{d}}e_{3}\left( \overline{\xi }_{d}(x){\hat{\overline{\theta }}_{d}}^{T}- \underline{\xi }_{d}(x){\hat{\underline{\theta }}_{d}}^{T}\right) -2\lambda _{\alpha _{d}}\hat{\alpha }_{d} \end{aligned}$$
(57)
$$\begin{aligned} \dot{\hat{\alpha }}_{h}= & {} \gamma _{\alpha _{h}}e_{4}\left( \overline{\xi }_{h}(x){\hat{\overline{\theta }}_{h}}^{T}- \underline{\xi }_{h}(x){\hat{\underline{\theta }}_{h}}^{T}\right) -2\lambda _{\alpha _{h}}\hat{\alpha }_{h} \end{aligned}$$
(58)

where \(\gamma =[\underline{\gamma }_{g},\overline{\gamma }_{g},\underline{\gamma }_{d},\overline{\gamma }_{d},\underline{\gamma }_{h}, \overline{\gamma }_{h},\gamma _{\alpha _{g}},\gamma _{\alpha _{d}},\gamma _{\alpha _{h}},\gamma _{m}]\) is the positive adaptation gain.

At the present stage, we summarize our main result in the following theorem, which shows that the designed controller guarantees the boundedness and stability of closed-loop system.

Theorem 1

Consider a flexible-joint manipulator system as shown in (3), the control input u in (39) with the IT2FNN-based adaptive laws given by (52)–(58) guarantees that all the signals in the resulting closed-loop systems are bounded. Furthermore, given an attenuation factor \(\rho \), the tracking performance of system will be satisfied

$$\begin{aligned} \sum _{i=1}^4\int _{0}^{T}e_{i}^2(s)\mathrm{d}s\le&\frac{1}{a_{0}}(V(0)+Tb_{0}\nonumber \\&+\sum _{i=2}^4\int _{0}^{T}\rho ^2J_{i}^2\mathrm{d}t, \quad T\in [0,\infty ] \end{aligned}$$
(59)

Proof of Theorem 1

To make the proof of stability more simple and clear, we define m(x), g(x), d(x), h(x) as \(f_{1}(x)\), \(f_{2}(x)\), \(f_{3}(x)\), \(f_{4}(x)\). Obviously, the approximation of m(x), g(x), d(x), h(x) is, respectively, \(\hat{f}_{1}(x)\), \(\hat{f}_{2}(x)\), \(\hat{f}_{3}(x)\), \(\hat{f}_{4}(x)\).

Consider the Lyapunov function candidate as

$$\begin{aligned} V=&\frac{1}{2}\sum _{i=1}^4 e_{i}^2+\frac{1}{2\gamma _{f_{1}}}{\tilde{f}_{1}}^{T}\tilde{f}_{1}+\sum _{i=2}^4\frac{1}{2\underline{\gamma }_{f_{i}}} {\tilde{\underline{\theta }}_{f_{i}}}^{T}\tilde{\underline{\theta }}_{f_{i}}\nonumber \\&+\sum _{i=2}^4\frac{1}{2\overline{\gamma }_{f_{i}}} {\tilde{\overline{\theta }}_{f_{i}}}^{T}\tilde{\overline{\theta }}_{f_{i}}+\sum _{i=2}^4\frac{1}{2\gamma _{\alpha _{f_{i}}}} {\tilde{\alpha }_{f_{i}}}^{T}\tilde{\alpha }_{f_{i}} \end{aligned}$$
(60)

The time derivative of V is

$$\begin{aligned} \dot{V}=&-\sum _{i=1}^4k_{i}{e_{i}}^{2}+\sum _{i=2}^4(f_{i}-\hat{f}_{i})e_{i}-\sum _{i=2}^4\frac{1}{\underline{\gamma }_{f_{i}}} \tilde{\underline{\theta }}_{f_{i}}^{T}\dot{\hat{\underline{\theta }}}_{f_{i}}\nonumber \\&-\sum _{i=2}^4\frac{1}{\overline{\gamma }_{f_{i}}}\tilde{\overline{\theta }}_{f_{i}}^{T}\dot{\hat{\overline{\theta }}}_{f_{i}} -\sum _{i=2}^4\frac{1}{\gamma _{\alpha _{f_{i}}}}{\tilde{\alpha }_{f_{i}}}^{T}\dot{\hat{\alpha }}_{f_{i}}\nonumber \\&+(f_{1}-\hat{f}_{1})ue_{4}-\frac{1}{\gamma _{f_{1}}}{\tilde{f}_{1}}^{T}\dot{\hat{f}}_{1} \end{aligned}$$
(61)

Applying (52), we get

$$\begin{aligned} \dot{V}\le&-\sum _{i=1}^4k_{i}{e_{i}}^{2}+\sum _{i=2}^4(f_{i}-\hat{f}_{i})e_{i}-\sum _{i=2}^4\frac{1}{\underline{\gamma }_{f_{i}}} \tilde{\underline{\theta }}_{f_{i}}^{T}\dot{\hat{\underline{\theta }}}_{f_{i}}\nonumber \\&-\sum _{i=2}^4\frac{1}{\overline{\gamma }_{f_{i}}}\tilde{\overline{\theta }}_{f_{i}}^{T}\dot{\hat{\overline{\theta }}}_{f_{i}} -\sum _{i=2}^4\frac{1}{\gamma _{\alpha _{f_{i}}}}{\tilde{\alpha }_{f_{i}}}^{T}\dot{\hat{\alpha }}_{f_{i}}\nonumber \\ \end{aligned}$$
(62)

Applying (49)–(51), \(\dot{V}\) can be rewritten as

$$\begin{aligned} \dot{V}\le&-\sum _{i=1}^4k_{i}{e_{i}}^{2}+\sum _{i=1}^4e_{i}\left[ (1-\hat{\alpha }_{f_{i}})\underline{\xi }_{f_{i}}(x){\tilde{\underline{\theta }}_{f_{i}}}^{T}\right. \nonumber \\&\left. +\,\hat{\alpha }_{f_{i}}\overline{\xi }_{f_{i}}(x){\tilde{\overline{\theta }}_{f_{i}}}^{T} +\left( \overline{\xi }_{f_{i}}(x){\hat{\overline{\theta }}_{f_{i}}}^{T}-\underline{\xi }_{f_{i}}(x){\hat{\underline{\theta }}_{f_{i}}}^{T}\right) \tilde{\alpha }_{f_{i}}\right] \nonumber \\&-\sum _{i=2}^4\frac{1}{\underline{\gamma }_{f_{i}}}\tilde{\underline{\theta }}_{f_{i}}^{T}\dot{\hat{\underline{\theta }}}_{f_{i}} -\sum _{i=2}^4\frac{1}{\overline{\gamma }_{f_{i}}}\tilde{\overline{\theta }}_{f_{i}}^{T}\dot{\hat{\overline{\theta }}}_{f_{i}}\nonumber \\&-\sum _{i=2}^4\frac{1}{\gamma _{\alpha _{f_{i}}}}{\tilde{\alpha }_{f_{i}}}^{T}\dot{\hat{\alpha }}_{f_{i}} +\sum _{i=2}^4e_{i}\left[ (1-\alpha _{f_{i}})\underline{\varepsilon }_{f_{i}}(x)\right. \nonumber \\&\left. +\,\alpha _{f_{i}}\overline{\varepsilon }_{f_{i}}(x) +\left( \overline{\xi }_{f_{i}}(x){\tilde{\overline{\theta }}_{f_{i}}}^{T}-\underline{\xi }_{f_{i}}(x) {\tilde{\underline{\theta }}_{f_{i}}}^{T}\right) \tilde{\alpha }_{f_{i}}\right] \nonumber \\ \le&-\sum _{i=1}^4k_{i}{e_{i}}^{2}+\sum _{i=1}^4\tilde{\underline{\theta }}_{f_{i}}^{T}\left[ e_{i}(1-\hat{\alpha }_{f_{i}})\underline{\xi }_{f_{i}}(x)\right. \nonumber \\&\left. -\frac{1}{\underline{\gamma }_{f_{i}}}\dot{\hat{\underline{\theta }}}_{f_{i}}\right] +\sum _{i=1}^4\tilde{\underline{\theta }}_{f_{i}}^{T}\left( e_{i} \hat{\alpha }_{f_{i}}\overline{\xi }_{f_{i}}(x) -\frac{1}{\overline{\gamma }_{f_{i}}}\dot{\hat{\overline{\theta }}}_{f_{i}}\right) \nonumber \\&+\sum _{i=2}^{4}{\tilde{\alpha }_{f_{i}}}^{T}\left[ e_{i}(\overline{\xi }_{f_{i}}(x){\hat{\overline{\theta }}_{f_{i}}}^{T}-\underline{\xi }_{f_{i}}(x) {\hat{\underline{\theta }}_{f_{i}}}^{T})-\frac{1}{\gamma _{\alpha _{f_{i}}}}\dot{\hat{\alpha }}_{f_{i}}\right] \nonumber \\&+\sum _{i=2}^4e_{i}\Bigg [(1-\alpha _{f_{i}})\underline{\varepsilon }_{f_{i}}(x)+\alpha _{f_{i}}\overline{\varepsilon }_{f_{i}}(x)\nonumber \\&+(\overline{\xi }_{f_{i}}(x){\tilde{\overline{\theta }}_{f_{i}}}^{T}-\underline{\xi }_{f_{i}}(x) {\tilde{\underline{\theta }}_{f_{i}}}^{T})\tilde{\alpha }_{f_{i}}\Bigg ] \end{aligned}$$
(63)

We define \(J_{i}=(1-\alpha _{f_{i}})\underline{\varepsilon }_{f_{i}}(x)+\alpha _{f_{i}}\overline{\varepsilon }_{f_{i}}(x) +(\overline{\xi }_{f_{i}}(x){\tilde{\overline{\theta }}_{f_{i}}}^{T}-\underline{\xi }_{f_{i}}(x) {\tilde{\underline{\theta }}_{f_{i}}}^{T})\tilde{\alpha }_{f_{i}},i=2,3,4\), and applying (53)–(58) into the time derivative of the Lyapunov function \(\dot{V}\), we have

$$\begin{aligned} \dot{V}\le&-\sum _{i=1}^4k_{i}{e_{i}}^{2}+\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} \tilde{\underline{\theta }}_{f_{i}}^{T}\hat{\underline{\theta }}_{f_{i}}+\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} \tilde{\overline{\theta }}_{f_{i}}^{T}\hat{\overline{\theta }}_{f_{i}}\nonumber \\&+\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}}{\tilde{\alpha }_{f_{i}}}^{T}\hat{\alpha _{f_{i}}}+\sum _{i=2}^{4}e_{i}J_{i} \end{aligned}$$
(64)

Setting \(c_{i}=k_{i}-\frac{1}{2\rho ^{2}}\), we have

$$\begin{aligned} \dot{V}\le&-\sum _{i=1}^4c_{i}{e_{i}}^{2}-\sum _{i=1}^4\frac{1}{2\rho ^{2}}{e_{i}}^{2}+\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} \tilde{\underline{\theta }}_{f_{i}}^{T}\hat{\underline{\theta }}_{f_{i}}\nonumber \\&+\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} \tilde{\overline{\theta }}_{f_{i}}^{T}\hat{\overline{\theta }}_{f_{i}} +\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}}{\tilde{\alpha }_{f_{i}}}^{T}\hat{\alpha }_{f_{i}}+\sum _{i=2}^{4}e_{i}J_{i} \end{aligned}$$
(65)

Since \(-\frac{1}{2a^2}b^2+bc \le \frac{1}{2}a^2c^2\), we have

$$\begin{aligned} \dot{V}\le&-\sum _{i=1}^4c_{i}{e_{i}}^{2}-\frac{1}{2\rho ^{2}}{e_{1}}^{2}+\sum _{i=2}^{4}\rho ^2J_{i}^2\nonumber \\&+\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} \tilde{\overline{\theta }}_{f_{i}}^{T}\hat{\overline{\theta }}_{f_{i}} +\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}}{\tilde{\alpha }_{f_{i}}}^{T}\hat{\alpha }_{f_{i}}\nonumber \\&+\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} \tilde{\underline{\theta }}_{f_{i}}^{T}\hat{\underline{\theta }}_{f_{i}}\nonumber \\ \le&-\sum _{i=1}^4c_{i}{e_{i}}^{2}-\frac{1}{2\rho ^{2}}{e_{1}}^{2}+\sum _{i=2}^{4}\rho ^2J_{i}^2\nonumber \\&+\sum _{i=2}^4\frac{\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} \left( 2{\underline{\theta }_{f_{i}}^{*}}^{T}\hat{\underline{\theta }}_{f_{i}}-2\hat{\underline{\theta }}_{f_{i}}^{T}\hat{\underline{\theta }}_{f_{i}}\right) \nonumber \\&+\sum _{i=2}^4\frac{\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} \left( 2{\overline{\theta }_{f_{i}}^{*}}^{T}\hat{\overline{\theta }}_{f_{i}}-2\hat{\overline{\theta }}_{f_{i}}^{T}\hat{\overline{\theta }}_{f_{i}}\right) \nonumber \\&+\sum _{i=2}^4\frac{\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} \left( 2{\alpha _{f_{i}}^{*}}^{T}\hat{\alpha }_{f_{i}}-2\hat{\alpha _{f_{i}}}^{T}\hat{\alpha }_{f_{i}}\right) \end{aligned}$$
(66)

Since \({a^{*}}^{T}a^{*}+\hat{a}^{T}\hat{a}\ge 2{a^{*}}^{T}\hat{a}\), and thus \(2{a^{*}}^{T}\hat{a}-2\hat{a}^{T}\hat{a}\le {a^{*}}^{T}a^{*}- \hat{a}^{T}\hat{a}\), we can get

$$\begin{aligned} \dot{V}\le&-\sum _{i=1}^4c_{i}{e_{i}}^{2}-\frac{1}{2\rho ^{2}}{e_{1}}^{2}+\sum _{i=2}^{4}\rho ^2J_{i}^2\nonumber \\&+\sum _{i=2}^4\frac{\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} \left( {\underline{\theta }_{f_{i}}^{*}}^{T}\underline{\theta }_{f_{i}}^{*}-\hat{\underline{\theta }}_{f_{i}}^{T}\hat{\underline{\theta }}_{f_{i}}\right) \nonumber \\&+\sum _{i=2}^4\frac{\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} ({\overline{\theta }_{f_{i}}^{*}}^{T}\overline{\theta }_{f_{i}}^{*}-\hat{\overline{\theta }}_{f_{i}}^{T}\hat{\overline{\theta }}_{f_{i}})\nonumber \\&+\sum _{i=2}^4\frac{\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} \left( {\alpha _{f_{i}}^{*}}^{T}\alpha _{f_{i}}^{*}-\hat{\alpha }_{f_{i}}^{T}\hat{\alpha }_{f_{i}}\right) \nonumber \\ \le&-\sum _{i=1}^4c_{i}{e_{i}}^{2}-\frac{1}{2\rho ^{2}}{e_{1}}^{2}+\sum _{i=2}^{4}\rho ^2J_{i}^2\nonumber \\&+\sum _{i=2}^4\frac{\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} (-{\underline{\theta }_{f_{i}}^{*}}^{T}\underline{\theta }_{f_{i}}^{*}-\hat{\underline{\theta }}_{f_{i}}^{T}\hat{\underline{\theta }}_{f_{i}})\nonumber \\&+\sum _{i=2}^4\frac{\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} \left( -{\overline{\theta }_{f_{i}}^{*}}^{T}\overline{\theta }_{f_{i}}^{*}-\hat{\overline{\theta }}_{f_{i}}^{T}\hat{\overline{\theta }}_{f_{i}}\right) \nonumber \\&+\sum _{i=2}^4\frac{\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} \left( -{\alpha _{f_{i}}^{*}}^{T}\alpha _{f_{i}}^{*}-\hat{\alpha }_{f_{i}}^{T}\hat{\alpha }_{f_{i}}\right) \nonumber \\&+\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} {\underline{\theta }_{f_{i}}^{*}}^{T}\underline{\theta }_{f_{i}}^{*} +\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} {\alpha _{f_{i}}^{*}}^{T}\alpha _{f_{i}}^{*}\nonumber \\&+\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} {\overline{\theta }_{f_{i}}^{*}}^{T}\overline{\theta }_{f_{i}}^{*} \end{aligned}$$
(67)

Since \(\tilde{a}^{T}\tilde{a}=(a^{*}-\hat{a})^{T}(a^{*}-\hat{a})={a^{*}}^{T}a^{*}-2{a^{*}}^{T}\hat{a}+\hat{a}^{T}\hat{a}\le 2{a^{*}}^{T}a^{*}+2\hat{a}^{T}\hat{a}\), we have \(-\frac{1}{2}\tilde{a}^{T}\tilde{a}\ge -\hat{a}^{T}\hat{a}-{a^*}^Ta^{*}\). The time derivative of the Lyapunov function V can be obtained as follows

$$\begin{aligned} \dot{V}\le&-\sum _{i=1}^4c_{i}{e_{i}}^{2}-\frac{1}{2\rho ^{2}}{e_{1}}^{2}+\sum _{i=2}^{4}\rho ^2J_{i}^2\nonumber \\&-\sum _{i=2}^4\frac{\underline{\lambda }_{f_{i}}}{2\underline{\gamma }_{f_{i}}} \tilde{\underline{\theta }}_{f_{i}}^{T}\tilde{\underline{\theta }}_{f_{i}} -\sum _{i=2}^4\frac{\overline{\lambda }_{f_{i}}}{2\overline{\gamma }_{f_{i}}} \tilde{\overline{\theta }}_{f_{i}}^{T}\tilde{\overline{\theta }}_{f_{i}}\nonumber \\&-\sum _{i=2}^4\frac{\lambda _{\alpha _{f_{i}}}}{2\gamma _{\alpha _{f_{i}}}} \tilde{\alpha }_{f_{i}}^{T}\tilde{\alpha }_{f_{i}} +\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} {\underline{\theta }_{f_{i}}^{*}}^{T}\underline{\theta }_{f_{i}}^{*}\nonumber \\&+\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} {\alpha _{f_{i}}^{*}}^{T}\alpha _{f_{i}}^{*} +\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} {\overline{\theta }_{f_{i}}^{*}}^{T}\overline{\theta }_{f_{i}}^{*} \end{aligned}$$
(68)

To guarantee \(k_{i}\ge \frac{1}{2\rho ^{2}}\), we define \(c_{i},i=1,\ldots ,4\) is a positive constant and \(c_{0}=\min \{\frac{2\rho ^2c_{1}+1}{\rho ^2},2c_{i},\underline{\lambda }_{f_{i}},\overline{\lambda }_{f_{i}},\lambda _{\alpha _{f_{i}}};i=2,3,4\}\). The time derivative of the Lyapunov function \(\dot{V}\) can be rewritten as

$$\begin{aligned} \dot{V}\le&-c_{0}\left( \sum _{i=1}^4 \frac{1}{2}e_{i}^2+\frac{1}{2\gamma _{f_{1}}}{\tilde{f}_{1}}^{T}\tilde{f}_{1}+\sum _{i=2}^4\frac{1}{2\underline{\gamma }_{f_{i}}} {\tilde{\underline{\theta }}_{f_{i}}}^{T}\tilde{\underline{\theta }}_{f_{i}}\right. \nonumber \\&\left. +\sum _{i=2}^4\frac{1}{2\overline{\gamma }_{f_{i}}} {\tilde{\overline{\theta }}_{f_{i}}}^{T}\tilde{\overline{\theta }}_{f_{i}}+\sum _{i=2}^4\frac{1}{2\gamma _{\alpha _{f_{i}}}} {\tilde{\alpha }_{f_{i}}}^{T}\tilde{\alpha }_{f_{i}}\right) \nonumber \\&+\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} {\underline{\theta }_{f_{i}}^{*}}^{T}\underline{\theta }_{f_{i}}^{*}+\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}}{\overline{\theta }_{f_{i}}^{*}}^{T}\overline{\theta }_{f_{i}}^{*}\nonumber \\&+\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} {\alpha _{f_{i}}^{*}}^{T}\alpha _{f_{i}}^{*} +\frac{c_{0}}{2\gamma _{f_{1}}}{\tilde{f}_{1}}^{T}\tilde{f}_{1}+\sum _{i=2}^{4}\rho ^2J_{i}^2 \end{aligned}$$
(69)

Define \(c_{V\max }=\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} {\underline{\theta }_{f_{i}}^{*}}^{T}\underline{\theta }_{f_{i}}^{*}+\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}}{\overline{\theta }_{f_{i}}^{*}}^{T}\overline{\theta }_{f_{i}}^{*} +\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} {\alpha _{f_{i}}^{*}}^{T}\alpha _{f_{i}}^{*} +\frac{c_{0}}{2\gamma _{f_{1}}}{\tilde{f}_{1}}^{T}\tilde{f}_{1}+\sum _{i=2}^{4}\rho ^2J_{i}^2\), we get

$$\begin{aligned} \dot{V}\le c_{0}V+c_{V\max } \end{aligned}$$
(70)

Integrating (70) over [0, t], we have

$$\begin{aligned} V(t)&\le V(0)\exp (-c_{0}t)+\frac{c_{V\max }}{c_{0}}(1-\exp (-c_{0}t))\nonumber \\&\le V(0)+\frac{c_{V\max }}{c_{0}},t\ge 0 \end{aligned}$$
(71)

We define the tight set \(\Omega _{0}=\{X|V(X)\le C_{0}\}\), where \(C_{0}=V(0)+\frac{c_{V\max }}{c_{0}}\). Then we can conclude that all the signals in the closed-loop system are bounded.

Setting \(a_{0}=\min \{c_{1}+\frac{1}{2\rho ^2},c_{i},i=2,3,4\}\) and (68) can be rewritten as

$$\begin{aligned} \dot{V}\le&-a_{0}\sum _{i=1}^4{e_{i}}^{2}+\sum _{i=2}^{4}\rho ^2J_{i}^2-\sum _{i=2}^4\frac{\underline{\lambda }_{f_{i}}}{2\underline{\gamma }_{f_{i}}} \tilde{\underline{\theta }}_{f_{i}}^{T}\tilde{\underline{\theta }}_{f_{i}}\nonumber \\&-\sum _{i=2}^4\frac{\overline{\lambda }_{f_{i}}}{2\overline{\gamma }_{f_{i}}} \tilde{\overline{\theta }}_{f_{i}}^{T}\tilde{\overline{\theta }}_{f_{i}} -\sum _{i=2}^4\frac{\lambda _{\alpha _{f_{i}}}}{2\gamma _{\alpha _{f_{i}}}} \tilde{\alpha }_{f_{i}}^{T}\tilde{\alpha }_{f_{i}}+b_{0}\nonumber \\ \le&-a_{0}\sum _{i=1}^4{e_{i}}^{2}+b_{0}+\sum _{i=2}^{4}\rho ^2J_{i}^2 \end{aligned}$$
(72)

where \(b_{0}=\sum _{i=2}^4\frac{2\underline{\lambda }_{f_{i}}}{\underline{\gamma }_{f_{i}}} {\underline{\theta }_{f_{i}}^{*}}^{T}\underline{\theta }_{f_{i}}^{*} +\sum _{i=2}^4\frac{2\lambda _{\alpha _{f_{i}}}}{\gamma _{\alpha _{f_{i}}}} {\alpha _{f_{i}}^{*}}^{T}\alpha _{f_{i}}^{*} +\sum _{i=2}^4\frac{2\overline{\lambda }_{f_{i}}}{\overline{\gamma }_{f_{i}}} {\overline{\theta }_{f_{i}}^{*}}^{T}\overline{\theta }_{f_{i}}^{*}\)

Integrating (72) over [0, t], we have

$$\begin{aligned} \int _{0}^{T}\dot{V}\hbox {d}t\le&-\int _{0}^{T}a_{0}\sum _{i=1}^4\int _{0}^{T}e_{i}^2(s)\hbox {d}s+Tb_{0}\nonumber \\&+\sum _{i=2}^4\int _{0}^{T}\rho ^2J_{i}^2\hbox {d}t \end{aligned}$$
(73)

Since \(\int _{0}^{T}\dot{V}\hbox {d}t=V(T)-V(0)\), we have

$$\begin{aligned} \sum _{i=1}^4\int _{0}^{T}e_{i}^2(s)\hbox {d}s\le&-\frac{1}{a_{0}}\left( V(0)-V(T)+Tb_{0}\right. \nonumber \\&\left. +\sum _{i=2}^4\int _{0}^{T}\rho ^2J_{i}^2\hbox {d}t\right) \end{aligned}$$
(74)

Since \(-\frac{1}{a_{0}}V(T)\le 0\), we get

$$\begin{aligned} \sum _{i=1}^4\int _{0}^{T}e_{i}^2(s)\hbox {d}s\le&\frac{1}{a_{0}}\left( V(0)+Tb_{0}\right. \nonumber \\&\left. +\sum _{i=2}^4\int _{0}^{T}\rho ^2J_{i}^2\hbox {d}t\right) \end{aligned}$$
(75)

That is, given an attenuation factor \(\rho \), the accuracy of tracking error is determined by the upper bound of approximation error. Thus, the theorem is proved. \(\square \)

5 Experimental results

In this section, we demonstrate the effectiveness of the proposed scheme on a single-link flexible-joint manipulator. The actual values of parameters for dynamic equations (3) are \(M=0.2\,\hbox {kg}\), \(L=0.02\,\hbox {m}\), \(I=1.35\times 10^{-3}\,\hbox {kg}\,\hbox {m}^{2}\), \(K=7.47 N \, \hbox {m/rad}\), \(J=2.16\times 10^{-1}\,\hbox {kg}\,\hbox {m}^{2}\). Three IT2FNN are used to approximate the nonlinear functions g(x), d(x) and h(x). \(x=[x_{1},x_{2},x_{3},x_{4}]\) is the vector input. For each input \(x_{i}\), type-2 Gaussian membership functions are chosen as

$$\begin{aligned} \hat{F}_{i}^{j}= {\left\{ \begin{array}{ll} \mu _{\underline{F}_{i}^{j}}(x_{i})=a\exp \left( -\frac{1}{2}\left( \frac{x_{i}+c_{j}}{\sigma _{j}}\right) \right) \\ \mu _{\overline{F}_{i}^{j}}(x_{i})=\exp \left( -\frac{1}{2}\left( \frac{x_{i}+c_{j}}{\sigma _{j}}\right) \right) \\ \end{array}\right. } \end{aligned}$$
(76)

where \(i=1,2,3,4\), \(j=1,2,3\), \(c=[c_{1},c_{2},c_{3}]=[1.25,0,-\,1.25]\), \(\sigma =[\sigma _{1},\sigma _{2}, \sigma _{3}]=[0.6,0.6,0.6]\) and \(a=0.8\).

The other design parameters are chosen as: \(\underline{m}=1\), \(\gamma =[\underline{\gamma }_{g},\overline{\gamma }_{g},\underline{\gamma }_{d},\overline{\gamma }_{d},\underline{\gamma }_{h}, \overline{\gamma }_{h},\gamma _{\alpha _{g}},\gamma _{\alpha _{d}},\gamma _{\alpha _{h}},\gamma _{m}]=[200,500,200,500,200,500,0.05,0.05,0.05,0.006]\), \(\lambda =[\underline{\lambda }_{g},\overline{\lambda }_{g},\underline{\lambda }_{d},\overline{\lambda }_{d},\underline{\lambda }_{h}, \overline{\lambda }_{h},\lambda _{\alpha _{g}},\lambda _{\alpha _{d}},\lambda _{\alpha _{h}}]=[10,10,10,10,11.25,11.25,0.001,0.001,0.001]\).

The initial conditions are set as: \(x(0)=[x_{1}(0),x_{2}(0),x_{3}(0),x_{4}(0)]=[0,0,0,0]\), \(\hat{\theta }_{g}(0)=[\hat{\underline{\theta }}_{g}(0),\hat{\overline{\theta }}_{g}(0)]=[1.2,1.2]\), \(\hat{\theta }_{d}(0)=[\hat{\underline{\theta }}_{d}(0),\hat{\overline{\theta }}_{d}(0)]=[1.2,1.2]\), \(\hat{\theta }_{h}(0)=[\hat{\underline{\theta }}_{h}(0),\hat{\overline{\theta }}_{h}(0)]=[1.2,1.2]\), \(\hat{\alpha }_{g}(0)=0\), \(\hat{\alpha }_{d}(0)=0\), \(\hat{\alpha }_{h}(0)=0\), \(\hat{m}(0)=500\). The desired output trajectory is designed to be \(y_{d}=0.2\sin (t)\). The control target is that the system output can track the desired trajectory even in the case of external disturbance \(d(t)=0.05\cos (2t)\).

To evaluate the performance of all controllers with difference approximator (T1FNN, NN, IT2FNN) clearly, we use the following performance criterions: the integral of square error (ISE), the integral of the absolute value of the error (IAE) and the integral of the time multiplied by the absolute value of the error (ITAE), which can be expressed as:

$$\begin{aligned} \hbox {ISE}&= \int \limits _0^\infty {{{\left[ {e(t)} \right] }^2}\hbox {d}t} \nonumber \\ \hbox {IAE}&= \int \limits _0^\infty {\left| {e(t)} \right| \hbox {d}t} \nonumber \\ \hbox {ITAE}&= \int \limits _0^\infty {t\left| {e(t)} \right| \hbox {d}t} \end{aligned}$$
(77)
Fig. 3
figure 3

Responses of \(q_1\) and \(\dot{q}_1\)

Fig. 4
figure 4

Responses of \(q_2\) and \(\dot{q}_2\)

Fig. 5
figure 5

Trajectories of control input

Fig. 6
figure 6

Position tracking performance

Simulation results demonstrate the superior performance of the proposed controller. The responses of the system states \(q_1\), \(\dot{q}_{1}\), \(q_2\), \(\dot{q}_{2}\) are illustrated in Figs. 3 and 4. The trajectory of control input is shown in Fig. 5. It is clear that the control input is bounded. Tracking curves are depicted in Fig. 6. We can find that three approximators have the ability to achieve precise tracking. However, the accuracy of each approximator is different. Figure 7 demonstrates that the proposed controller has better steady-state performance, less fluctuation and higher approximation accuracy than T1FNN and NN. It is noted that the tracking error tends to a small neighborhood of zero even though there is external disturbance.

Table 1 lists the ISE, IAE and ITAE for all controllers. The values of the ISE, the IAE and the ITAE for the proposed controller with the IT2FNN approximator are lower than those obtained for the T1FNN and the NN approximator. It is clear that the adaptive backstepping controller with the IT2FNN approximator is able to achieve better tracking performance and high accuracy.

Fig. 7
figure 7

Position tracking error

Table 1 Performance index

Compared to the existing control method for flexible-joint manipulator in recent years [46], the adaptive backstepping controller with the IT2FNN approximator has the obvious advantages in transient tracking performance and high accuracy. From the results of simulation without external disturbance in [46], it can be seen that the maximum tracking error is about 0.25 rad and the tracking error tends to a small neighborhood of zero at about 3.75 s. In this study, the maximum tracking error is 0.02 rad and reduces to less than 0.005 rad at 2.57 s even though there is external disturbance. Starting from 3.08 s, the range of the tracking error is stable between 0.002 rad and \(2.1\hbox {e}^{-5}\,\hbox {rad}\). The time it takes for the tracking error to a small neighborhood of zero is about 0.31 times shorter than the control method in [46]. It can demonstrate that the proposed method significantly eliminates the undesirable overshoot and reduces the settling time. Therefore, the proposed adaptive backstepping controller with the IT2FNN approximator can obtain better steady-state performance and improve approximation accuracy.

6 Conclusions

In this paper, an adaptive backstepping control scheme based on IT2FNN approximator has been proposed for a flexible-joint manipulator with mismatched uncertainties. The IT2FNN is used to approximate the unknown functions. The stability analysis of the proposed scheme is derived. The proposed adaptive controller guarantees that all the signals in the resulting closed-loop systems are bounded. Finally, the simulation results of the comparative study illustrate that the performances of adaptive backstepping control based on IT2FNN approximator over the others.