1 Introduction

While the Interval type-2 (IT2) systems arise as an alternative to manage uncertainties present in all industrial processes, the type-1 (T1) models such as T1 singleton fuzzy logic or their equivalent radial basis function neural network (RBFNN), also called T1 RBFNN, both cannot manage uncertainties [1]. T1 models require several cycles, epochs or iterations of training, and adjustment to get an acceptable result or an adequate level of precision. For example in [2], 1300 epochs of training were required to get an adequate result. Therefore, as mentioned by [3], the main problem to obtain better precision output relies on the modeling of the system. Most researchers base their modeling by intuition and as a thumb rule because there are not specific criteria for determining how many rules are needed to model this kind of systems.

Specialized literature shows a couple of techniques to model intelligent systems [4,5,6,7,8,9,10,11], but all of them are applied to T1 models. Only a pair of proposals was found that uses the IT2 [2, 11]. In [4, 5] has been proposed a model that uses linguistic labels as fuzzy sets, but in [5] the center label is missing to create the universe of discourse (UOD) and with them the fuzzy rule base. In [4] the model is based in the Gaussian data distribution, but they do not consider the mean which means that the model is restricted to have odd input variables, while in [5] the model is restricted to even input variables.

In [6] a method that uses control charts to model the rule base was developed, but it was restricted to use a genetic algorithm in order to generate the membership functions (MF), the support or width of the fuzzy sets is non-uniform, and finally a combination of trapezoidal and triangular MFs is produced to get the UOD. In [7] a similar method presented in [6] was used, but it only uses triangular MFs and presents the presence of blank spaces in the UOD. In [8] five linguistic labels have been used which are restricted to a constant rule base that is inflexible and unadaptable with blank spaces in the UOD.

The statistical properties of the Gaussian distribution were used in [3] and [9] to model the fuzzy rule base. The central composite design (CCD) of the design of experiments was used in [10, 11] to obtain a simplified and compact rule base to assemble the UOD. In [10] a T1 RBFNN that needs to be trained was proposed. Finally, in [11] the technique of CCD was used to model an IT2 rule base that considers uncertainties.

In the classic approach of the RBFNN or T1 RBFNN, only two types of membership functions can be used, the logistic function, see Fig. 1, or the Gaussian function, see Fig. 2, that are defined as receptive fields or fuzzifiers [12] (which were presented by [3]).

Fig. 1
figure 1

Logistic function

Fig. 2
figure 2

Gaussian function

1.1 Related works

Using the phrase of Mendel [1], the IT2 RBFNN technology is going through its infancy phase because it is a new technique developed in 2015 by Rubio-Solis and Panoutsos in [2], therefore, there are very few related works. Between the theoretical proposals for modeling and assembling IT2 RBFNN network, it can be mentioned the framework presented in [2] to model an IT2 network for the first time, where the necessity of expert knowledge to assemble the IT2 is mentioned. Nonetheless, in [1] is stated that “Rules may be provided by experts or can be extracted from numerical data”, which is demonstrated in [3, 4, 9,10,11].

Theoretical basis about IT2 was developed since 2015 in [2, 10, 13,14,15,16]; however, your application has barely been shown in works as [13,14,15,16,17,18,19,20,21,22,23,24,25,26,27,28,29]; this fact shows that this is an emerging and current technology. Among the works found in the current literature appears: a general IT2 network presented in [13], which is restricted to only one type of membership function, as a Gaussian function. A recurrent self-evolving RBFNN is presented in [14] to adapt the neurons in the hidden layer of the network in a non-linear dynamical system. In [15] a model to convert the IT2 Mandami model into an IT2 Takagi–Sugeno model via genetic algorithms was proposed, Baklouti et al. [16] applied an IT2 RBFNN to type Takagi–Sugeno-Kang (TSK) in order to evaluate time series. From [17] could be obtained the basis of interval type-2 fuzzy logic systems that serves as basis to generate the IT2 RBFNN.

In [18], the use of an IT2 RBFNN to avoid the numerous arithmetic operations required by the interval type-2 fuzzy models, which reduce their complexities and their type of reduction process was proposed. In [16], a theoretical method called Beta Basis Function Interval type-2 fuzzy neural network (BIT2FNN) in a TSK model to manage the uncertainties and noise was proposed. In [19], a classifier using IT2 RBFNN to recognize alphabets and manage the noise present in the upcoming signals was presented. The classifier uses the IT2 RBFNN to avoid the problem of the non-linearities and noise present in time series with a recurrent interval type-2 fuzzy neural network or RIT2IFNN with a TSK model [20]. In [21], a multilayer type-2 extreme learning machine method for classifying and recognizing walking was presented. In [22], a method to forecast and adjust the neural network using backpropagation algorithm based on general type-2 model was developed. In [23], the IT2 radial basis function network was used in a rail manufacturing process to classify. In [24], the IT2 RBFNN was used in a rotary steerable system and in directional drilling for oil and gas exploration in prospection. In [25], a classifier with IT2 RBFNN based on clustering was developed, it was used to deal with the uncertainties present in the data and adjust the connection of the weights. In [26], a recurrent IT2 RBFNN was used as an adaptive network for a micro-electro-mechanical system of a gyroscope to control the non-singular sliding control. In [27], the IT-2 RBFNN was used to overcome the control of ammonia flow in a selective catalytic reduction, the IT2 RBFNN aided to modeling uncertainties and constrain the predictive control. In [28], the IT2 RBFNN was used to identify and online prediction of the glucose level on diabetes patients. In [29], the IT2 fuzzy radial basis function neural network (IT2 FRBFNN) was used to sliding mode controls of non-linear systems to approximate the sliding. Finally, in [30], a general type-2 radial basis function neural network (GT2 RBFNN) was developed to manage the trajectory of a remote operated underwater vehicle (ROV).

1.2 Contributions

The main contribution of this paper is the presentation of a novel method for modeling IT2 RBFNN using the CCD technique that is part of a methodology used in the design of experiments to model a compact and simple fuzzy rule base. Other contribution is the implementation of an application of the IT2 RBFNN in a real industrial process using image processing to evaluate quality features, this combination of processes is not documented in the current literature. Additionally, the implementation of two membership functions from fuzzy logic: the Huygen’s tractrix or unnamed function (Fig. 3) and the triangular function (Fig. 4), called in this form in [17]. Both used for the first time as RBF in this work since, according to our knowledge, there is no evidence that they have been used in the literature as RBF.

Fig. 3
figure 3

Huygen’s tractrix function

Fig. 4
figure 4

Triangular function

2 Theoretical foundations

2.1 T1 Radial basis function neural networks

This kind of neural network is a type of interpolation focused on getting approximations based on a distance between a pattern and a sample [12] and is graphically depicted in Fig. 5. The T1 RBFNN is similar to the K-nearest neighbor’s (KNN) classifier.

Fig. 5
figure 5

Two input case topology of a T1 RBFNN

The T1 RBFNN operates with assigned weights in the inputs, each of them is placed in a multidimensional vector given by Eq. (1) and is based on a Gaussian distribution function Eq. (2) or by a logistic distribution function Eq. (3). The Gaussian distribution function is equivalent to a membership function in fuzzy models. The output of the T1 RBFNN model could be calculated in many forms Eqs. (4)–(7).

$${w}_{i}={R}_{i}\left(x\right)={R}_{i}\left(\Vert x-{u}_{x}\Vert \frac{{u}_{i}}{{\sigma }_{x}}\right) \forall x,u \in U$$
(1)

where ux is the mean in the radial basis function, x is the input of the function, and \({\sigma }_{x}\) is the spread of the function.

$$R\left(x\right)={e}^{- \frac{\Vert {x-{u}_{x}\Vert }^{2}}{2{{\sigma }_{x}}^{2}}}\forall x,u \in U$$
(2)
$$R\left(x\right)=\frac{1}{1+\frac{{e}^{[{\Vert x-{u}_{i}\Vert ]}^{2}}}{{\sigma }_{i}^{2}}} \forall x,u \in U$$
(3)
$$d\left(x\right)=\sum_{i=1}^{H}{c}_{i}{w}_{i}\forall c,R \in x$$
(4)
$$d\left(x\right)=\sum_{i=1}^{H}{c}_{i}{R}_{i}\forall c,R \in x$$
(5)
$$d\left(x\right)=\frac{\sum_{i=1}^{H}{c}_{i}{w}_{i}}{\sum_{i=1}^{H}{w}_{i}}$$
(6)
$$d\left(x\right)=\frac{{c}_{i}{R}_{i}(x)}{{R}_{i}(x)}\forall c,R \in x$$
(7)

where ci is the output in the dataset for the prediction and the final approximation is obtained by Eqs. (45, 6 or 7) which can be reinterpreted as a center of gravity deffuzzifier in fuzzy model Eq. (8).

$${y}_{c}\left(x\right)=\frac{\sum_{1}^{n}{y}_{i}{\mu }_{B}({y}_{i})}{\sum_{1}^{n}{\mu }_{B}({y}_{i})}$$
(8)

2.2 IT2 fuzzy logic system

The IT2 arises from the union of two T1 fuzzy sets represented by Eq. (9) to generate an interval. An IT2 fuzzy set is given by Eq. (10) and is characterized by \(\tilde{A }\) with a membership function \({\mu }_{\tilde{A }}\left(x,u\right)\),

$$A=\left\{\left(x,{\mu }_{A}\left(x\right)\right)/\forall x\in X\right\}$$
(9)

where A represents the set and \({\mu }_{A}\) is the grade of membership of some x′ in A.

$$\tilde{A }= \left\{\left(\left(x,u\right),{\mu }_{\tilde{A }}\left(x,u\right)\right)|\forall x\in X,\forall u\in {J}_{x}\subseteq \left[\mathrm{0,1}\right]\right\}$$
(10)

The classic approach for the IT2 models is the Gaussian fuzzifier defined on Eq. (11), showing a difference against the T1 membership function that presents two means,

$${{\mu }_{\tilde{A }}}_{k}^{i}\left({x}_{k}\right)={\tilde{F }}_{k}^{i}=\mathrm{exp}\left[-\frac{1}{2}{\left[\frac{{x}_{k}-{m}_{k}^{i}}{{\sigma }_{k}^{i}}\right]}^{2}\right]$$
(11)

where \({m}_{k}^{i}\in \left[{m}_{k1}^{i},{m}_{k2}^{i}\right]\) is the uncertain mean, k = 1,2,…,p (p is the number of inputs) and = 1,2,..M (the number of M rules), and \({\sigma }_{k}^{i}\) is the standard deviation.

2.3 IT2 Radial basis function neural networks

IT2 RBFNN has been presented for the first time in 2015 by [2], since then it has evolved into new models such as the ones presented by Mandami [2], Takagi–Sugeno [14], and Takagi–Sugeno-Kang model [15].

The basis of this model is the extension of the model shown by [12] to obtain the left and right functions (Fig. 6). Basically the similitudes of the T1 fuzzy model are equivalent to the ones in T1 RBFNN, which are well known since 1993 in [16,17,18].

Fig. 6
figure 6

IT2 RBFNN topology, extracted from [2]

Fig. 7
figure 7

2 k CCD model

The IT2 RBFNN algorithm is generated with the adjustment of T1 RBFNN. First, Eq. (2) is replaced by Eq. (11). Second, Eqs. (47) are converted to Eqs. (1215) to get the lower output and similarly Eqs. (1619) to get the upper output.

$$\underline{d(x)}={\textstyle\sum_{i=1}^H}\underline{c_iR_i}(x)\forall c,R\in x$$
(12)
$$\underline{d(x)}={\textstyle\sum_{i=1}^H}\underline{c_iw_i}\forall c,R\in x$$
(13)
$$\underline{d(x)}=\frac{\sum_{i=1}^H\underline{c_iw_i}}{\sum_{i=1}^H\underline{w_i}}$$
(14)
$$\underline{d(x)}=\frac{\underline{c_iR_i}(x)}{\underline{R_i}(x)}\forall c,R\in x$$
(15)
$$\overline{d\left(x\right)}={\textstyle\sum_{i=1}^H}{\overline c}_i{\overline R}_i(x)\forall c,R\in x$$
(16)
$$\overline{d\left(x\right)}={\textstyle\sum_{i=1}^H}{\overline c}_i{\overline w}_i\forall c,R\in x$$
(17)
$$\overline{d\left(x\right)}=\frac{\sum_{i=1}^{H}{\overline{c}}_{i}{\overline{w}}_{i}}{\sum_{i=1}^{H}\overline{{w}_{i}}}$$
(18)
$$\overline{d\left(x\right)}=\frac{{\overline{c}}_{i}{\overline{R}}_{i}(x)}{{\overline{R}}_{i}(x)}\forall c,R \in x$$
(19)

where ci is the output in the database for the prediction and the low and high are obtained by applying Eqs. (2021). The final approximation is obtained by Eq. (22) which can be reinterpreted as Eq. (23).

$$d\left(x_i\right)=\frac{\overline{d\left(x_i\right)}+\underline{d(x_i)}}2$$
(20)
$$d\left({x}_{i}\right)=\frac{\left(\frac{{\overline{c}}_{i}{\overline{R}}_{i}({x}_{i})}{{\overline{R}}_{i}({x}_{i})}\right)+\left(\frac{{\underline{c}}_{i}{\underline{R}}_{i}({x}_{i})}{{\underline{R}}_{i}({x}_{i})}\right)}{2}$$
(21)
$$d\left({x}\right)=\frac{\sum_{i=1}^{n}d({x}_{i})}{2}$$
(22)
$$d\left(x\right)=\frac{\left(\frac{{\overline{c}}_{i}{\overline{R}}_{i}(x)}{{\overline{R}}_{i}(x)}\right)+\left(\frac{{\underline{c}}_{i}{\underline{R}}_{i}(x)}{{\underline{R}}_{i}(x)}\right)}{2}$$
(23)

A. Central composite design

The factorial design or CCD is a technique used to analyze the factors and their possible correlation. These factors present limits to define the universe, and their levels are called low and high. In order to establish the model, a combinatorial via permutations is needed to define a pattern where one of the variables must change, meanwhile the rest remain constant. The simplest model is called 2 k (Fig. 7).

In Fig. 7, each sing at a corner represents different levels of variables. For (1), both variables are on low level, for (a) the first variable is on high level, and the second one is on the low level, for (b) the first variable is on the low level and the second on the high level, and for (ab) both variables are on the high level.

3 Revisited equivalence of the T1 FLS and T1 RBFNN

In [2], a series of restrictions have been established to this equivalence as follows:

  • The number of perceptive fields in the hidden layer is equal to the number of fuzzy rules.

First, a definition for the receptive field is required.

3.1 Definition 1: a receptive field is a neuron in the network that represents a mathematical operation such as Eq. (2) or Eq. (3).

But this restriction needs to be enhanced, and Fig. 5 needs to be redrawn since the receptive field on this figure only represents a part of the fuzzy rule, e.g., in T1 model a fuzzy rule is defined by Eq. (24), every variable in the T1 RBFNN requires a receptive field represented mathematically by Eq. (2) or Eq. (3), then the number of receptive fields (rf) required is equal to the number of variables (V) multiplied by the number of rules (r), given by Eq. (25). The enhanced graphical representation is depicted in Fig. 8,

$$Rule i:IF\times 1\;is\;a\;and\;\times 2\;is\;b\;then\;y\;is\;G$$
(24)

where: × 1, × 2 are the inputs, and G is the output of the rule.

Fig. 8
figure 8

T1 RBFNN of T1 topology, two input case expanded

$$rf=V*r$$
(25)

The IT2 RBFNN topology of [2] shown in Fig. 6 needs to be redrawn (Fig. 9) because Fig. 6 has additional receptive fields; this condition requires additional calculus and training to tuning precise output, also require additional receptive fields to achieve a precise output. Then, the quantity of receptive fields is given by Eq. (26), and the topology is reorganized with the addition of the new receptive fields,

$$rf=2(V\times r)$$
(26)

where: rf are the receptive fields, V is the number of variables and, r is the number of rules.

Fig. 9
figure 9

IT2 RBFNN topology, two input case expanded

  • The membership function within each rule is chosen as Gaussian.

But this fact requires testing because in [12] is mentioned that the logistic function is an alternative model of RBF that needs to be tested. Therefore, a definition of membership function is required in the neural network.

3.2 Definition 2: The membership function for neural networks is a function that defines the weight for the input variable.

  • The T-norm operator used to compute each rule’s firing strength is the multiplication.

The T-norm has been defined in [31] as the intersection of two sets. But the intersection in the case of T1 RBFNN is not used in the receptive fields as occurs with the fuzzy rules before fuzzification in the FLS.

4 Proposal

A. Assemble and calculation of IT2 RBFNN parameters based on the CCD

In CCD and [11], the IT2 CCD is obtained and used to generate the fuzzy rule base that serves as input for the IT2 RBFNN.

The initial calculi of the rule base or the antecedents for the IT2 RBFNN are given by Eq. (27), as were proposed in [11], where 2 k is the matrix that conforms the CCD model, and every pair in the matrix is formed by the possible combinations of the lower and upper limits of control for the variables. These parameters or receptive fields for the rules or the rf are obtained from the process control specifications, and their equivalences to the CCD 2 k model are presented in Table 1.

Table 1 Equivalences of CCD states and quality control limits
$${2}^{k}=\left[\begin{array}{c} \begin{array}{cc}LCLa& LCLb\end{array}\\ \begin{array}{cc}LCLa& UCLB\end{array}\\ \begin{array}{cc}UCLA& LCLb\end{array}\\ \begin{array}{cc}UCLA& UCLB\end{array}\end{array}\right]$$
(27)

From Eq. (27), the inputs are obtained for the receptive fields on the IT2 RBFNN. The input for the first receptive field in the T1 RBFNN is LCLa, and for calculating the receptive fields for the first variable in IT2 RBFNN, additional calculations are needed. Firstly, it is required the spread of the data specifications, which is given by Eq. (28),

$${\sigma }_{{x}_{i}}=\sum_{i=1}^{n}\left(\frac{{x}_{i}-{\overline{x} }_{i}}{n}\right)$$
(28)

The IT2 matrix is calculated using the matrix presented in Eq. (27) and \({\sigma }_{{x}_{i}}\) from Eq. (28). The lower limit of interval represented by \(\underline{L}\) is given by Eq. (29), and the upper interval limit \(\overline{R}\) is given by Eq. (30), and their respective solutions \(\overline{y}\) upper and \(\underline{y}\) lower are obtained by interpolation, which are given by Eqs. (31) and (32).

Equation (27) is converted into Eq. (33) to get data in the receptive field of the IT2 RBFNN,

$$\underline{L}={x}_{i}{-\sigma }_{{x}_{i}}$$
(29)
$$\overline{R}={x}_{i}+{\sigma }_{{x}_{i}}$$
(30)
$$\underline{y}={y}_{i}-{\sigma }_{{y}_{i}}$$
(31)
$$\overline{y}={y}_{i}{+\sigma }_{{y}_{i}}$$
(32)
$$Rf=\left[\begin{array}{c}\underline{a} \\ \underline{a} \\ \underset{\_}{\begin{array}{c}\underline{A} \\ A\end{array}}\end{array}\begin{array}{c}\overline{a}\\ \overline{a}\\ \overline{\begin{array}{c}A\\ \overline{A}\end{array}}\end{array}\begin{array}{c} \underline{b}\\ \underset{\_}{ B}\\ \underset{\_}{\begin{array}{c} \underline{b}\\ B\end{array}}\end{array}\begin{array}{c} \overline{b}\\ \overline{B}\\ \overline{\begin{array}{c}b\\ \overline{B}\end{array}}\end{array}\begin{array}{c}\underset{\_}{{y}_{1}}\\ \underset{\_}{{y}_{b}}\\ \underset{\_}{\begin{array}{c}\underset{\_}{{y}_{a}}\\ {y}_{ab}\end{array}}\end{array}\begin{array}{c}\overline{{y}_{1}}\\ \overline{{y}_{b}}\\ \overline{\begin{array}{c}{y}_{a}\\ \overline{{y}_{ab}}\end{array}}\end{array}\right]$$
(33)
$$For\; i=1,a,b,ab$$

where: \(Rf\) represents the receptive field universe, \(\underline{a}\) represents the lower limit of the IT2 for the variable a, and this variable (a) is the first receptive field or RF1 in the RBFNN. The same case is for the rest of the elements in the matrix \(Rf\). The interval for the output response is obtained from \(\underset{\_}{{Y}_{i}},{\overline{Y}}_{i}\). Those values represent the lower and upper output or response for a specific receptive field of the inputs.

B. Enhancement and changes to the classic IT2 RBFNN

First, the RBF was changed from logistic to a Gaussian function form, see eq. (3).

Second, the use of the CCD IT2 from [11] is adapted to model the IT2 RBFNN using Eqs. (25)–(31).

C. Improvements

The use of the Huygen’s tractrix (Fig. 3) given by Eq. (34) and the triangular radial basis function (Fig. 4) given by Eq. (35), as RBFs.

$${\mu }_{{X}_{i}=}\frac{1}{1+{\left|\frac{({{x}_{i}-x^{\prime}_{i})}}{c}\right|}^{2}}$$
(34)

where: \(x_{i}\) represents the mean of the variable, \(x^{\prime}_{i}\) represents the input, and c represents the spread of the set.

$${\mu }_{{X}_{i}=\mathrm{MAX}((\mathrm{MIN} \left(\frac{{x}_{i}-A}{B-A}, \frac{C-{x}_{i}}{C-B}\right)), 0)}$$
(35)

where:\(x_{i}\) represents the mean of the variable, A and \(C\) represent the lower and upper limits, respectively, and B represents the mean of the set.

D. Input–output data

For generating the T1 RBFNN and the IT2 RBFNN, 19 data pairs were used, see Table 2.

Table 2 Input–output data for tests

E. Conformation of the T1 RBFNN and IT2 RBFNN architectures

The basic architecture for the T1 RBFNN is assembled by three layers defined in the following way: input layer, hidden layer, and output layer, the conformation and the number of neurons needed on every layer can be seen in Table 3.

Table 3 Description of T1 RBFNN architecture

The architecture for the IT2 RBFNN is assembled by three layers defined in the following way: input layer, hidden layer, and output layer, the conformation and the number of neurons needed on every layer can be seen in Table 4. In this case, a four layer was added for the type reduction and the output.

Table 4 Description of IT2 RBFNN architecture

5 Results

The results of this proposal are organized as follows. First, the test for the logistic function as radial basis or receptive field in IT2 RBFNN without type reduction (Fig. 10) and IT2 RBFNN with type reduction (Fig. 11) is presented. Second, the test for the Gaussian RBF as receptive field in IT2 RBFNN without type reduction (Fig. 12) and IT2 RBFNN with type reduction (Fig. 13) is shown. Third, the test of the Huygen’s tractrix without type reduction (Fig. 14) and with type reduction (Fig. 15) is presented. Fourth, the tests of the triangular RBF without type reduction (Fig. 16) and with type reduction (Fig. 17) are presented.

Fig. 10
figure 10

IT2 RBFNN approximations with logistic RBFs without type reduction

Fig. 11
figure 11

IT2 RBFNN approximations with logistic RBFs, with type reduction

Fig. 12
figure 12

IT2 RBFNN approximations with Gaussian RBFs without type reduction

Fig. 13
figure 13

IT2 RBFNN approximations with Gaussian RBFs, with type reduction

Fig. 14
figure 14

IT2 RBFNN approximations with Huygen’s tractrix RBFs without type reduction

Fig. 15
figure 15

IT2 RBFNN approximations with Huygen’s tractrix RBFs, with type reduction

Fig. 16
figure 16

IT2 RBFNN approximations with triangular RBFs, without type reduction

Fig. 17
figure 17

IT2 RBFNN approximations with triangular RBFs, with type reduction

To calculate the accuracy and the enhancement of the proposal the mean squared error was used to document the variations given by Eq. (36),

$$\mathrm{MSE}=\frac{\sum_{i=1}^{n}{({\widehat{Y}}_{i}-{Y}_{i})}^{2}}{n}$$
(36)

where: \({\widehat{Y}}_{i}\) is the goal or expected value, \({Y}_{i}\) is the obtained value by the model, and n is the total of samples tested. The values obtained in the experiments are shown in Table 5.

Table 5 MSE obtained in the experiments

The use of the logistic RBF, as is shown in Table 5 and Fig. 10, demonstrates that this function produces a big error rate with an MSE value of 5.144 for T1 RBFNN and 9.278 for IT2 RBFNN that increases the error using the IT2 model (see Fig. 11).

The goal values appear inside the interval (Fig. 12) and with the type reduction the values obtained as outputs of IT2 RBFNN are equal or too close to the expected values of prediction (Fig. 13). The classic Gaussian RBF provides good results for the IT2 model (Fig. 13), when it is compared to their counterpart of T1, in comparison to the logistic RBFNN (Fig. 11), in both types, the classic Gaussian RBF reduced de error in a proportion of 5.22 times for T1 and 47 times for IT2.

Huygen’s tractrix used as RBF shows a particular case, the obtained results appear inside the low and high values of the type-2 interval, similar to the results obtained using the Gaussian RBF (Fig. 12), as shown in (Fig. 14). This function never has been used in literature, but their results are excellent due to a reduction near to the 50% when contrasted with the Gaussian RBF in both types (type-1 and type-2 RBFNN), as is shown in (Fig. 15).

The triangular RBF presents an interval of predictions above the expected value (Fig. 16), it can be said that this phenomenon occurs because the IT2 RBFNN system has a non-linear behavior in contrast to the triangular RBF prediction values. Therefore, the error is increased in a linear way. The error rate achieved by the IT2 RBFNN system represents the third part of the obtained by the Logistic RBF and the double of the Gaussian RBF (see Table 5). The predictions are so similar in both models type-1 and type-2 (Fig. 17), nonetheless, T1 RBFNN achieves better results.

6 Conclusion

The use of CCD provides a better method to assemble the fuzzy rule base in a simplified and compact manner with the advantage of a compact base with a few rules that provide precise approximations as outputs.

A simplified and compact rule base reduces the computational time expend in the calculations.

The use of the logistic RBF turned out not to be suitable for a network without training, and this is part of future work. The logistic function does not produce better results by their shape that is non-symmetrical, and these are future research for the application of this class of systems.

The adapted model of Fig. 9 provides an accurate result without training.

The classic Gaussian RBF provides good results for the IT2 model when it is compared to their counterpart of T1, achieving a prediction five times better.

The most important enhancement is the use of a RBFNN which does not need training and provides accurate results.

The results have shown a significant enhancement with the application of the Huygen’s tractrix as RBF, obtaining a 50% in error reduction in comparison to the Gaussian T1 RBFNN and IT2 RBFNN.