1 Introduction

Regression analysis is a widely used approach for characterizing the relationship between response and explanatory variables. Due to the characteristics of practical observations, shortage of information, or decision-makers’ subjective judgment, observations are usually expressed as linguistic terms characterized by membership functions based on fuzzy sets (Zadeh 1965). For fuzzy observations, Tanaka et al. (1982) proposed mathematical programming for formulating a fuzzy regression model using numerical explanatory variables and fuzzy responses. A number of fuzzy approaches have been proposed to establish fuzzy regression models with crisp/fuzzy parameters using various types of explanatory and response variables (Celmins 1987; Chang and Lee 1994; D’Urso and Santoro 2006; Chen and Hsueh 2009; Kelkinnama and Taheri 2012; Chen et al. 2017).

Based on the concept of fuzzy sets, Atanassov (1986) proposed intuitionistic fuzzy sets (IFSs), which include both a membership degree and a non-membership degree to express positive and negative information, respectively. IFSs, which contain more information than do fuzzy sets, have been widely studied and applied in various fields (Atannasov 1999). For solving time series problems, IFSs have been applied to neuron network techniques, such as support vector regression (Lin et al. 2016; Hung and Lin 2013), intuitionistic fuzzy inference (Eyoh et al. 2018; Hájek and Olej 2012), and semi-parametric partially logistic regression (Hesamian and Akbari 2017); however, overfitting may occur and the influence power of explanatory variables cannot be known. A few studies have proposed approaches for formulating intuitionistic fuzzy regression (IFR) models (Parvathi et al. 2013; Arefi and Taheri 2015). Parvathi et al. (2013) applied a linear programming problem to determine the symmetric triangular intuitionistic fuzzy number (TIFN) coefficients of an IFR model. In their study, a key task was to determine the upper and lower bounds of observed crisp data using an IFR model in which the intuitionistic fuzziness is minimized by minimizing the support of the determined coefficients. Based on the concept proposed by Tanaka et al. (1982), the approach presented by Parvathi et al. (2013) produces the crisp parameters of explanatory variables for an objective function of the linear programming problem. Arefi and Taheri (2015) proposed an IFR model based on the least-squares method in which the response and explanatory variables are symmetric TIFNs. In their approach, to simplify computation, a multiplication operation of two symmetric TIFNs is used to approximately produce a symmetric TIFN. This operation is used to obtain a general solution formulation to determine symmetric TIFN parameters. In addition, the formulation was derived based on the premise that explanatory variables and the parameters to be determined are all positive; however, a negative parameter was produced in their example.

The present study proposes an approach for formulating IFR models. Acknowledging the developments of fuzzy regression models, some of their advantages are adopted in this study, since IFSs are the extended version of fuzzy sets. For example, as verified in many studies (Kelkinnama and Taheri 2012) fuzzy regression approaches that adopt the least absolute deviation of the distance between the observed and predicted datasets can produce a more robust estimator than that of those based on least-squares deviation (Stahel and Weisberg 2012). In addition, the signs of the determined parameters can greatly affect the performance of the established fuzzy models based on the fuzzy arithmetic operations. However, some approaches presume that all parameters in the model are positive, although negative parameters can be produced (Chen and Hsueh 2009; Arefi and Taheri 2015), which may affect the interpretation of the explanatory variables and result in poor model performance. Particularly, when least-squares approaches are used for fuzzy regression in the formulation process, the signs of the parameters should be predetermined for deriving the solution equations; however, this is impractical for fuzzy regression analyses with multiple explanatory variables.

With the above considerations, the present study proposes an approach for formulating IFR models. Mathematical programming problems with an objective function for minimizing the absolute deviation of distance are built up based on the definitions of intuitionistic fuzzy numbers (IFNs). The signs of the parameters in the IFR model can be determined for the proposed mathematical programming problem to reflect the corresponding IFN operations. In the following section, some basic definitions of IFSs/IFNs and their properties, such as arithmetic operations and a distance measure, are described. In Sect. 3, a general IFR model is formulated, with the signs of the parameters determined in the formulation process. An example is used to demonstrate the proposed approach and for comparison with an existing approach in Sect. 4. Finally, the conclusions are provided in Sect. 5.

2 Background

This section introduces the basic definitions and properties of IFSs (Atanassov 1986), which are a generalization of those for fuzzy sets. IFSs include membership and non-membership degrees.

Definition 1 (Guha and Chakraborty 2010)

Let X denote a universe of discourse. An IFS \( \tilde{A} \) in X is given by:

$$ \tilde{A} = \left\{ {\left( {x,\mu_{A} (x),v_{A} (x)} \right)|x \in R} \right\} $$
(1)

where \( \mu_{A} (x),v_{A} (x):X \to \left[ {0,1} \right] \) are functions that satisfy \( 0 \le \mu_{A} (x) + v_{A} (x) \le 1 \) for all \( x \in X \). As shown in Fig. 1, the values of \( \mu_{A} (x) \) and \( v_{A} (x) \) represent membership and non-membership degrees, respectively; then, the hesitancy degree can be defined as \( \pi_{A} (x) = 1 - \mu_{A} (x) - v_{A} (x) \).

Fig. 1
figure 1

Intuitionistic fuzzy set

Definition 2 (Guha and Chakraborty 2010)

An IFN is an IFS characterized by:

(1) An IFN is an intuitionistic fuzzy subset defined on the real line.

(2) A unique value \( m \in X \) exists, and \( \mu_{A} (m) = 1 \) and \( v_{A} (m) = 0 \) are met where \( m \) is called the mean value of \( \tilde{A} \).

(3) The convexity of the membership function \( \mu_{A} (x) \) is defined as:

$$ \mu_{A} \left( {\lambda x_{1} + (1 - \lambda )x_{2} } \right) \ge \hbox{min} \left( {\mu_{A} (x_{1} ),\mu_{A} (x_{2} )} \right),\; {\text{where}}\quad x_{1} ,x_{2} \in R\;{\text{and}}\;\lambda \in \left[ {0,1} \right] $$
(2)

(4) The concavity of the non-membership function \( v_{A} (x) \) is defined as:

$$ v_{A} \left( {\lambda x_{1} + (1 - \lambda x_{2} )} \right) \le \hbox{max} \left( {v_{A} (x_{1} ),v_{A} (x_{2} )} \right),\;{\text{where}}\quad x_{1} ,x_{2} \in R\;{\text{and}}\;\lambda \in \left[ {0,1} \right] $$
(3)

Definition 3 (Mahapatra and Roy 2009)

A TIFN \( \tilde{A} = (a^{VL} ,a^{ML} ,a^{C} ,a^{MU} ,a^{VU} ) \) is an IFS in R with the following membership function, \( \mu_{A} (x) \), and non-membership function, \( v_{A} (x) \), respectively:

$$ \mu_{A} (x) = \left\{ {\begin{array}{*{20}l} {\frac{{x - a^{ML} }}{{a^{C} - a^{ML} }}, \, } \hfill & {a^{ML} \le x \le a^{C} } \hfill \\ {\frac{{a^{MU} - x}}{{a^{MU} - a^{C} }},} \hfill & {a^{C} \le x \le a^{MU} } \hfill \\ {0,} \hfill & {\text{otherwise}} \hfill \\ \end{array} } \right.\;{\text{and}}\;v_{A} (x) = \left\{ {\begin{array}{*{20}l} {1 - \frac{{x - a^{VL} }}{{a^{C} - a^{VL} }},} \hfill & {a^{VL} \le x \le a^{C} } \hfill \\ {1 - \frac{{a^{VU} - x}}{{a^{VU} - a^{C} }},} \hfill & {a^{C} \le x \le a^{VU} } \hfill \\ {1,} \hfill & {\text{otherwise}} \hfill \\ \end{array} } \right. $$
(4)

For the TIFN shown in Fig. 2, \( a^{C} \) is called the central value; \( a^{ML} \) and \( a^{MU} \) are the lower and upper bounds of membership, respectively; \( a^{VL} \) and \( a^{VU} \) are the lower and upper bounds of non-membership, respectively.

Fig. 2
figure 2

Triangular intuitionistic fuzzy set

Definition 4 (Chakraborty et al. 2014)

A TIFN \( \tilde{A} = (a^{VL} ,a^{ML} ,a^{C} ,a^{MU} ,a^{VU} ) \) is called positive, i.e., \( \tilde{A} > 0 \), if \( a^{VL} > 0 \); it is called negative, i.e., \( \tilde{A} < 0 \), if \( a^{VU} < 0 \). In addition, \( \tilde{A} \ge 0 \) implies \( a^{VL} \ge 0 \).

Definition 5 (Guha and Chakraborty 2010)

The α-cuts of an IFN \( \tilde{A} \) are defined as:

$$ \tilde{A}_{\alpha } = \left\{ {\left\langle {x,\mu (x),v(x)} \right\rangle |\mu (x) \ge \alpha \;{\text{and}}\;v(x) \le 1 - \alpha ,\;\alpha \in \left[ {0,1} \right]} \right\} $$
(5)

The inequality \( v_{A} (x) \le 1 - \alpha \) is equivalent to \( 1 - v_{A} (x) \ge \alpha \) and thus \( \tilde{A}_{\alpha } \) can be expressed as the crisp sets \( \tilde{A}_{\mu } (\alpha ) \) = \( \left\{ {x:\mu_{A} (x) \ge \alpha } \right\} \) and \( \tilde{A}_{1 - v} (\alpha ) \) = \( \left\{ {x:1 - v_{A} (x) \ge \alpha } \right\} \). Alternatively, \( \tilde{A}_{\alpha } \) can be represented by the following pair of intervals:

$$ \tilde{A}_{\alpha } = \left\{ {[\tilde{A}_{\mu }^{L} (\alpha ),\tilde{A}_{\mu }^{R} (\alpha )],[\tilde{A}_{1 - v}^{L} (\alpha ),\tilde{A}_{1 - v}^{R} (\alpha )]} \right\} $$
(6)

Figure 3 shows that the two crisp sets \( \left\{ {x:v_{A} (x) \le 1 - \alpha } \right\} \) and \( \tilde{A}_{1 - v} (\alpha ) \) = \( \left\{ {x:1 - v_{A} (x) \ge \alpha } \right\} \) have the same intervals. For simplicity and compatibility with a previous study (Arefi and Taheri 2015), the notation of \( \tilde{A}_{1 - v} (\alpha ) \) is adopted hereafter.

Fig. 3
figure 3

α-cuts of a TIFN

Based on this definition, the α-cuts of a TIFN \( \tilde{A} \) can be formulated in the following general form:

$$ \begin{aligned} \tilde{A}_{\alpha } &= \left\{ {\left[ {a^{ML} + \alpha (a^{C} - a^{ML} ),a^{MU} - \alpha (a^{MU} - a^{C} )} \right],}\right.\\&\quad\left.{\quad\left[ {a^{VL} + \alpha (a^{C} - a^{VL} ),a^{VU} - \alpha (a^{VU} - a^{C} )} \right]} \right\} \end{aligned} $$
(7)

The two extreme cases are \( \tilde{A}_{\alpha = 0} = \{ [a^{ML} ,a^{MU} ],[a^{VL} ,a^{VU} ]\} \) and \( \tilde{A}_{\alpha = 1} = \{ [a^{C} ],[a^{C} ]\} \).

Definition 6 (Mahapatra and Roy 2009)

Let \( \tilde{A} = (a^{VL} ,a^{ML} ,a^{C} ,a^{MU} ,a^{VU} ) \) and \( \tilde{B} = (b^{VL} , \)\( b^{ML} , \)\( b^{C} , \)\( b^{MU} , \)\( b^{VU} ) \) be two TIFNs. Based on the extension principle, the sum of two TIFNs can be formulated as:

$$ \tilde{A} \oplus \tilde{B} = (a^{VL} + b^{VL} ,a^{ML} + b^{ML} ,a^{C} + b^{C} ,a^{MU} + b^{MU} ,a^{VU} + b^{VU} ) $$
(8)

The multiplication of a TIFN and a constant k can be expressed as:

$$ k\tilde{A} = (ka^{VL} ,ka^{ML} ,ka^{C} ,ka^{MU} ,ka^{VU} ),\quad {\text{if}}\;k \ge 0 $$
(9)
$$ k\tilde{A} = (ka^{VU} ,ka^{MU} ,ka^{C} ,ka^{ML} ,ka^{VL} ),\quad {\text{if}}\;k < 0 $$
(10)

The multiplication of two TIFNs can be approximately determined using the following equations based on the signs of two TIFNs:

$$ \tilde{A} \otimes \tilde{B} \cong (a^{VL} b^{VL} ,a^{ML} b^{ML} ,a^{C} b^{C} ,a^{MU} b^{MU} ,a^{VU} b^{VU} ),\quad {\text{if}}\;\tilde{A} \ge 0\;{\text{and}}\;\tilde{B} \ge 0 $$
(11)
$$ \tilde{A} \otimes \tilde{B} \cong (a^{VU} b^{VL} ,a^{MU} b^{ML} ,a^{C} b^{C} ,a^{ML} b^{MU} ,a^{VL} b^{VU} ),\quad {\text{if}}\;\tilde{A} \ge 0\;{\text{and}}\;\tilde{B} \le 0 $$
(12)
$$ \tilde{A} \otimes \tilde{B} \cong (a^{VU} b^{VU} ,a^{MU} b^{MU} ,a^{C} b^{C} ,a^{ML} b^{ML} ,a^{VL} b^{VL} ),\quad {\text{if}}\;\tilde{A} \le 0\;{\text{and}}\;\tilde{B} \le 0 $$
(13)

Property 1

Define a zero TIFN as \( \tilde{0} = (0,0,0,0,0) \). Based on the definition of the arithmetic operators of TIFNs, the multiplication of any two TIFNs with different signs, i.e., \( \tilde{A} \)\( \ge 0 \) and \( \tilde{B} \)\( \le 0 \), will result in zero if and only if \( \tilde{A} \) or \( \tilde{B} \) is zero.

Proof

The case of and \( \tilde{B} \) indicates that \(0 \le a^{VL} \le a^{C} \le \)\( a^{MU} \le \)\( a^{VU} \) and \( b^{VL} \)\( \le b^{ML} \le \)\( b^{C} \le \)\( b^{MU} \le \)\( b^{VU} \le 0 \) based on the above definition. In addition, the multiplication of \( \tilde{A} \otimes \tilde{B} \) = \( (a^{VU} b^{VL} , \)\( a^{MU} b^{ML} , \)\( a^{C} b^{C} , \)\( a^{ML} b^{MU} , \)\( a^{VL} b^{VU} ) \)\( = \{ 0\} \) implies that \( a^{VU} b^{VL} = \)\( a^{MU} b^{ML} \)\( = \)\( a^{C} b^{C} = \)\( a^{ML} b^{MU} = \)\( a^{VL} b^{VU} = 0 \). Based on the definition of TIFNs, the inequalities \( a^{VU} b^{VL} \le \)\( a^{MU} b^{ML} \)\( \le \)\( a^{C} b^{C} \le \)\( a^{ML} b^{MU} \le \)\( a^{VL} b^{VU} \) hold. \( \tilde{A} \)\( \ne \{ 0\} \), i.e., \( a^{C} > 0 \), implies that \( b^{C} = 0 \). In addition, the constraint of \( 0 < \)\( a^{C} \le \)\( a^{MU} \le \)\( a^{VU} \) makes \( b^{VL} \)\( = b^{ML} = 0 \) and then \( a^{VU} b^{VL} = \)\( a^{MU} b^{ML} \)\( = 0 \) is satisfied. The constraint of \( b^{C} \le \)\( b^{MU} \le \)\( b^{VU} \le 0 \) implies that \( b^{MU} = \)\( b^{VU} = 0 \) since \( b^{C} = 0 \). Therefore, the TIFN \( \tilde{B} \) is a zero TIFN.

Definition 7 (Grzegorzewski 2003)

The distance between two IFNs, \( \tilde{A} \) and \( \tilde{B} \), can be measured by calculating the integral of the average absolute difference of all α-cuts with a parameter p, where \( 1 \le p \le \infty \). The distance measure, \( D_{p} (\tilde{A},\tilde{B}) \), can be denoted as:

$$ \begin{aligned} D_{p} (\tilde{A},\tilde{B}) &= \Bigg({\frac{1}{4}\int_{0}^{1} {\left| {\tilde{A}_{\mu }^{L} (\alpha ) - \tilde{B}_{\mu }^{L} (\alpha )} \right|^{p} d\alpha } + \frac{1}{4}\int_{0}^{1} {\left| {\tilde{A}_{\mu }^{R} (\alpha ) - \tilde{B}_{\mu }^{R} (\alpha )} \right|^{p} d\alpha }} \\ & \quad \left. {+ \frac{1}{4}\int_{0}^{1} {\left| {\tilde{A}_{1 - v}^{L} (\alpha ) - \tilde{B}_{1 - v}^{L} (\alpha )} \right|^{p} d\alpha } + \frac{1}{4}\int_{0}^{1} {\left| {\tilde{A}_{1 - v}^{R} (\alpha ) - \tilde{B}_{1 - v}^{R} (\alpha )} \right|^{p} d\alpha } } \right.^{{{\raise0.7ex\hbox{$1$} \!\mathord{\left/ {\vphantom {1 p}}\right.\kern-0pt} \!\lower0.7ex\hbox{$p$}}}} \Bigg). \end{aligned} $$
(14)

Based on Eq. (14), the distance measure is the average of the absolute distance difference between the two-side membership (non-membership) functions of the two IFNs. When the IFNs \( \tilde{A} \) and \( \tilde{B} \) are triangular, i.e., TIFNs, and \( p = 1 \), the integral of \( |\tilde{A}_{\mu }^{L} (\alpha ) - \tilde{B}_{\mu }^{L} (\alpha )| \) with \( 0 \le \alpha \le 1 \), i.e., the absolute distance difference between the left-hand-side membership functions of \( \tilde{A} \) and \( \tilde{B} \), will yield either a trapezoidal area (Fig. 4a) or two triangular areas (Fig. 4b), where the top-side length is \( |\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)| \), the bottom-side length is \( |\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)| \), and the height is 1. Based on basic geometry, if the signs of (\( \tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1) \)) and (\( \tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0) \)) are the same, i.e., \( \left( {\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)} \right) \times \left( {\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)} \right) \ge 0 \), a trapezoidal area will be produced; otherwise, two triangular areas are obtained. The area of the former can be expressed as \( \tfrac{1}{2}(|\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)| + |\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)|) \), and that of the latter is \( \tfrac{1}{4}(|\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)| \) + \( |\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)|) \). Let \( D^{ML} (\tilde{A},\tilde{B}) \) denote the integral of \( |\tilde{A}_{\mu }^{L} (\alpha ) - \tilde{B}_{\mu }^{L} (\alpha )| \) with 0 ≤ α ≤ 1; it can be formulated as:

Fig. 4
figure 4

Integral of the difference of the left spread of membership between two TIFNs

$$ \begin{aligned} D^{ML} (\tilde{A},\tilde{B}) & = \int_{0}^{1} {\left| {\tilde{A}_{\mu }^{L} (\alpha ) - \tilde{B}_{\mu }^{L} (\alpha )} \right|d\alpha } \\ \, & = \left\{ \begin{aligned} \frac{1}{2}\left( {\left| {\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)} \right| + \left| {\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)} \right|} \right) ,\hfill\\ \quad {\text{if}}\;\left( {\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)} \right)\left( {\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)} \right) \ge 0 \hfill \\ \frac{1}{4}\left( {\left| {\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)} \right| + \left| {\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)} \right|} \right) ,\\\quad {\text{if}}\;\left( {\tilde{A}_{\mu }^{L} (1) - \tilde{B}_{\mu }^{L} (1)} \right)\left( {\tilde{A}_{\mu }^{L} (0) - \tilde{B}_{\mu }^{L} (0)} \right) < 0 \hfill \\ \end{aligned} \right. \\ \, & = \left\{ \begin{aligned} \frac{1}{2}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{ML} - b^{ML} } \right|} \right) ,\quad {\text{if}}\; (a^{C} - b^{C} ) (a^{ML} - b^{ML} )\ge 0 \hfill \\ \frac{1}{4}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{ML} - b^{ML} } \right|} \right) ,\quad {\text{if}}\; (a^{C} - b^{C} ) (a^{ML} - b^{ML} )< 0 \hfill \\ \end{aligned} \right. \\ \end{aligned} $$
(15)

Similarly, the other three components of the distance measure can be determined as:

$$ D^{VL} (\tilde{A},\tilde{B}) = \left\{ \begin{aligned} \frac{1}{2}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{VL} - b^{VL} } \right|} \right) ,\quad {\text{if}}\; (a^{C} - b^{C} ) (a^{VL} - b^{VL} )\ge 0 \hfill \\ \frac{1}{4}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{VL} - b^{VL} } \right|} \right) ,\quad {\text{if}}\; (a^{C} - b^{C} ) (a^{VL} - b^{VL} )< 0 \hfill \\ \end{aligned} \right. $$
(16)
$$ D^{MU} (\tilde{A},\tilde{B}) = \left\{ \begin{aligned} \frac{1}{2}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{MU} - b^{MU} } \right|} \right) ,\quad {\text{if}}\; (a^{C} - b^{C} ) (a^{MU} - b^{MU} )\ge 0 \hfill \\ \frac{1}{4}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{MU} - b^{MU} } \right|} \right) ,\quad {\text{if}}\; (a^{C} - b^{C} ) (a^{MU} - b^{MU} )< 0 \hfill \\ \end{aligned} \right. $$
(17)
$$ D^{VU} (\tilde{A},\tilde{B}) = \left\{ \begin{aligned} \frac{1}{2}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{VU} - b^{VU} } \right|} \right) ,\quad {\text{if}}\;{ (}a^{C} - b^{C} ) (a^{VU} - b^{VU} )\ge 0 \hfill \\ \frac{1}{4}\left( {\left| {a^{C} - b^{C} } \right| + \left| {a^{VU} - b^{VU} } \right|} \right) ,\quad {\text{if}}\; (a^{C} - b^{C} ) (a^{VU} - b^{VU} )< 0 \hfill \\ \end{aligned} \right. $$
(18)

Therefore, the distance measure of two TIFNs can be reformulated as the average of the above four kinds of distance measure as follows:

$$ D_{TIFN} (\tilde{A},\tilde{B}) = \frac{1}{4}\left( {D^{VL} (\tilde{A},\tilde{B}) + D^{ML} (\tilde{A},\tilde{B}) + D^{MU} (\tilde{A},\tilde{B}) + D^{VU} (\tilde{A},\tilde{B})} \right) $$
(19)

The above formulation can be considered as a general distance measure for measuring the distance between two TIFNs.

Definition 8 (Arefi and Taheri 2015)

A similarity measure of two TIFNs \( \tilde{A} \) and \( \tilde{B} \) is defined as:

$$ \begin{aligned} S(\tilde{A},\tilde{B}) &= 1 - \frac{1}{2}\left[ {\frac{{\int_{ - \infty }^{\infty } {\left| {\mu_{A} (x) - \mu_{B} (x)} \right|dx} }}{{\int_{ - \infty }^{\infty } {\mu_{A} (x)dx} + \int_{ - \infty }^{\infty } {\mu_{B} (x)dx} }}}\right. \\&\left.{\quad+ \frac{{\int_{ - \infty }^{\infty } {\left| {v_{A} (x) - v_{B} (x)} \right|dx} }}{{\int_{ - \infty }^{\infty } {(1 - v_{A} (x))dx} + \int_{ - \infty }^{\infty } {(1 - v_{B} (x))dx} }}} \right] \end{aligned} $$
(20)

with the value of \( 0 \le S(\tilde{A},\tilde{B}) \le 1 \). This index is measured in terms of the area of the average difference between the membership and non-membership functions of two IFNs. However, the degree of difference cannot be determined if the two IFNs have no interactions.

Definition 9

The general formula of a different distance measure based on squared errors between two TIFNs, proposed by Arefi and Taheri (2015), is:

$$ \begin{aligned} d^{2} (\tilde{A},\tilde{B}) & = \left( {a^{C} - b^{C} } \right)^{2} + \frac{1}{24}\left[ {\left( {s_{B}^{ML} - s_{A}^{ML} } \right)^{2} + \left( {s_{B}^{MR} - s_{A}^{MR} } \right)^{2} + \left( {s_{B}^{VL} - s_{A}^{VL} } \right)^{2}}\right. \\&\left.{\quad+ \left( {s_{B}^{VR} - s_{A}^{VR} } \right)^{2} } \right] + \frac{1}{6}\left( {m_{A} - m_{B} } \right)\left[ {\left( {s_{B}^{ML} - s_{A}^{ML} } \right) - \left( {s_{B}^{MR} - s_{A}^{MR} } \right) }\right.\\&\left.{\quad+ \left( {s_{B}^{VL} - s_{A}^{VL} } \right) - \left( {s_{B}^{VR} - s_{A}^{VR} } \right)} \right] \\ \end{aligned} $$
(21)

where \( s_{A}^{ML} = a^{C} - a^{ML} \) and \( s_{A}^{MR} = a^{MU} - a^{C} \) are called the left and right spreads of the membership function, respectively; similarly, \( s_{A}^{VL} = a^{C} - a^{VL} \) and \( s_{A}^{VR} = a^{VU} - a^{C} \) are the left and right spreads of the non-membership function, respectively.

Useful measures are critical for evaluating IFR model performance. Arefi and Taheri (2015) adopted the similarity measure \( S(\tilde{A},\tilde{B}) \) and squared error distance \( d^{2} (\tilde{A},\tilde{B}) \) to evaluate the performance of their proposed IFR approach with TIFN explanatory and response datasets. Besides these two measures, the distance measure \( D_{TIFN} (\tilde{A},\tilde{B}) \) in terms of the average absolute difference of two TIFNs is used in this study to compare the performance of the proposed approach with that proposed by Arefi and Taheri (2015).

3 Formulations

This study builds up an IFR model based on the criterion of the least absolute difference of distance. With the least absolute deviations criterion, mathematical programming problems are formulated to determine the optimal parameters of TIFNs to minimize the total distance between the observation and prediction variables. To achieve this end, the general distance measure, \( D_{TIFN} (\tilde{A},\tilde{B}) \), expressed in Eq. (19), is used as the objective function in the mathematical programming problems. In addition, for comparison with an existing approach (Arefi and Taheri 2015), the observed, predicted, and explanatory variables and parameters are expressed as TIFNs in Definition 3.

Consider the intuitionistic fuzzy observation set \( (\tilde{Y}_{i} ,\tilde{X}_{1i} ,\tilde{X}_{2i} , \ldots ,\tilde{X}_{ji} , \ldots ,\tilde{X}_{pi} ) \), where \( \tilde{Y}_{i} \) = \( (y_{i}^{VL} , \)\( y_{i}^{ML} , \)\( y_{i}^{C} , \)\( y_{i}^{MU} , \)\( y_{i}^{VU} ) \) is the response variable and \( \tilde{X}_{ji} = \)\( (x_{ji}^{VL} , \)\( x_{ji}^{ML} , \)\( x_{ji}^{C} , \)\( x_{ji}^{MU} , \)\( x_{ji}^{VU} ) \) represents the jth explanatory variable in the form of TIFNs. The general IFR model can be expressed as:

$$ \tilde{Y}_{i} = \tilde{B}_{0} \oplus \tilde{B}_{1} \otimes \tilde{X}_{1i} \oplus \tilde{B}_{2} \otimes \tilde{X}_{2i} \oplus \cdots \oplus \tilde{B}_{p} \otimes \tilde{X}_{pi} = \sum\limits_{j = 0}^{p} {\tilde{B}_{j} \otimes \tilde{X}_{ji} } $$
(22)

where \( \tilde{B}_{j} = (b_{j}^{VL} ,b_{j}^{ML} ,b_{j}^{C} ;b_{j}^{MU} ,b_{j}^{VU} ) \) is the corresponding intuitionistic fuzzy parameters, and \( \tilde{X}_{0i} = \left( {1,1,1,1,1} \right) \) is specified. Let the predicted fuzzy response variable be denoted as \( \hat{\tilde{Y}}_{i} \) = \( (\hat{y}_{i}^{VL} , \)\( \hat{y}_{i}^{ML} , \)\( \hat{y}_{i}^{C} , \)\( \hat{y}_{i}^{MU} , \)\( \hat{y}_{i}^{VU} ) \); then, the predicted IFR model can be formulated as:

$$ \hat{\tilde{Y}}_{i} = \hat{\tilde{B}}_{0} \oplus (\hat{\tilde{B}}_{1} \otimes \tilde{X}_{1i} ) \oplus (\hat{\tilde{B}}_{2} \otimes \tilde{X}_{2i} ) \oplus \cdots \oplus (\hat{\tilde{B}}_{p} \otimes \tilde{X}_{pi} ) = \sum\limits_{j = 0}^{p} {\hat{\tilde{B}}_{j} \otimes \tilde{X}_{ji} } $$
(23)

where \( \hat{\tilde{B}}_{j} = (\hat{b}_{j}^{VL} ,\hat{b}_{j}^{ML} ,\hat{b}_{j}^{C} ;\hat{b}_{j}^{MU} ,\hat{b}_{j}^{VU} ) \) is the jth estimated intuitionistic fuzzy parameter of the TIFNs. Consider a model with one explanatory variable, i.e., \( \hat{\tilde{Y}}_{i} = \hat{\tilde{B}}_{0} \oplus (\hat{\tilde{B}}_{1} \otimes \tilde{X}_{i} ) \). Suppose that the TIFN parameter \( \hat{\tilde{B}}_{1} \) is negative, i.e., \( \hat{\tilde{B}}_{1} \le 0 \); then, based on the arithmetic operator in Definition 6, the predicted response TIFN \( \hat{\tilde{Y}}_{i} \) is determined as:

$$ \left\{ {\begin{array}{*{20}l} {\hat{y}_{i}^{VL} = \hat{b}_{0}^{VL} + \hat{b}_{1}^{VL} x_{i}^{VU} } \hfill \\ {\hat{y}_{i}^{ML} = \hat{b}_{0}^{ML} + \hat{b}_{1}^{ML} x_{i}^{MU} } \hfill \\ {\hat{y}_{i}^{C} = \hat{b}_{0}^{C} + \hat{b}_{1}^{C} x_{i}^{C} } \hfill \\ {\hat{y}_{i}^{MU} = \hat{b}_{0}^{MU} + \hat{b}_{1}^{MU} x_{i}^{ML} } \hfill \\ {\hat{y}_{i}^{VU} = \hat{b}_{0}^{VU} + \hat{b}_{1}^{VU} x_{i}^{VL} } \hfill \\ \end{array} } \right. $$
(24)

Alternatively, suppose that this TIFN parameter is positive and denoted as \( \hat{\tilde{B}}_{2} \), i.e., \( \hat{\tilde{B}}_{2} \ge 0 \); then, the TIFN \( \hat{\tilde{Y}}_{i} \) can be expressed as:

$$ \left\{ {\begin{array}{*{20}l} {\hat{y}_{i}^{VL} = \hat{b}_{0}^{VL} + \hat{b}_{2}^{VL} x_{i}^{VL} } \hfill \\ {\hat{y}_{i}^{ML} = \hat{b}_{0}^{ML} + \hat{b}_{2}^{ML} x_{i}^{ML} } \hfill \\ {\hat{y}_{i}^{C} = \hat{b}_{0}^{C} + \hat{b}_{2}^{C} x_{i}^{C} } \hfill \\ {\hat{y}_{i}^{MU} = \hat{b}_{0}^{MU} + \hat{b}_{2}^{MU} x_{i}^{MU} } \hfill \\ {\hat{y}_{i}^{VU} = \hat{b}_{0}^{VU} + \hat{b}_{2}^{VU} x_{i}^{VU} } \hfill \\ \end{array} } \right. $$
(25)

The signs of the explanatory TIFN parameters are unknown, which influences the IFR model performance. To overcome this problem, this study sets two dummy TIFNs with different signs, i.e., \( \hat{\tilde{B}}_{1} \le 0 \) and \( \hat{\tilde{B}}_{2} \ge 0 \). Based on Property 1, if an IFR model has the formulation \( \hat{\tilde{Y}}_{i} = \hat{\tilde{B}}_{0} \oplus ((\hat{\tilde{B}}_{1} \oplus \hat{\tilde{B}}_{2} ) \otimes \tilde{X}_{i} ) \) subject to \( \hat{\tilde{B}}_{1} \otimes \hat{\tilde{B}}_{2} = 0 \), then one dummy TIFN parameter will be determined as the optimal parameter for the explanatory variable and the other one will be zero. Therefore, the predicted TIFN response will have two dummy TIFN parameters for each explanatory TIFN variable in the mathematical programming problems, in which the constraint of their product being zero is added. For the predicted TIFN response with one explanatory variable, the formulations in the mathematical programming problems are:

$$ \left\{ {\begin{array}{*{20}l} {\hat{y}_{i}^{VL} = \hat{b}_{0}^{VL} + \hat{b}_{1}^{VL} x_{i}^{VU} + \hat{b}_{2}^{VL} x_{i}^{VL} } \hfill \\ {\hat{y}_{i}^{ML} = \hat{b}_{0}^{ML} + \hat{b}_{1}^{ML} x_{i}^{MU} + \hat{b}_{2}^{ML} x_{i}^{ML} } \hfill \\ {\hat{y}_{i}^{C} = \hat{b}_{0}^{C} + \hat{b}_{1}^{C} x_{i}^{C} + \hat{b}_{2}^{C} x_{i}^{C} } \hfill \\ {\hat{y}_{i}^{MU} = \hat{b}_{0}^{MU} + \hat{b}_{1}^{MU} x_{i}^{ML} + \hat{b}_{2}^{MU} x_{i}^{MU} } \hfill \\ {\hat{y}_{i}^{VU} = \hat{b}_{0}^{VU} + \hat{b}_{1}^{VU} x_{i}^{VL} + \hat{b}_{2}^{VU} x_{i}^{VU} } \hfill \\ \end{array} } \right. $$
(26)

With multiple TIFN explanatory variables, the mathematical programming problems are formulated as Eq. (27), in which the objective function \( D_{TIFN} (\tilde{A},\tilde{B}) \) in Eq. (19) is adopted to determine the optimal parameters in order to minimize the distance between observed and predicted TIFN responses.

$$ \begin{aligned} & \hbox{min} \sum\limits_{i = 1}^{n} {D_{TIFN} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} )} = \frac{1}{4}\sum\limits_{i = 1}^{n} {\left( {D^{VL} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) + D^{ML} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) + D^{MU} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) + D^{VU} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} )} \right)} \\ & {\text{s}} . {\text{t}} .\quad \hat{y}_{i}^{VL} = \hat{b}_{0}^{VL} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VL} x_{ji}^{VU} + \hat{b}_{j2}^{VL} x_{ji}^{VL} } \right]} \\ \, & \quad \quad \hat{y}_{i}^{ML} = \hat{b}_{0}^{ML} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{ML} x_{ji}^{MU} + \hat{b}_{j2}^{ML} x_{ji}^{ML} } \right]} \\ & \quad \quad \hat{y}_{i}^{C} \, = \hat{b}_{0}^{C} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{C} + \hat{b}_{j2}^{C} } \right]x_{ji}^{C} } \\ \, & \quad \quad \hat{y}_{i}^{MU} = \hat{b}_{0}^{MU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{MU} x_{ji}^{ML} + \hat{b}_{j2}^{MU} x_{ji}^{MU} } \right]} \\ \, & \quad \quad \hat{y}_{i}^{VU} = \hat{b}_{0}^{VU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VU} x_{ji}^{VL} + \hat{b}_{j2}^{VU} x_{ji}^{VU} } \right]} \, \\ \, & \quad \quad \hat{b}_{j1}^{C} \le 0; \, \hat{b}_{j2}^{C} \ge 0 \\ \, & \quad \quad \hat{b}_{jk}^{VL} \le \hat{b}_{jk}^{ML} \le \hat{b}_{jk}^{C} \le \hat{b}_{jk}^{MU} \le \hat{b}_{jk}^{VU} \, \\ \, & \quad \quad \hat{b}_{j1}^{VL} \hat{b}_{j2}^{VU} = \hat{b}_{j1}^{ML} \hat{b}_{j2}^{MU} = \hat{b}_{j1}^{C} \hat{b}_{j2}^{C} = \hat{b}_{j1}^{MU} \hat{b}_{j2}^{ML} = \hat{b}_{j1}^{VU} \hat{b}_{j2}^{VL} = 0 \, \\ & \quad \quad i = 1, \cdots ,n, \, j = 1, \ldots ,p, \, k = 1,2 \\ \end{aligned} $$
(27)

The last three constraints in the above model restrict the two dummy parameters, \( \hat{\tilde{B}}_{j1} \), and \( \hat{\tilde{B}}_{j2} \), for each explanatory variable with different signs. The zero restriction, \( \hat{\tilde{B}}_{j1} \otimes \hat{\tilde{B}}_{j2} = 0 \), holds for all dummy parameters. The resulting parameters should follow the definition of a TIFN.

In addition, the objective function in Eq. (27) is the average of \( D^{VL} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \), \( D^{ML} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \), \( D^{MU} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \), and \( D^{VU} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \), in which correct formulations should be decided for each measure in the resolution process based on Eqs. (15)–(18). To deal with this problem, a pair of dummy binary variables, \( d_{1i}^{VL} \) and \( d_{2i}^{VL} \), is added in \( D^{VL} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \); they can be reformulated as:

$$ \begin{aligned} D^{VL} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) & = d_{1i}^{VL} \frac{1}{2}\left( {\left| {y_{i}^{C} - \hat{y}_{i}^{C} } \right| + \left| {y_{i}^{VL} - \hat{y}_{i}^{VL} } \right|} \right) + d_{2i}^{VL} \frac{1}{4}\left( {\left| {y_{i}^{C} - \hat{y}_{i}^{C} } \right| + \left| {y_{i}^{VL} - \hat{y}_{i}^{VL} } \right|} \right) \\ \, & = \left( {\frac{1}{2}d_{1i}^{VL} + \frac{1}{4}d_{2i}^{VL} } \right)\left( {\left| {y_{i}^{C} - \hat{y}_{i}^{C} } \right| + \left| {y_{i}^{VL} - \hat{y}_{i}^{VL} } \right|} \right) \\ \end{aligned} $$
(28)

Additional constraints of Eq. (29) are also added in the mathematical programming problems.

$$ \begin{aligned} & (d_{1i}^{VL} - d_{2i}^{VL} \, )(y_{i}^{C} - \hat{y}_{i}^{C} )(y_{i}^{VL} - \hat{y}_{i}^{VL} ) \ge 0 \\ & d_{1i}^{VL} + d_{2i}^{VL} = 1 , { }d_{1i}^{VL} ,d_{2i}^{VL} \in \{ 0,1\} \\ \end{aligned} $$
(29)

The constraints guarantee that when \( (y_{i}^{C} - \hat{y}_{i}^{C} )(y_{i}^{VL} - \hat{y}_{i}^{VL} ) \ge 0 \), then \( d_{1i}^{VL} = 1 \) and \( d_{2i}^{VL} = 0 \); otherwise, \( d_{1i}^{VL} = 0 \) and \( d_{2i}^{VL} = 1 \), and \( D^{VL} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) in Eq. (28) is obtained. Similarly, \( D^{ML} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \), \( D^{MU} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \), and \( D^{VU} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) can be determined. Therefore, Eq. (27) becomes:

$$ \begin{aligned} & \hbox{min} \sum\limits_{i = 1}^{n} {D_{TIFN} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} )} = \frac{1}{4}\sum\limits_{i = 1}^{n} {\left\{ {\left( {\frac{{d_{1i}^{VL} }}{2} + \frac{{d_{2i}^{VL} }}{4}} \right)\left( {\left| {y_{i}^{C} - \hat{y}_{i}^{C} } \right| + \left| {y_{i}^{VL} - \hat{y}_{i}^{VL} } \right|} \right) } \right.} \\ & \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \left. {+ \left( {\frac{{d_{1i}^{ML} }}{2} + \frac{{d_{2i}^{ML} }}{4}} \right)\left( {\left| {y_{i}^{C} - \hat{y}_{i}^{C} } \right| + \left| {y_{i}^{ML} - \hat{y}_{i}^{ML} } \right|} \right)} \right. \\ & \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \left. { + \left( {\frac{{d_{1i}^{MU} }}{2} + \frac{{d_{2i}^{MU} }}{4}} \right)\left( {\left| {y_{i}^{C} - \hat{y}_{i}^{C} } \right| + \left| {y_{i}^{MU} - \hat{y}_{i}^{MU} } \right|} \right)} \right. \\ & \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \quad \left. { + \left( {\frac{{d_{1i}^{VU} }}{2} + \frac{{d_{2i}^{VU} }}{4}} \right)\left( {\left| {y_{i}^{C} - \hat{y}_{i}^{C} } \right| + \left| {y_{i}^{VU} - \hat{y}_{i}^{VU} } \right|} \right)} \right\} \\ & {\text{s}} . {\text{t}} .\quad \hat{y}_{i}^{VL} = \hat{b}_{0}^{VL} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VL} x_{ji}^{VU} + \hat{b}_{j2}^{VL} x_{ji}^{VL} } \right]} \\ & \quad \quad \hat{y}_{i}^{ML} = \hat{b}_{0}^{ML} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{ML} x_{ji}^{MU} + \hat{b}_{j2}^{ML} x_{ji}^{ML} } \right]} \\ & \quad \quad \hat{y}_{i}^{C} \, = \hat{b}_{0}^{C} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{C} + \hat{b}_{j2}^{C} } \right]x_{ji}^{C} } \\ & \quad \quad \hat{y}_{i}^{MU} = \hat{b}_{0}^{MU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{MU} x_{ji}^{ML} + \hat{b}_{j2}^{MU} x_{ji}^{MU} } \right]} \\ & \quad \quad \hat{y}_{i}^{VU} = \hat{b}_{0}^{VU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VU} x_{ji}^{VL} + \hat{b}_{j2}^{VU} x_{ji}^{VU} } \right]} \\ & \quad \quad (d_{1i}^{VL} - d_{2i}^{VL} \, )(y_{i}^{C} - \hat{y}_{i}^{C} )(y_{i}^{VL} - \hat{y}_{i}^{VL} ) \ge 0 \\ & \quad \quad (d_{1i}^{ML} - d_{2i}^{ML} \, )(y_{i}^{C} - \hat{y}_{i}^{C} )(y_{i}^{ML} - \hat{y}_{i}^{ML} ) \ge 0 \\ & \quad \quad (d_{1i}^{MU} - d_{2i}^{MU} \, )(y_{i}^{C} - \hat{y}_{i}^{C} )(y_{i}^{MU} - \hat{y}_{i}^{MU} ) \ge 0 \\ & \quad \quad (d_{1i}^{VU} - d_{2i}^{VU} \, )(y_{i}^{C} - \hat{y}_{i}^{C} )(y_{i}^{VU} - \hat{y}_{i}^{VU} ) \ge 0 \\ & \quad \quad d_{1i}^{VL} + d_{2i}^{VL} = 1 , { }d_{1i}^{ML} + d_{2i}^{ML} = 1 , { }d_{1i}^{MU} + d_{2i}^{MU} = 1 , { }d_{1i}^{VU} + d_{2i}^{VU} = 1 \, \\ & \quad \quad d_{1i}^{VL} ,d_{2i}^{VL} ,d_{1i}^{ML} ,d_{2i}^{ML} ,d_{1i}^{MU} ,d_{2i}^{MU} ,d_{1i}^{VU} ,d_{2i}^{VU} \in \{ 0,1\} \\ & \quad \quad \hat{b}_{j1}^{C} \le 0; \, \hat{b}_{j2}^{C} \ge 0 \\ & \quad \quad \hat{b}_{jk}^{VL} \le \hat{b}_{jk}^{ML} \le \hat{b}_{jk}^{C} \le \hat{b}_{jk}^{MU} \le \hat{b}_{jk}^{VU} \\ & \quad \quad \hat{b}_{j1}^{VL} \hat{b}_{j2}^{VU} = \hat{b}_{j1}^{ML} \hat{b}_{j2}^{MU} = \hat{b}_{j1}^{C} \hat{b}_{j2}^{C} = \hat{b}_{j1}^{MU} \hat{b}_{j2}^{ML} = \hat{b}_{j1}^{VU} \hat{b}_{j2}^{VL} = 0 \\ & \quad \quad i = 1, \ldots ,n, \, j = 1, \ldots ,p, \, k = 1,2 \\ \end{aligned} $$
(30)

Furthermore, considering that the objective function in Eq. (30) is expressed as the sum of the absolute difference between observed and predicted TIFN responses, this will increase computational efforts. To deal with this problem, an efficient approach can be applied to enhance the computational efficiency. For example, let \( M_{i}^{1} \) denote \( \hbox{max} \{ y_{i}^{C} - \hat{y}_{i}^{C} ,0\} \) and \( M_{i}^{2} \) be \( \hbox{max} \{ \hat{y}_{i}^{C} - y_{i}^{C} ,0\} \). It is easy to show that \( M_{i}^{1} + M_{i}^{2} \) is equivalent to \( |y_{i}^{C} - \hat{y}_{i}^{C} | \) and that \( M_{i}^{1} - M_{i}^{2} \) is equivalent to \( y_{i}^{C} - \hat{y}_{i}^{C} \). Thus, the model can be reformulated as:

$$ \begin{aligned} & \hbox{min} \sum\limits_{i = 1}^{n} {D_{TIFN} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} )} = \frac{1}{4}\sum\limits_{i = 1}^{n} \left\{ {\left( {\frac{{d_{1i}^{VL} }}{2} + \frac{{d_{2i}^{VL} }}{4}} \right)\left( {M_{i}^{1} + M_{i}^{2} + VL_{i}^{1} + VL_{i}^{2} } \right)} \right. \hfill \\ &\left. {\quad + \left( {\frac{{d_{1i}^{ML} }}{2} + \frac{{d_{2i}^{ML} }}{4}} \right)\left( {M_{i}^{1} + M_{i}^{2} + ML_{i}^{1} + ML_{i}^{2} } \right)} \right. \hfill \\ &\left. {\quad + \left( {\frac{{d_{1i}^{MU} }}{2} + \frac{{d_{2i}^{MU} }}{4}} \right)\left( {M_{i}^{1} + M_{i}^{2} + MU_{i}^{1} + MU_{i}^{2} } \right) } \right. \hfill \\ &\left. {\quad+ \left( {\frac{{d_{1i}^{VU} }}{2} + \frac{{d_{2i}^{VU} }}{4}} \right)\left( {M_{i}^{1} + M_{i}^{2} + VU_{i}^{1} + VU_{i}^{2} } \right)} \right\} \hfill \\ & {\text{s}} . {\text{t}} .\quad M_{i}^{1} - M_{i}^{2} = y_{i}^{C} - \hat{y}_{i}^{C} \\ & \quad \quad VL_{i}^{1} - VL_{i}^{2} = y_{i}^{VL} - \hat{y}_{i}^{VL} \\ & \quad \quad ML_{i}^{1} - ML_{i}^{2} = y_{i}^{ML} - \hat{y}_{i}^{ML} \\ & \quad \quad MU_{i}^{1} - MU_{i}^{2} = y_{i}^{MU} - \hat{y}_{i}^{MU} \\ & \quad \quad VU_{i}^{1} - VU_{i}^{2} = y_{i}^{VU} - \hat{y}_{i}^{VU} \\ & \quad \quad \hat{y}_{i}^{VL} = \hat{b}_{0}^{VL} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VL} x_{ji}^{VU} + \hat{b}_{j2}^{VL} x_{ji}^{VL} } \right]} \\ & \quad \quad \hat{y}_{i}^{ML} = \hat{b}_{0}^{ML} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{ML} x_{ji}^{MU} + \hat{b}_{j2}^{ML} x_{ji}^{ML} } \right]} \\ & \quad \quad \hat{y}_{i}^{C} \, = \hat{b}_{0}^{C} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{C} + \hat{b}_{j2}^{C} } \right]x_{ji}^{C} } \\ & \quad \quad \hat{y}_{i}^{MU} = \hat{b}_{0}^{MU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{MU} x_{ji}^{ML} + \hat{b}_{j2}^{MU} x_{ji}^{MU} } \right]} \\ & \quad \quad \hat{y}_{i}^{VU} = \hat{b}_{0}^{VU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VU} x_{ji}^{VL} + \hat{b}_{j2}^{VU} x_{ji}^{VU} } \right]} \\ & \quad \quad (d_{1i}^{VL} - d_{2i}^{VL} \, )(M_{i}^{1} - M_{i}^{2} )(VL_{i}^{1} - VL_{i}^{2} ) \ge 0 \\ & \quad \quad (d_{1i}^{ML} - d_{2i}^{ML} \, )(M_{i}^{1} - M_{i}^{2} )(ML_{i}^{1} - ML_{i}^{2} ) \ge 0 \\ & \quad \quad (d_{1i}^{MU} - d_{2i}^{MU} \, )(M_{i}^{1} - M_{i}^{2} )(MU_{i}^{1} - MU_{i}^{2} ) \ge 0 \\ & \quad \quad (d_{1i}^{VU} - d_{2i}^{VU} \, )(M_{i}^{1} - M_{i}^{2} )(VU_{i}^{1} - VU_{i}^{2} ) \ge 0 \\ & \quad \quad d_{1i}^{VL} + d_{2i}^{VL} = 1 , { }d_{1i}^{ML} + d_{2i}^{ML} = 1 , { }d_{1i}^{MU} + d_{2i}^{MU} = 1 , { }d_{1i}^{VU} + d_{2i}^{VU} = 1 \, \\ & \quad \quad d_{1i}^{VL} ,d_{2i}^{VL} ,d_{1i}^{ML} ,d_{2i}^{ML} ,d_{1i}^{MU} ,d_{2i}^{MU} ,d_{1i}^{VU} ,d_{2i}^{VU} \in \{ 0,1\} \\ & \quad \quad \hat{b}_{j1}^{C} \le 0; \, \hat{b}_{j2}^{C} \ge 0 \\ & \quad \quad \hat{b}_{jk}^{VL} \le \hat{b}_{jk}^{ML} \le \hat{b}_{jk}^{C} \le \hat{b}_{jk}^{MU} \le \hat{b}_{jk}^{VU} \\ & \quad \quad \hat{b}_{j1}^{VL} \hat{b}_{j2}^{VU} = \hat{b}_{j1}^{ML} \hat{b}_{j2}^{MU} = \hat{b}_{j1}^{C} \hat{b}_{j2}^{C} = \hat{b}_{j1}^{MU} \hat{b}_{j2}^{ML} = \hat{b}_{j1}^{VU} \hat{b}_{j2}^{VL} = 0 \\ & \quad \quad M_{i}^{1} ,M_{i}^{2} ,VL_{i}^{1} ,VL_{i}^{2} ,ML_{i}^{1} ,ML_{i}^{2} ,MU_{i}^{1} ,MU_{i}^{2} ,VU_{i}^{1} ,VU_{i}^{2} \ge 0 \\ & \quad \quad \, i = 1, \ldots ,n, \, j = 1, \ldots ,p, \, k = 1,2 \\ \end{aligned} $$
(31)

In addition, sometimes an observation dataset contains two types of explanatory variable, i.e., crisp and TIFN explanatory variables. For example, suppose that p explanatory variables are adopted to build up an IFR model, among which \( \tilde{X}_{1} ,\tilde{X}_{2} , \ldots ,\tilde{X}_{k} \) are TIFNs and \( X_{k + 1} ,X_{k + 2} , \ldots ,X_{p} \) are crisp numbers, i.e., \( X_{ji}^{VL} \), \( X_{ji}^{ML} \), \( X_{ji}^{MU} \), and \( X_{ji}^{VU} \) are equal to \( X_{ji}^{C} \) for j = k+1 to p. The formulations of the predicted TIFN responses in Eq. (30) become:

$$ \begin{aligned} \hat{y}_{i}^{VL} & = \hat{b}_{0}^{VL} + \sum\limits_{j = 1}^{k} {\left[ {\hat{b}_{j1}^{VL} x_{ji}^{VU} + \hat{b}_{j2}^{VL} x_{ji}^{VL} } \right]} + \sum\limits_{j = k + 1}^{p} {\left[ {\hat{b}_{j1}^{VL} + \hat{b}_{j2}^{VL} } \right]x_{ji}^{C} } \\ \hat{y}_{i}^{ML} & = \hat{b}_{0}^{ML} + \sum\limits_{j = 1}^{k} {\left[ {\hat{b}_{j1}^{ML} x_{ji}^{MU} + \hat{b}_{j2}^{ML} x_{ji}^{ML} } \right] + \sum\limits_{j = k + 1}^{p} {\left[ {\hat{b}_{j1}^{ML} + \hat{b}_{j2}^{ML} } \right]x_{ji}^{C} } } \\ \hat{y}_{i}^{C} \, & = \hat{b}_{0}^{C} + \sum\limits_{j = 1}^{k} {\left[ {\hat{b}_{j1}^{C} + \hat{b}_{j2}^{C} } \right]x_{ji}^{C} } \\ \hat{y}_{i}^{MU} & = \hat{b}_{0}^{MU} + \sum\limits_{j = 1}^{k} {\left[ {\hat{b}_{j1}^{MU} x_{ji}^{ML} + \hat{b}_{j2}^{MU} x_{ji}^{MU} } \right]} + \sum\limits_{j = k + 1}^{p} {\left[ {\hat{b}_{j1}^{MU} + \hat{b}_{j2}^{MU} } \right]x_{ji}^{C} } \\ \hat{y}_{i}^{VU} & = \hat{b}_{0}^{VU} + \sum\limits_{j = 1}^{k} {\left[ {\hat{b}_{j1}^{VU} x_{ji}^{VL} + \hat{b}_{j2}^{VU} x_{ji}^{VU} } \right]} + \sum\limits_{j = k + 1}^{p} {\left[ {\hat{b}_{j1}^{VU} + \hat{b}_{j2}^{VU} } \right]x_{ji}^{C} } \\ \end{aligned} $$
(32)

If all explanatory variables are crisp numbers, the above formulations become:

$$ \begin{aligned} \hat{y}_{i}^{VL} & = \hat{b}_{0}^{VL} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VL} + \hat{b}_{j2}^{VL} } \right]x_{ji}^{C} } \\ \hat{y}_{i}^{ML} & = \hat{b}_{0}^{ML} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{ML} + \hat{b}_{j2}^{ML} } \right]x_{ji}^{C} } \\ \hat{y}_{i}^{C} \, & = \hat{b}_{0}^{C} + \sum\limits_{j = 1}^{k} {\left[ {\hat{b}_{j1}^{C} + \hat{b}_{j2}^{C} } \right]x_{ji}^{C} } \\ \hat{y}_{i}^{MU} & = \hat{b}_{0}^{MU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{MU} + \hat{b}_{j2}^{MU} } \right]x_{ji}^{C} } \\ \hat{y}_{i}^{VU} & = \hat{b}_{0}^{VU} + \sum\limits_{j = 1}^{p} {\left[ {\hat{b}_{j1}^{VU} + \hat{b}_{j2}^{VU} } \right]x_{ji}^{C} } \\ \end{aligned} $$
(33)

Based on the above formulations, the proposed approach can deal with explanatory variables of various types, increasing flexibility. The signs of the TIFN parameters are determined in the solution process of a mathematical programming problem based on the criterion of the minimum distance between the observed and predicted TIFN responses. In addition, the mathematical programming model can be easily solved using commercial software, such as LINGO (Anderson et al. 2017).

4 Example and comparison

This study builds up a linear IFR model from mathematical programming problems with the criterion of least absolute deviation between the observed and predicted TIFN responses. Studies on IFR models are very limited. The approach proposed by Parvathi et al. (2013) uses crisp observation and explanatory variables and attempts to determine the IFR model with the least intuitionistic fuzziness, where all the given data can be included. As such, their approach cannot be compared with the approach proposed here. In this section, to demonstrate the proposed approach, the dataset from Arefi and Taheri (2015) is used to formulate an IFR model, and the performance of the model is compared to that of Arefi and Taheri (2015). The performance criteria include the similarity measure and distance measure proposed by Arefi and Taheri (2015). In addition, the distance measure proposed in this study is adopted.

Arefi and Taheri (2015) demonstrated their model using the TIFN dataset (see Table 1) given by Mohammadi and Taheri (2004). They fitted the least-squares regression model \( \hat{\tilde{Y}}_{AT} \) as:

Table 1 Dataset used in example (Mohammadi 2004)
$$ \begin{aligned} \hat{\tilde{Y}}_{AT} & = (19.9929,21.0878,21.9811,22.8744,23.9693) \\ & \quad \oplus ( - 0.2339, - 0.2338, - 0.2221, - 0.2104, - 0.2103) \otimes \tilde{X}_{1} \oplus (2.4701) \otimes \tilde{X}_{2} \\ \end{aligned} $$
(34)

Using the approach proposed here, the TIFN parameters can be solved from the model, as shown in Table 2. The IFR model \( \hat{\tilde{Y}}_{CN} \) is expressed as:

Table 2 TIFN parameters for the proposed approach
$$ \hat{\tilde{Y}}_{CN} = (20.7126,21.0663,21.0663,21.1706,21.7505) \oplus ( - 0.1969) \otimes \tilde{X}_{1} \oplus (2.6922) \otimes \tilde{X}_{2} $$
(35)

As shown in Table 2, only one dummy TIFN variable was obtained with the corresponding sign. In addition, the parameters of explanatory variables are crisp values to produce the smallest absolute deviation.

Examining the two models, namely Eqs. (34) and (35), the signs of the determined parameters from Arefi and Taheri (2015) and the proposed approach are the same. If traditional regression analysis is applied to build up a regression model using the central value of the TIFN data in the example, the regression estimators of the explanatory will be \( \hat{b}_{0} = 21.9767 \), \( \hat{b}_{1} = - 0.2221 \), and \( \hat{b}_{2} = 2.4727 \), indicating that the two approaches can produce the same signs of parameters and approximately equivalent values compared to those obtained using traditional regression analysis. However, the outcomes from Arefi and Taheri’s approach (Arefi and Taheri 2015) are questionable since the model formulation is based on least-squares regression analysis under the assumption that the parameters and TIFN explanatory variables are positive. In contrast, in the proposed approach, the signs of parameters are determined in the model formulation process.

Furthermore, Arefi and Taheri’s approach (2015) was developed based on symmetric TIFNs, and thus the estimated parameters and predicted response are also symmetric TIFNs. However, based on the definitions of the product operator for TIFNs given in Eqs. (11)–(13), the product of two TIFNs does not produce a symmetric TIFN, even if they are symmetric. The proposed predicted TIFN responses are not symmetric, which is more reasonable in theory.

Performance comparisons between Arefi and Taheri’s approach (2015) and the proposed approach based on the similarity measure \( SM(\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) [Eq. (20)], the distance measure \( d^{2} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) [Eq. (21)], and the absolute distance measure \( D_{TIFN} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) [Eq. (19)] were conducted item by item. The results are listed in Table 3. Although the similarity measure \( SM(\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) of the proposed model is 0.2% lower than that of the model obtained using Arefi and Taheri’s approach (2015), the proposed approach outperforms Arefi and Taheri’s approach (2015) in terms of distance measures \( d^{2} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) and \( D_{TIFN} (\tilde{Y}_{i} ,\hat{\tilde{Y}}_{i} ) \) by 13.2% and 13.7%, respectively.

Table 3 Results of performance comparison

These results show the feasibility and applicability of the proposed approach. The proposed approach can deal with different TIFN types in a dataset whether they are symmetric or asymmetric. The signs of TIFN parameters are known prior to formulating the model in the proposed approach. In addition, the linear formulation of the proposed mathematical programming problems increases computational performance. In general, the proposed approach is more generalized than existing ones.

5 Conclusion

This study used mathematical programming problems to build up IFR models. The least absolutely deviations between the predicted and observed TIFN responses are considered as the objective function, making the models more robust. The linear formulation of the proposed mathematical programming problems increases computational performance. The formulation of IFR models is derived based on the main components of an IFR model, i.e., the central value and lower and upper bounds of membership and non-membership functions. Unlike existing methods, the proposed approach does not limit observations to be symmetrical TIFNs. More importantly, the signs of parameters can be determined in the resolution process of finding the optimal parameters simultaneously. The proposed approach is general and can be used with TIFNs or crisp numbers. A performance comparison showed that the present IFR model outperforms an existing one in terms of distance measures. In future research, a more robust approach will be developed, and more applications will be used to demonstrate the applicability of the IFR model.