1 Introduction

The evolution of fuzzy set theory by Zadeh (1965) has drawn the attention of authors worldwide. Before this invention, probability was used to quantify the uncertainty. But to quantify the uncertainty using probability theory, the uncertainty should have been expressed as precise number. The vague terms for example ‘very, more, slightly’, etc. could not be quantified using probability theory. To quantify the uncertainty associated with such vague terms, the fuzzy set theory proposed by Zadeh (1965) proved to be an effective tool. Opposite to the classical set theory in which an element either belongs to the set or not, in fuzzy set, each element is assigned a number lying between 0 and 1 known as its membership degree. The practical relevance of fuzzy set in dealing with real world problems is now a proven fact (Chen and Chen (2014), Chen et al. (2009), Chen and Wang (2010), Chen et al. (2012a)). Fuzzy entropy is an important term related with fuzzy set. Fuzzy entropy measures the degree of fuzziness arising due to ambiguity that an element belongs to the set or not. Zadeh (1968) for the first time attempted to measure the fuzzy entropy as weighted Shannon entropy. But this proposal failed to serve the purpose. Another successful attempt to quantify the degree of fuzziness arising due to the ambiguity of being or not being the element of a set was made by Luca and Termini (1972). Also, they introduced a new fuzzy entropy based on Shannon entropy. In addition to this, Luca and Termini (1972) axiomatized the fuzzy entropy. With this axiomatization of fuzzy entropy, several authors proposed fuzzy entropies from their viewpoints and applied them in distinct fields like pattern recognition, image processing, etc. (Higashi and Klir (1982), Joshi and Kumar (2017a), Kaufman (1980), Yager (1979)).

Decision-making is another important field where the concept of fuzzy entropy is widely applied. Decision-making problems involve a set of alternatives satisfying a certain set of criteria. The main aim is to select an alternative which optimally satisfies the whole criteria. Such type of problems are also named as Multi-Criteria Decision-Making (MCDM) problems. Several methods have been proposed by researchers in the literature to solve MCDM problems (Arya and Kumar (1989) Chen et al. (2012b), Chen and Chang (2016), Zeng et al. (2019), Wang and Chen (2017), Chen and Chiou (2015), Chen et al. (2012c)). Criteria weights play an eminent role in the solution of a MCDM problem. Therefore, evaluation of criteria weights needs utmost attention. Estimating the importance of criteria weights in the solution od MCDM problems, Chen and Li (2010) bifurcated them into two parts: subjective weights and objective weights. Subjective weights are provided by the experts according to their preferences (Chu et al. (1979), Hwang and Lin (1987), Saaty (1980)). On the other hand, objective criteria weights are obtained by solving mathematical models without considering the experts’ viewpoint. Entropy method, principle element analysis (Choo and Wedley (1985), Fan (1996)) etc. are the examples of objective criteria weights. Among the objective methods, the entropy method is one of the most trusted approaches of determining criteria weights and has gained much popularity with researchers. In present communication, we will be using the entropy method for determining the criteria weights. Several MCDM methods based on this approach have been proposed so far.

TOPSIS (Technique for Order Preference by Similarity to Ideal Solution) is one of the well-known techniques used for solving MCDM problems (Joshi and Kumar 2017b, 2018b, c, d, e, c, Lalotra and Singh 2018, Singh et al. 2019, Lalotra and Singh 2020). In conventional TOPSIS method, an alternative which is nearest to the best solution and farthest from the worst solution is supposed to be the best alternative. The decision is made on the basis of relative closeness coefficients which are obtained by dividing the distance of an alternative from the worst solution by sum of the distances of the same alternative from the best and the worst solutions. To compute the distance between alternatives and the best as well as worst solutions, different distance measures are used. But the distance measures used to calculate the distances of an alternative from the best and the worst solutions just represent the distance between two numbers and do not give the sufficient information about their correlations. For instance, consider an example of a machine that is not functioning properly. Intuitively, we know that there is a strong relationship between symptoms and faulty parts of the machine. Now, in such a case, if we correlate the symptoms and the corresponding faulty part of the machine with the help of fuzzy sets then distance measure used will merely compute the distance between two numbers but not their correlation. This may cause the loss of necessary information and we may get erroneous results. To avoid such situation, in this paper, we propose a TOPSIS method based on weighted correlation coefficients introduced by Gerstenkorn and Manko (1991).

The parametric generalization of Shannon entropy ( Shannon (1948)) proposed by Renyi (1961) caused the attention of authors from across the world. For the first time, the researchers observed the role of parameters in an information measure. The presence of parameters in an information measure makes it flexible from application point of view and widens its scope of applications. For example, if we face a problem based on some environmental issue, different parameters may be used to represent the different environmental factors such as humidity, pressure, and temperature. More the number of parameters an information measure has, more will it be flexible for applications in diverse fields. Therefore, depending upon the nature of problem, we need to develop more and more flexible information measures which not only cater the need of the hour but also are the better performers. This study is a sincere effort in this direction. In this communication, a new bi-parametric exponential fuzzy information measure is proposed. The proposed information measure is a two parametric extension of exponential entropy studied by Pal and Pal (1989). The prime aims of introducing this communication are as follows:

  • to introduce a new bi-parametric exponential fuzzy information measure,

  • to introduce a new MCDM method based on entropy weights based correlation coefficients, and

  • to utilize the proposed information measure and proposed MCDM method in detecting the fault in a machine.

With this aim, the subject contents of this paper are managed as follows: the background literature of topic under consideration is covered in Sect. 1. The relevant definitions and concepts necessary related to topic are given in Sect. 2. A new fuzzy information measure along with its validation is proposed in Section   3. Besides discussing some major properties, the performance of proposed fuzzy information measure is also demonstrated through numerical examples in Sect. 3. A new MCDM method based on weighted correlation coefficients along with its justification is introduced in Sect. 4. The proposed method is explained using numerical examples in Sect.,5. Finally, the paper is concluded with conclusions in Sect. 6.

In next Section, we present some relevant concepts and definitions.

2 Preliminaries

Definition 2.1

(Fuzzy set (Zadeh (1965))). Let \(\sqcup = (\ell _1, \ell _2, \ldots , \ell _n)\) be a finite universe of discourse. A fuzzy set (FS) \(\odot \) in \(\sqcup \) is defined as

$$\begin{aligned} \odot =\{\langle \ell _i, \mu _\odot (\ell _i)\rangle |\ell _i\in \sqcup \}, \end{aligned}$$
(1)

where \(\mu _\odot : \sqcup \rightarrow [0, 1]\) is the membership function and \(\mu _\odot (\ell _i)\) represents the membership degree of \(\ell _i \in \sqcup \) to \(\odot \).

Definition 2.2

(Set operations on FSs (Zadeh (1965))). For any two FSs \(\odot _1\) and \(\odot _2\), some basic set operations are given by

  1. 1.

    \(\odot _1\cup \odot _2=\{\langle \ell _i, \max (\mu _{\odot _1} (\ell _i), \mu _{\odot _2} (\ell _i))\rangle | \ell _i\in \sqcup \}\);

  2. 2.

    \(\odot _1\cap \odot _2=\{\langle \ell _i, \min (\mu _{\odot _1} (\ell _i), \mu _{\odot _2} (\ell _i))\rangle | \ell _i\in \sqcup \}\);

  3. 3.

    \(\odot ^c=\{\langle \ell _i, (1-\mu _\odot (\ell _i)\rangle |\ell _i\in \sqcup \}\).

Definition 2.3

(Sharpened version (Luca and Termini (1972))). A FS \(\odot _1\) is said to be a sharpened version of another fuzzy set \(\odot _2\) if

  1. 1.

    \(\mu _{\odot _1} \,(\ell _i)\) is less than or equal to \(\mu _{\odot _2}\, (\ell _i)\) if \(\mu _{\odot _2}\, (\ell _i)\le \frac{1}{2}\).

  2. 2.

    \(\mu _{\odot _1} \,(\ell _i)\) is greater than or equal to \(\mu _{\odot _2} (\ell _i)\) if \(\mu _{\odot _2} (\ell _i)\ge \frac{1}{2}\).

Definition 2.4

(Fuzzy entropy (Luca and Termini (1972))). For any FS \(\odot \), a measure \(H(\odot )\) is said to be fuzzy information measure if it satisfies the following properties:

Property 1. The value of \(H(\odot )\) is zero if and only if \(\odot \) ia a crisp set. This property is called Sharpness.

Property 2. The value of \(H(\odot )\) is maximum if and only if \(\odot \) is most fuzzy set. This property is called Maximality.

Property 3. The value of \(H(\odot )\) is greater than equal to the value of \(H({\tilde{\odot }})\) if and only if \({\tilde{\odot }}\) is the sharpened version of \(\odot \). This property is known as Resolution.

Property 4. The value of  \(H(\odot )\) is equal to the value of  \(H(\odot ^c)\), where \(\odot ^c\) denotes the complement of \(\odot \). This property is known as Symmetry.

Definition 2.5

(Correlation coefficients (Gerstenkorn and Manko (1991))) For any two FSs  \(\odot _1\) and  \(\odot _2\), the correlation coefficient  \(\sqsubset (\odot _1, \odot _2)\) is given by

$$\begin{aligned} \sqsubset (\odot _1, \odot _2)=\sum _{i=1}^n \frac{\Lambda (\odot _1, \odot _2)}{\sqrt{\delta (\odot _1)\cdot \delta (\odot _2)}}, \end{aligned}$$
(2)

where  \(\Lambda (\odot _1, \odot _2)=\displaystyle \sum\nolimits _{i=1}^n\left( \mu _{\odot _1} (\ell _i)\mu _{\odot _2} (\ell _i)\right) \), \(\delta (\odot _1)=\displaystyle \sum\nolimits _{i=1}^n(\mu _{\odot _1} (\ell _i))^2\) and \(\delta (\odot _2)=\displaystyle \sum\nolimits _{i=1}^n(\mu _{\odot _2} (\ell _i))^2\).

The correlation coefficients  \(\sqsubset (\odot _1, \odot _2)\) satisfies the following properties:

  1. 1.

    \(0\le \sqsubset (\odot _1, \odot _2)\le 1\).

  2. 2.

    \(\sqsubset (\odot _1, \odot _2)=\sqsubset (\odot _2, \odot _1)\).

  3. 3.

    \(\sqsubset (\odot _1, \odot _2)=1\) if \(\odot _1=\odot _2\).

3 A new bi-parametric exponential fuzzy information measure

3.1 Background

Let \(\triangle _n=\{\hbar =(\hbar _1, \hbar _2, \ldots , \hbar _n);\hbar _i\ge 0, \sum _{i=1}^n \hbar _i=1\}\),\(n\ge 2\) be a set of complete probability distributions. For some \(\hbar \in \triangle _n\), Shannon entropy ( Shannon (1948)) is defined by

$$\begin{aligned} {}_{Sh}H (\hbar )=-\sum _{i=1}^n \hbar _i\log (\hbar _i). \end{aligned}$$
(3)

Afterwards, Renyi (1961) introduced a parametric extension of Shannon’s concept given by

$$\begin{aligned} {}_{Re}H^\varsigma (\hbar )= {\left\{ \begin{array}{ll} \frac{1}{1-\varsigma }\log \left( \sum _{i=1}^n \hbar _i^\varsigma \right) ,~ \varsigma >0 (\ne 1);\\ -\sum _{i=1}^n \hbar _i\log (\hbar _i), \varsigma =1. \end{array}\right. } \end{aligned}$$
(4)

This development caused the several authors to propose parametric extensions of Shannon entropy (Boekee and Lubbe (1980), Havdra and Charavat (1967), Joshi and Kumar (2016, 2018g), Tsallis (1988)). Pointing out the limitations of Shannon (1948), Pal and Pal (1989) introduced a new entropy known as exponential entropy given by

$$\begin{aligned} {}_{PP} H (\hbar )= \sum _{i=1}^n \hbar _i(e^{1-\hbar _i}-1);~\hbar \in \triangle _n. \end{aligned}$$
(5)

The introduction of FS by Zadeh (1965) altered the way of quantifying the fuzziness. Moreover, the pioneering study of fuzzy entropy by Luca and Termini (1972) inspired the authors worldwide to come up with several fuzzy information measures. Extending the Shannon’s concept, Luca and Termini (1972) introduced a fuzzy entropy. Following the way by Luca and Termini (1972), Pal and Pal (1989) proposed a new fuzzy entropy defined by

$$\begin{aligned} {}^{PP} H (\odot )=&\frac{1}{n(\sqrt{e}-1)}\sum _{i=1}^n\left( \mu _\odot (\ell _i)e^{1-\mu _\odot (\ell _i)}\right. \nonumber \\&\left. +(1-\mu _\odot (\ell _i))e^{\mu _\odot (\ell _i)}-1\right) . \end{aligned}$$
(6)

The concept of Pal and Pal (1989) was further generalized by Gupta et al. (2014) by introducing a parameter as follows:

$$\begin{aligned} {}^G H_\varsigma (\odot )=&\frac{1}{n\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }}-1\right) }\sum _{i=1}^n \left( \mu _\odot (\ell _i)^\varsigma e^{1-\mu _\odot (\ell _i)^\varsigma }\right. \nonumber \\&\left.+(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma }-1\right) ;~ 0<\varsigma <1. \end{aligned}$$
(7)

From the above discussion, the importance of parameters in an information measure can be easily estimated. In fact, the problems involving several parameters may be conveniently represented by a multi-parametric information measure. For example, an environmental problem may involve the parameters such as temperature, pressure, and humidity. Motivated by this, we now propose a new bi-parametric exponential fuzzy information measure.

3.2 Definition

For some FS \(\odot \), define

$$\begin{aligned} {}_\varsigma ^\varrho H (\odot )= {\left\{ \begin{array}{ll} \displaystyle \frac{1}{n\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}\right) }\sum _{i=1}^n \left[ \left( \mu _\odot (\ell _i)^\varsigma e^{(1-\mu _\odot (\ell _i)^\varsigma )}+(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma }\right) \right. \\ \displaystyle \left. -\left( \mu _\odot (\ell _i)^\varrho e^{(1-\mu _\odot (\ell _i)^\varrho )}+(1-\mu _\odot (\ell _i))^\varrho e^{1-{(1-\mu _\odot (\ell _i))}^\varrho }\right) \right] ,\\ \qquad \qquad \qquad \qquad \qquad \qquad \text {either}\, 0<\varsigma<1, 1<\varrho<\infty \,\text {or}\, 0<\varrho<1, 1<\varsigma <\infty ;\\ \displaystyle \frac{1}{n\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }}-2\right) }\sum _{i=1}^n \left( \mu _\odot (\ell _i)^\varsigma e^{(1-\mu _\odot (\ell _i)^\varsigma )}+(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma }-2\right) , \varsigma =0 \,\text {or}\, \varrho =0;\\ \displaystyle \frac{1}{n(\sqrt{e}-2)}\sum _{i=1}^n\left( \mu _\odot (\ell _i)e^{1-\mu _\odot (\ell _i)}+(1-\mu _\odot (\ell _i))e^{\mu _\odot (\ell _i)}-2\right) ; \varsigma =1, \varrho =0 \,\text {or}\,\varsigma =0, \varrho =1. \end{array}\right. } \end{aligned}$$
(8)

In (8), second and third cases, respectively, denote fuzzy entropies analogous to that of studied by Gupta et al. (2014) and Pal and Pal (1989).

The proposed information measure (8) being symmetric in nature, we will study only one case, that is, \(0<\varsigma<1, 1<\varrho <\infty \).

In the next section, we validate the existence of (8) by satisfying the axioms in Definition (2.4).

3.3 Validation of proposed measure (8)

Theorem 3.1

The proposed information measure (8) is a valid fuzzy information measure.

Proof

Property 1. First, we suppose that \(\odot \) is a crisp set, that is, \(\mu _\odot (\ell _i)=1\) or 0 for all \(\ell _i\in \sqcup \). Therefore, (8) gives

$$\begin{aligned} {}_\varsigma ^\varrho H(\odot )=0 ~\text {for~all} ~0<\varsigma<1, 1<\varrho <\infty . \end{aligned}$$
(9)

Conversely, suppose that  \({}_\varsigma ^\varrho H(\odot )=0\). This implies

$$\begin{aligned}&\frac{1}{\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}\right) }\nonumber \\&\quad \sum _{i=1}^n \left[ \left( \mu _\odot (\ell _i)^\varsigma e^{(1-\mu _\odot (\ell _i)^\varsigma )}+(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma }\right) \right. \nonumber \\&\quad \left. -\left( \mu _\odot (\ell _i)^\varrho e^{(1-\mu _\odot (\ell _i)^\varrho )}+(1-\mu _\odot (\ell _i))^\varrho e^{1-{(1-\mu _\odot (\ell _i))}^\varrho }\right) \right] =0. \end{aligned}$$
(10)

Since \(0<\varsigma<1, 1<\varrho <\infty \), therefore, we have

$$\begin{aligned} 2^{1-\varsigma }e^{1-2^{-\varsigma }}>2^{1-\varrho }e^{1-2^{-\varrho }}. \end{aligned}$$
(11)

Moreover, \(0\le \mu _\odot (\ell _i)\le 1\) implies

$$\begin{aligned}&\mu _\odot (\ell _i)^\varsigma e^{(1-\mu _\odot (\ell _i)^\varsigma )}+(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma }\nonumber \\&\quad \ge \mu _\odot (\ell _i)^\varrho e^{(1-\mu _\odot (\ell _i)^\varrho )}+(1-\mu _\odot (\ell _i))^\varrho e^{1-{(1-\mu _\odot (\ell _i))}^\varrho }. \end{aligned}$$
(12)

Therefore, from (11) and (12), we have

$$\begin{aligned} 0< \frac{A_1-B_1}{2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}}\le 1, \end{aligned}$$
(13)

where

$$\begin{aligned} A_1=\mu _\odot (\ell _i)^\varsigma e^{(1-\mu _\odot (\ell _i)^\varsigma )}+(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma };\\ B_1=\mu _\odot (\ell _i)^\varrho e^{(1-\mu _\odot (\ell _i)^\varrho )}+(1-\mu _\odot (\ell _i))^\varrho e^{1-{(1-\mu _\odot (\ell _i))}^\varrho }. \end{aligned}$$

This shows that  \({}_\varsigma ^\varrho H (\odot )=0\) implies

$$\begin{aligned}&\left( \mu _\odot (\ell _i)^\varsigma e^{(1-\mu _\odot (\ell _i)^\varsigma )}+(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma }\right) \nonumber \\&\quad =\left( \mu _\odot (\ell _i)^\varrho e^{(1-\mu _\odot (\ell _i)^\varrho )}+(1-\mu _\odot (\ell _i))^\varrho e^{1-{(1-\mu _\odot (\ell _i))}^\varrho }\right) . \end{aligned}$$
(14)

Since \(\varsigma , \varrho >0\) and \(\varsigma \ne \varrho \), then (14) will hold only if  \(\mu _\odot (\ell _i)=1\) or 0. This proves Property 1.

Property 2. Consider a function \(\psi \) given by

$$\begin{aligned} \psi (x, \varsigma , \varrho )=\frac{\left( x^\varsigma e^{1-x^\varsigma }+(1-x)^\varsigma e^{1-(1-x)^\varsigma }\right) -\left( x^\varrho e^{1-x^\varrho }+(1-x)^\varrho e^{1-(1-x)^\varrho }\right) }{2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}}, \end{aligned}$$
(15)

where \( 0<\varsigma<1, 1<\varrho <\infty \) and \(0\le x\le 1\).

Differentiating, (15) with respect to x, we get

$$\begin{aligned} \frac{\partial \psi }{\partial x}= \frac{A_2-B_2}{2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}}, \end{aligned}$$
(16)

where

$$\begin{aligned} A_2=&&\varrho x^{\varrho -1}(1-x^\varrho )e^{1-x^\varrho }-\varrho (1-x)^{\varrho -1}(1-(1-x)^\varrho )e^{(1-(1-x)^\varrho )},\\ B_2=&&\varsigma x^{\varsigma -1}(1-x^\varsigma )e^{1-x^\varsigma }-\varsigma (1-x)^{\varsigma -1}(1-(1-x)^\varsigma )e^{(1-(1-x)^\varsigma )}. \end{aligned}$$

To determine the stationary points, we express (16) as

$$\begin{aligned} \frac{\partial \psi }{\partial x}=\frac{{\widehat{A}}+{\widehat{B}}+{\widehat{C}}+{\widehat{D}}}{2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}}, \end{aligned}$$
(17)

where

$$\begin{aligned} {\widehat{A}}=&&\varrho x^{\varrho -1}(1-x^\varrho )e^{1-x^\varrho };\nonumber \\ {\widehat{B}}=&&-\varrho (1-x)^{\varrho -1}(1-(1-x)^\varrho )e^{(1-(1-x)^\varrho )};\nonumber \\ {\widehat{C}}=&&-\varsigma x^{\varsigma -1}(1-x^\varsigma )e^{1-x^\varsigma };\nonumber \\ {\widehat{D}}=&\varsigma (1-x)^{\varsigma -1}(1-(1-x)^\varsigma )e^{(1-(1-x)^\varsigma )}. \end{aligned}$$
(18)

On replacing x with \(1-x\) in \({\widehat{A}}\) and \({\widehat{C}}\), we find that \({\widehat{A}}+{\widehat{B}}=0={\widehat{C}}+{\widehat{D}}\). Therefore, \(\displaystyle \frac{\partial \psi }{\partial x}=0\) at  \(x=1-x\). This implies that  \(x=\frac{1}{2}\) is the stationary point.

To prove that  \(\psi \) has maximum value \(x=\frac{1}{2}\), we prove that  \(\psi \) is concave at \(x=\frac{1}{2}\). For this, differentiating (15) twice with respect to x and computing its value at \(x=\frac{1}{2}\), we get

$$\begin{aligned} \left( \frac{\partial ^2\psi }{\partial x^2}\right) _{x=\frac{1}{2}}= \frac{A_3-B_3}{2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}}, \end{aligned}$$
(19)

where

$$\begin{aligned} A_3=&& e^{1-.5^\varrho }\left( 4\varrho ^22^{2-2\varrho }-2\varrho ^22^{2-3\varrho }- 2\varrho (\varrho -1)2^{2-\varrho }\right. \\&&\left. +2\varrho (\varrho -1)2^{2-2\varrho }\right) ,\\ B_3=&&e^{1-.5^\varsigma }\left( 4\varsigma ^22^{2-2\varsigma }- 2\varsigma ^22^{2-3\varsigma }-2\varsigma (\varsigma -1)2^{2-\varsigma }\right. \\&& \left. +2\varsigma (\varsigma -1)2^{2-2\varsigma }\right) . \end{aligned}$$

Now, (19) is negative for all \(0<\varsigma<1, 1<\varrho <\infty \) which shows that the function  \(\psi \) is concave with maximum value at the point \(x=\frac{1}{2}\). This proves that \({}_\varsigma ^\varrho H(\odot )\) has maximum value when \(\odot \) is the most fuzzy set.

Property 3. In Property 2, it is proved that the measure \({}_\varsigma ^\varrho H(\odot )\) is a concave function with point of inflexion at \(x=\frac{1}{2}\). From this, it may be easily concluded that the measure \({}_\varsigma ^\varrho H(\odot )\) is an increasing function in the interval [0, 0.5) and decreasing function in the interval (0.5, 1]. This proves Property 3.

Property 4. This property follows directly from the definition.

Thus, the existence of proposed measure \({}_\varsigma ^\varrho H(\odot )\) is established.

In next section, we discuss some major properties of proposed measure (8).

3.4 Properties of measure (8)

Theorem 3.2

For any two fuzzy sets \(\odot _1\) and \(\odot _2\),  \({}_\varsigma ^\varrho H(\odot _1\cup \odot _2)+{}_\varsigma ^\varrho H(\odot _1\cap \odot _2)={}_\varsigma ^\varrho H(\odot _1)+{}_\varsigma ^\varrho H(\odot _2)\).

Corollary

For any fuzzy set \(\odot \) and its complement \(\odot ^c\),  \({}_\varsigma ^\varrho H(\odot \cup \odot ^c)+{}_\varsigma ^\varrho H(\odot \cap \odot ^c)={}_\varsigma ^\varrho H(\odot )+{}_\varsigma ^\varrho H(\odot ^c).\)

Theorem 3.3

The maximum value of  \({}_\varsigma ^\varrho H(\odot )\) takes place when  \(\odot \) is most fuzzy set and minimum value takes place when  \(\odot \) is least fuzzy set. Moreover, both of the values are independent of \(\varsigma \) and \(\varrho \).

Proof

The proofs of the abovementioned theorems as well as corollary are given in Appendix ‘A’. \(\square \)

3.5 A demonstration of performance

The adjectives like ‘slightly’, ‘more or less’ are used to modify the linguistic variables. Fuzzy sets proposed by Zadeh (1965) have been proved to be an effective tool to characterize the linguistic variables. Therefore, the adjectives such as very and slightly can be supposed as operations on fuzzy sets. Using these operations on fuzzy sets, in this section, we compare the performance of proposed measure (8) with some other existing fuzzy information measures in the literature. For a fuzzy set \(\odot \), the modifier of \(\odot \) is defined as

$$\begin{aligned} \odot ^n=\{\langle \ell _i, \mu _\odot (\ell _i)^n\rangle |\ell _i\in \sqcup \}. \end{aligned}$$
(20)

Hwang and Yang (2008) and Hung and Yang (2008) define the concentration of a fuzzy set \(\odot \) as  \(CON (\odot )=\odot ^2\) and dilation of a fuzzy set \(\odot \) as \(DIL (\odot )=\odot ^\frac{1}{2}\). The two terms, that is, concentration and dilation are frequently used to modify the fuzzy sets. Thus, we use these terms as mathematical operators to characterize the linguistic variables as follows:

$$\begin{aligned} \text {More or Less} (\odot )=&&DIL (\odot )=\odot ^\frac{1}{2}, ~\text {Very} A=CON (\odot )=\odot ^2;\nonumber \\ \text {Quite Very} (\odot )=&& \odot ^3, \text {Very Very} (\odot )=\odot ^4. \end{aligned}$$
(21)

Many authors (De et al. (2001), Hung and Yang (2006), Hwang and Yang (2008), Joshi and Kumar (2018a), Xia and Xu (2012)) have utilized these operators as a criteria of performance of a fuzzy set as follows:

$$\begin{aligned} H (\odot ^\frac{1}{2})>H(\odot )>H(\odot ^2)>H(\odot ^3)>H (\odot ^4), \end{aligned}$$
(22)

where \(H(\cdot )\) denotes the fuzzy entropy of a fuzzy set \(\odot \). Therefore, we use (22) as a criteria to compare the performance of proposed fuzzy information measure with other measures. Before comparison, we characterize the fuzzy sets \(\odot ^\frac{1}{2}, \odot ^2, \odot ^3, \odot ^4\) in terms of linguistic variables as follows:

$$\begin{aligned}&\text {Characterize}\,\odot ^\frac{1}{2}\,as\, ``{\rm More\, or \,Less\, Large}"\nonumber \\&\text {Characterize}\,\odot ^2 \,as\, ``{\rm Very\, Large}" \nonumber \\&\text{Characterize} \, \odot ^3 \,as\, ``{\rm Quite\, Very\, Large}"\nonumber \\&\text{Characterize}\,\odot ^4 \,as\, ``{\rm Very\, Very\, Large}". \end{aligned}$$
(23)

For comparison sake, we take two examples adapted from Joshi and Kumar (2018a).

Example 1

Consider a fuzzy set  \(\odot \)  defined on a universe of discourse \(X=(a, b, c, d, e)\) given by

$$\begin{aligned} \odot =\{(a, .1), (b, .3), (c, .4), (d, .9), (e, 1)\}. \end{aligned}$$
(24)

Characterizing the linguistic variables, we tag \(\odot \) as “Large". Using this characterization, following fuzzy sets may be generated:

$$\begin{aligned} \odot ^\frac{1}{2}=&\{(a, .316), (b, .548), (c, .632), (d, .949), (e, 1)\};\nonumber \\ \odot ^2=&\{(a, .010), (b, .090), (c, .160), (d, .810), (e, 1)\};\nonumber \\ \odot ^3=&\{(a, .001), (b, .027), (c, .064), (d, .0729), (e, 1)\};\nonumber \\ \odot ^4=&\{(a, 0), (b, .008), (c, .026), (d, .656), (e, 1)\}. \end{aligned}$$
(25)

To compare the performance of proposed measure, we consider the following well-known fuzzy information measures proposed by various authors given by

  1. 1.

    Fuzzy information measure proposed by Hwang and Yang (2008),

    $$\begin{aligned} H_{hy} (\odot )=&\frac{\sqrt{e}}{\sqrt{e}-1}\sum _{i=1}^n \left[ \left( 1-e^{-\mu _{\odot ^c} (\ell _i)}\right) I_{[\mu _{\odot ^c} (\ell _i)\ge \frac{1}{2}]}\right. \nonumber \\&\left. +\left( 1-e^{-\mu _{\odot } (\ell _i)}\right) I_{[\mu _{D^c} (\ell _i)< \frac{1}{2}]}\right] . \end{aligned}$$
    (26)
  2. 2.

    Fuzzy information measure introduced by Pal and Pal (1989),

    $$\begin{aligned} {}^{PP} H(\odot )=&\frac{1}{n}\sum _{i=1}^n \left[ \mu _\odot (\ell _i)e^{1-\mu _\odot (\ell _i)}\right. \nonumber \\&\left. +(1-\mu _\odot (\ell _i))e^{\mu _\odot (\ell _i)}-1\right] . \end{aligned}$$
    (27)
  3. 3.

    Fuzzy information measure proposed by Li and Liu (2008),

    $$\begin{aligned} H_{ll} (\odot )=\sum _{i=1}^n S(Cr (\xi _\odot =x_i)). \end{aligned}$$
    (28)
  4. 4.

    Fuzzy information measure introduced by Yager (1979),

    $$\begin{aligned} H_{yager} (\odot )=1-\frac{d_p (\odot , \odot ^c)}{n^\frac{1}{p}}. \end{aligned}$$
    (29)
  5. 5.

    Fuzzy information studied by Gupta et al. (2014),

    $$\begin{aligned}&{}^G H_\varsigma (\odot )\nonumber \\&\quad =\frac{1}{n\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }}-1\right) }\sum _{i=1}^n \left( \mu _\odot (\ell _i)^\varsigma e^{(1-\mu _\odot (\ell _i)^\varsigma )}\right. \nonumber \\&\qquad \left. +(1-\mu _\odot (\ell _i))^\varsigma e^{1-{(1-\mu _\odot (\ell _i))}^\varsigma }-1\right) ;\, 0<\varsigma <1. \end{aligned}$$
    (30)
  6. 6.

    Fuzzy information measure studied by Kosko (1986),

    $$\begin{aligned} H_{KO} (\odot )=\frac{d_p (\odot , \odot _{near})}{d_p (\odot , \odot _{far})}. \end{aligned}$$
    (31)

The computed values of fuzzy information measures (26) to (31) and proposed information measure (8) for the fuzzy sets defined in (25) are displayed in Table 1.

Table 1 Computed values of information measures (26)–(31) and (8)

The observations obtained from Table 1 are summarized as follows:

$$\begin{aligned}&H_{hy} (\odot ^\frac{1}{2})>H_{hy} (\odot )>H_{hy} (\odot ^2)> H_{hy} (\odot ^3)> H_{hy} (\odot ^4);\nonumber \\&{}^{PP}H (\odot ^\frac{1}{2})>{}^{PP}H (\odot )>{}^{PP}H (\odot ^2)> {}^{PP}H (\odot ^3)> {}^{PP}H (\odot ^4);\nonumber \\&H_{KO} (\odot ^\frac{1}{2})<H_{KO} (\odot )>H_{KO} (\odot ^2)> H_{KO} (\odot ^3)> H_{KO} (\odot ^4);\nonumber \\&H_{ll} (\odot ^\frac{1}{2})>H_{ll} (\odot )>H_{ll} (\odot ^2)> H_{ll} (\odot ^3)> H_{ll} (\odot ^4);\nonumber \\&H_{yager} (\odot ^\frac{1}{2})>H_{yager} (\odot )>H_{yager} (\odot ^2)> H_{yager} (\odot ^3)> H_{yager} (\odot ^4);\nonumber \\&{}^GH_\varsigma (\odot ^\frac{1}{2})>{}^GH_\varsigma (\odot )>{}^GH_\varsigma (\odot ^2)> {}^GH_\varsigma (\odot ^3)> {}^GH_\varsigma (\odot ^4);\nonumber \\&{}_\varsigma ^\varrho H (\odot ^\frac{1}{2})>{}_\varsigma ^\varrho H (\odot )>{}_\varsigma ^\varrho H (\odot ^2)> {}_\varsigma ^\varrho H (\odot ^3)> {}_\varsigma ^\varrho H (\odot ^4). \end{aligned}$$
(32)

From (32), it may be observed that \(H_{hy}, {}^{PP} H, H_{ll}, H_{yager}, {}^G H_\varsigma \) and \({}_\varsigma ^\varrho H\) follow the pattern (22) whereas \(H_{KO}\) do not. This shows that the performance of measure \(H_{KO}\) is not better as compared to other measures.

For further comparison, we consider another Example.

Example 2

On a universe of discourse \(X=\{{\widehat{a}}, {\widehat{b}}, {\widehat{c}}, {\widehat{d}}, {\widehat{e}}\}\), consider another fuzzy set defined by

$$\begin{aligned} {\widetilde{\odot }}=\{({\widehat{a}}, .2), ({\widehat{b}}, .3), ({\widehat{c}}, .4), ({\widehat{d}}, .7), ({\widehat{e}}, 1)\}. \end{aligned}$$
(33)

Again, using (20), we generate the following fuzzy sets.

$$\begin{aligned} {\widetilde{\odot }}^\frac{1}{2}=&\{({\widehat{a}}, .447), ({\widehat{b}}, .548), ({\widehat{c}}, .632), ({\widehat{d}}, .837), ({\widehat{e}}, 1)\};\nonumber \\ {\widetilde{\odot }}^2=&\{({\widehat{a}}, .040), ({\widehat{b}}, .090), ({\widehat{c}}, .160), ({\widehat{d}}, .490), ({\widehat{e}}, 1)\};\nonumber \\ {\widetilde{\odot }}^3=&\{({\widehat{a}}, .008), ({\widehat{b}}, .027), ({\widehat{c}}, .064), ({\widehat{d}}, .0343), ({\widehat{e}}, 1)\};\nonumber \\ {\widetilde{\odot }}^4=&\{({\widehat{a}}, .002), ({\widehat{b}}, .008), ({\widehat{c}}, .026), ({\widehat{d}}, .240), ({\widehat{e}}, 1)\}. \end{aligned}$$
(34)

The computed values of \(H_{hy}, {}^{PP} H, {}^G H_\varsigma \) and \({}_\varsigma ^\varrho H\) are shown in Table 2.

Table 2 Computed values of information measures (26)–(31) and (8)

A Discussion on performance: Thus, all the four measures, that is, \(H_{hy}, {}^{PP} H, {}^G H_\varsigma \) and \({}_\varsigma ^\varrho H\), considered in Table 2 follow the sequence given in (22). This conclusion generates a natural query in mind that which information measure is better than others. On analyzing the eight fuzzy information measures considered so far, we observe that all the information measures except the proposed one are either rigid or contains one parameter. Therefore, the measures (26), (27), (28), (29) and (31) are not suitable for the applications which involve parameters. Also, the measure (30) is one parametric which restricts its scope of application to the problems involving one parameter only. On the other hand, the proposed measure (8) is a bi-parametric information measure which enhances its scope of applications. This establishes the fact that the proposed measure (8) is effective and its performance is considerably good.

4 Application of proposed measure in decision-making

A multi-criteria decision-making (MCDM) is the process of selecting the most desirable alternative satisfying the established criteria. As discussed earlier, the criteria weights play a decisive role in the selection of the most appropriate alternative. The improper evaluation of criteria weights may lead to the selection of an inferior alternative which ultimately may appear in the form of loss. Therefore, the evaluation of criteria weights needs utmost attention while solving a MCDM problem. In the present study, we adopt two approaches to evaluate criteria weights: First approach includes the case when we have partial information about criteria weights and second approach is based on when the criteria weights are incompletely known or completely unknown.

4.1 A new TOPSIS method based on weighted correlation coefficients

4.1.1 Justification of proposed method

Technique for Order Preference by Similarity to Ideal Solutions (TOPSIS) is one of the widely used techniques to solve MCDM problems. In TOPSIS method, two extreme solutions, that is, the best possible solution and the worst possible solution are determined and an alternative closest to the best solution and farthest from the worst solution is chosen as the most desirable alternative. To check the degree of closeness with extreme solutions, different distance measures are used. Now, it has been proved that results of TOPSIS method vary with the change of distance measure used (Joshi and Kumar (2018f)). Therefore, instead of using distance measure based TOPSIS method, a new improved TOPSIS method based on correlation coefficients is being proposed in this study. Further, Ye (2010) argued that the techniques used in fuzzy decision-making either use score function or accuracy functions that do not give sufficient information about alternatives. Therefore, these methods are not reliable and a substitute is necessary for decision-making. With this viewpoint, Ye (2010) proposed an alternative method for fuzzy decision-making. But in his proposed technique, Ye (2010) considered the correlation between alternatives and the best solution only thus neglecting the correlation between alternatives and the worst solution.

In the next Subsection, we propose a MCDM method by considering the correlation of alternatives with the best and the worst possible solutions.

4.1.2 The proposed method

The procedural steps of proposed MCDM method are as follows:

Step 1. Preparation of fuzzy decision matrix

Represent the given MCDM problem in matrix form with rows representing alternatives  \(\curlyvee _i (i=1, 2, \ldots , m)\)  and columns denoting criteria  \(\Psi _j (j=1, 2, \ldots , n)\)  as follows:

$$\begin{aligned} \begin{array}{cccc} \Psi _1&\Psi _2&\ldots&\Psi _n \end{array}\qquad \nonumber \\ {\widetilde{M}}_{m\times n}= \begin{array}{c} \curlyvee _1\\ \curlyvee _2 \\ \vdots \\ \curlyvee _m \end{array} \left( \begin{array}{cccc} {\widehat{\mu }}_{11} &{} {\widehat{\mu }}_{12} &{} \ldots &{} {\widehat{\mu }}_{1n}\\ {\widehat{\mu }}_{21} &{} {\widehat{\mu }}_{22} &{} \ldots &{} {\widehat{\mu }}_{2n}\\ \vdots &{} \vdots &{} \ddots &{} \vdots \\ {\widehat{\mu }}_{m1} &{}{\widehat{\mu }}_{m2} &{} \ldots &{} {\widehat{\mu }}_{mn} \end{array} \right) ; \end{aligned}$$
(35)

where \({\widehat{\mu }}_{ij}=\mu \left( \curlyvee _i, \Psi _j\right) \) denotes the degree of satisfaction of the alternative  \(\curlyvee _i\) to the criteria  \(\Psi _j\). To determine the degree of satisfaction of  \(\curlyvee _i\)’s corresponding to criteria  \(\Psi _j\)’s, we apply the following method:

$$\begin{aligned} \mu _{ij}=\frac{n_y(i, j)}{N} \end{aligned}$$
(36)

where \(n_y (i, j)\) denotes the number of experts who support  \(\curlyvee _i\) corresponding to criteria  \(\Psi _j\) and N represents the total number of experts.

Step 2. Normalization of decision matrix

To treat all the criteria at a par, normalize the matrix obtained in Step 1 as follows:

$$\begin{aligned} \mu _{ij}= {\left\{ \begin{array}{ll} {\widehat{\mu }}_{ij} & \text {for Benefit Criteria;}\\ (1-{\widehat{\mu }}_{ij}) & \text {for Cost Criteria}. \end{array}\right. } \end{aligned}$$
(37)

Let the matrix so obtained be represented by  M

$$\begin{aligned} \begin{array}{cccc} \Psi _1&\Psi _2&\ldots&\Psi _n \end{array}\qquad \nonumber \\ M_{m\times n}= \begin{array}{c} \curlyvee _1\\ \curlyvee _2 \\ \vdots \\ \curlyvee _m \end{array} \left( \begin{array}{cccc} \mu _{11} &{} \mu _{12} &{} \ldots &{} \mu _{1n}\\ \mu _{21} &{} \mu _{22} &{} \ldots &{} \mu _{2n}\\ \vdots &{} \vdots &{} \ddots &{} \vdots \\ \mu _{m1} &{} \mu _{m2} &{} \ldots &{} \mu _{mn} \end{array} \right) ; \end{aligned}$$
(38)

. Step 3. Determination of criteria weights

As discussed earlier, the criteria weights play an important role in the solution of a MCDM problem. This implies that evaluation of criteria weights needs due attention. With this point in mind, we bifurcate the process of determination of criteria weights into two parts as follows:

Approach I: If the Information Available about criteria Weights is Partial

The advice of experts involved in decision-making process matters very much. But all the experts may not be well-versed with all the aspects of a problem. Therefore, it is not possible to have such reliable advice every time. This may also be due to lack of time, limited expertise about problem domain etc. that experts do not express themselves in the form of precise numbers. In fact, they prefer to express themselves in the form of intervals. We compile this available partial information about criteria weights in the form of a set denoted by \(\Sigma \). To determine the criteria weights, we use the principle of minimum entropy suggested by Wang and Wang (2012) as follows:

The total entropy of an alternative \(\curlyvee _i\) across all the criteria  \(\Psi _j\) is given by

$$\begin{aligned}&{}_\varsigma ^\varrho H(\curlyvee _i) =\sum _{j=1}^n {}_\varsigma ^\varrho H(\mu _{ij})\nonumber \\&\quad =\sum _{j=1}^n\left( \frac{1}{n\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }} -2^{1-\varrho }e^{1-2^{-\varrho }}\right) }\right. \nonumber \\&\qquad \left. \left[ \left( \mu _{ij}^\varsigma e^{(1-\mu _{ij}^\varsigma )}+(1-\mu _{ij})^\varsigma e^{1-{(1-\mu _{ij})}^\varsigma }\right) \right. \right. \nonumber \\&\qquad \left. \left. -\left( \mu _{ij}^\varrho e^{(1-\mu _{ij}^\varrho )} +(1-\mu _{ij})^\varrho e^{1-{(1-\mu _{ij})}^\varrho }\right) \right] \right) ;\nonumber \\&\quad =\frac{1}{n\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }} -2^{1-\varrho }e^{1-2^{-\varrho }}\right) }\nonumber \\&\qquad \sum _{j=1}^n \left[ \left( \mu _{ij}^\varsigma e^{(1-\mu _{ij}^\varsigma )}+(1-\mu _{ij})^\varsigma e^{1-{(1-\mu _{ij})}^\varsigma }\right) \right. \nonumber \\&\qquad \left. -\left( \mu _{ij}^\varrho e^{(1-\mu _{ij}^\varrho )}+(1-\mu _{ij})^\varrho e^{1-{(1-\mu _{ij})}^\varrho }\right) \right] . \end{aligned}$$
(39)

Since, each alternative is a fair competition, then, the weight coefficients corresponding to same criteria should also be same. Therefore, to determine the optimal criteria weights, we construct the following programming model:

$$\begin{aligned}&\min (I)=\sum _{i=1}^m \left( w_j \left( {}_\varsigma ^\varrho H \left( \curlyvee _i\right) \right) \right) \nonumber \\&\quad =\frac{1}{n\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }}-2^{1-\varrho }e^{1-2^{-\varrho }}\right) }\nonumber \\&\qquad \sum _{i=1}^m\sum _{j=1}^n \left( w_j \left[ \left( \mu _{ij}^\varsigma e^{(1-\mu _{ij}^\varsigma )}+(1-\mu _{ij})^\varsigma e^{1-{(1-\mu _{ij})}^\varsigma }\right) \right. \right. \nonumber \\&\qquad \left. \left. -\left( \mu _{ij}^\varrho e^{(1-\mu _{ij}^\varrho )}+(1-\mu _{ij})^\varrho e^{1-{(1-\mu _{ij})}^\varrho }\right) \right] \right) \end{aligned}$$
(40)

subject to the condition  \(\sum _{j=1}^n w_j=1, w_j\in \Sigma \). On solving (40), we get the criteria weight vector in the form   \(\arg \min E= (w_1, w_2, \ldots , w_n)'\) as an optimal solution, where  \('\)  denotes the transpose.

Approach II: If the criteria Weights are Unknown

To obtain the criteria weights in this case, we employ the method suggested by Chen and Li (2010) as follows:

$$\begin{aligned} w_j=\frac{1- {}_\varsigma ^\varrho H (\mu _{ij})}{n-\sum _{j=1}^n \left( {}_\varsigma ^\varrho H (\mu _{ij})\right) }, \end{aligned}$$
(41)

where

$$\begin{aligned}{}_\varsigma ^\varrho H(\mu _{ij})=&\frac{1}{\left( 2^{1-\varsigma }e^{1-2^{-\varsigma }} -2^{1-\varrho }e^{1-2^{-\varrho }}\right) } \nonumber \\& \sum _{i=1}^m\left[ \left( \mu _{ij}^\varsigma e^{(1-\mu _{ij}^\varsigma )}+(1-\mu _{ij})^\varsigma e^{1-{(1-\mu _{ij})}^\varsigma }\right) \right. \nonumber \\& \left. -\left( \mu _{ij}^\varrho e^{(1-\mu _{ij}^\varrho )}+(1-\mu _{ij})^\varrho e^{1-{(1-\mu _{ij})}^\varrho }\right) \right] . \end{aligned}$$
(42)

Step 4. Determination of extreme solutions

Determine the best solution \((\curlyvee ^\star )\) as well as worst solution  \((\curlyvee _\star )\)  as follows:

$$\begin{aligned} \curlyvee ^\star =\max {(\mu _{ij})} \,\text {and}\, \curlyvee _\star =\min {(\mu _{ij})}; \,\text {for all}\, i=1, 2, \ldots m; j=1, 2, \ldots , n. \end{aligned}$$
(43)

Step 5. Determination of weighted correlation coefficients

Determine the weighted correlation coefficients  \(\curlyvee _i\)’s\( (i=1, 2, \ldots , m)\) with \(\curlyvee ^\star \) and \(\curlyvee _\star \) using Definition (2.5) as follows:

$$\begin{aligned} \sqsubset (\curlyvee _i, \curlyvee ^\star )=\frac{w_j\Lambda (\curlyvee _i, \curlyvee ^\star )}{\sqrt{w_j \left( \delta (\curlyvee _i)\cdot \delta (\curlyvee ^\star )\right) }};\nonumber \\ \sqsubset (\curlyvee _i, \curlyvee _\star )=\frac{w_j\Lambda (\curlyvee _i, \curlyvee _\star )}{\sqrt{w_j\left( \delta (\curlyvee _i)\cdot \delta (\curlyvee _\star )\right) }}. \end{aligned}$$
(44)

Step 6. Determination of relative closeness coefficients

Determine the relative correlation coefficients  \(\chi _i (i=1, 2, \ldots , m)\) as follows:

$$\begin{aligned} \chi _i=\frac{\sqsubset (\curlyvee _i, \curlyvee _\star )}{\sqsubset (\curlyvee _i, \curlyvee _\star )+\sqsubset (\curlyvee _i, \curlyvee ^\star )}. \end{aligned}$$
(45)

Step 7. Ranking of alternatives

Arranging the  \(\curlyvee _i\)’s according to the values of  \(\chi _i\)’s\((i=1, 2, \ldots , m)\) in descending order, we obtain the preferential sequence of alternatives.

In subsequent Section, we explain the proposed MCDM method through numerical examples.

5 Illustrated numerical examples

Vagueness in scientific studies is presenting a challenging dimensions. Fuzzy set theory proposed by Zadeh (1965) has proved to be effective tool to characterize such vagueness. To counter the menace of vagueness in scientific studies, there is a need to develop information measures and relevant methods which can measure the vagueness in the underlying characterizing fuzzy sets. The example considered in this Section highlights the need of such measures and methods. Consider an example of a factory in which a machine is not functioning properly. In large-scale units, such occurrences are frequent and needs immediate attention otherwise it may slow down the production rate which ultimately may appear in the form of loss to the company. To keep the machine going on smoothly, the immediate detection of fault in the machine is necessary. To locate the fault in the machine at the earliest can be one of the best examples of the applied intelligence. Since, symptoms the machine is showing and the part of the machine in which fault is located are strongly correlated, therefore, the proposed MCDM method can be applied successfully in this case.

5.1 For partially known criteria weights

Example

To perform the task, we broadly divide the machine under observation into five parts say  \(\curlyvee _1, \curlyvee _2, \curlyvee _3, \curlyvee _4, \curlyvee _5\). Let  \((\Psi _1, \Psi _2, \Psi _3, \Psi _4)\) denote the set of symptoms bearing certain relationship with faulty sections that the machine is indicating on the basis of which the faulty section is to be fixed. Sometimes, different faults have common symptoms which give rise to fuzziness in determining the exact faulty part of the machine. Fuzzy sets proposed by Zadeh (1965) are helpful in dealing with such vagueness. Therefore, we use the proposed fuzzy information measure for the purpose. Using (36), let the degree of relationship that a symptom bears with a particular part of the machine under observation be represented by the fuzzy decision matrix given by

Table 3 Fuzzy decision matrix  \((M)_{5\times 4}\)

Let the information available about criteria weights be denoted by the set  \(\Sigma \) given by

$$\begin{aligned} \Sigma =\{0\le w_1\le .3, .1\le w_2\le .2, .2\le w_3\le .5, .1\le w_4\le .3\} \end{aligned}$$
(47)

subject to the condition  \(\sum _{j=1}^4 w_j=1\).

Now, we apply the proposed MCDM method to select the faulty section of the machine. The computational procedure is as follows:

Step 1. The given MCDM problem is represented by the matrix (46).

Step 2. Since, we are dealing with the problem of detection of fault in the machine, therefore, there is no benefit or cost criteria. Therefore, the normalized fuzzy decision matrix is represented by the fuzzy decision matrix (46) itself.

Step 3. To determine the weight-age of symptoms, we proceed as follows:

  1. 1.

    Using proposed information measure (8), the computed values of information for each symptom \(\Psi _j; (j=1, 2, 3, 4)\), that is, \({}_\varsigma ^\varrho H (\Psi _j)\) are shown in last row of Table 1.

  2. 2.

    Using (40), we construct the the following programming model:

    $$\begin{aligned} \min (I)=&.9601 w_1+.8742 w_2+ .9221 w_3+ .9821 w_4\nonumber \\&\text {subject to}~ {\left\{ \begin{array}{ll} 0\le w_1\le .3;\\ .1\le w_2\le .2;\\ .2\le w_3\le .5;\\ .1\le w_4\le .3;\\ \sum _{i=1}^4 w_i=1. \end{array}\right. } \end{aligned}$$
    (48)
  3. 3.

    Solving the linear programming model (48) using MATLAB, the criteria weights so obtained are given by

    $$\begin{aligned} w_1=.2, w_2=.2, w_3=.5, w_4=.1. \end{aligned}$$
    (49)

    The syntax for MATLAB code is provided in Appendix ‘B’.

Step 4. Using (43), the compute values of  \(\curlyvee ^\star \) and  \(\curlyvee _\star \) are given by

$$\begin{aligned} \curlyvee ^\star =(.8, .8, .6, .7)\, {\rm and}\, \curlyvee _\star =(.6, .1, .1, .3). \end{aligned}$$
(50)

Step 5. The computed values of weighted correlation coefficients using (44) are given by

$$\begin{aligned} \sqsubset (\curlyvee _1, \curlyvee ^\star )=&.9636, \sqsubset (\curlyvee _2, \curlyvee ^\star )=.9536, \sqsubset (\curlyvee _3, \curlyvee ^\star )\nonumber \\ =&.9420, \sqsubset (\curlyvee _4, \curlyvee ^\star )=.8482, \sqsubset (\curlyvee _5, \curlyvee ^\star )=.7525;\nonumber \\ \sqsubset (\curlyvee _1, \curlyvee _\star )=&.6742, \sqsubset (\curlyvee _2, \curlyvee _\star )=.7023, \sqsubset (\curlyvee _3, \curlyvee _\star )=.6725, \sqsubset (\curlyvee _4, \curlyvee _\star )\nonumber \\ =&.5999, \sqsubset (\curlyvee _5, \curlyvee _\star )=.5785. \end{aligned}$$
(51)

Step 6. The computed values of relative correlation coefficients  \(\chi _i (i=1, 2, 3, 4, 5)\) using (45) are given by

$$\begin{aligned} \chi _1=.4117, \chi _2=.4241, \chi _3=.4166, \chi _4=.4143, \chi _5=.4346. \end{aligned}$$
(52)

Step 7. Arranging the sections  \(\curlyvee _1, \curlyvee _2, \curlyvee _3, \curlyvee _4, \curlyvee _5\) of the machine, according to the values of  \(\chi _i\)s in descending order, the preferential sequence of faulty sections so obtained is given by

$$\begin{aligned} \curlyvee _5\succ \curlyvee _2\succ \curlyvee _3\succ \curlyvee _4\succ \curlyvee _1. \end{aligned}$$
(53)

Thus,  \(\curlyvee _5\) is the section of the machine in which fault is located.

A comparative analysis: To compare the performance of proposed method, the same Example was computed using the weighted correlation coefficients method proposed by Ye (2010) and the output so obtained is given by  \(\curlyvee _1\succ \curlyvee _2\succ \curlyvee _3\succ \curlyvee _4\succ \curlyvee _5\) with  \(\curlyvee _1\) as the best alternative. If we compute the above example by using conventional fuzzy TOPSIS method and using the Hamming distance as the distance measure, the preferential sequence of alternatives so obtained is given by:  \(\curlyvee _2\succ \curlyvee _1\succ \curlyvee _3\succ \curlyvee _4\succ \curlyvee _5\) with  \(\curlyvee _2\) as the best alternative. Thus, all the three methods used to determine the best alternative in same example are producing the different outputs. Now, a natural question that arise in mind is that which output is most reliable. This is an established fact that different algorithms are defined with different viewpoints ( Joshi and Kumar (2017c)). Due to this difference in viewpoints, the weight-age given to different criteria in a problem may vary which ultimately may change the output. We start with Ye’s method ( Ye (2010)). Ye’s method ( Ye (2010)) is based on the correlation of alternatives with the best solution only whereas the correlation with the worst solution is not considered. Even in TOPSIS method, which is one of the widely used method, the best alternative is decided on the basis of relative closeness coefficients in which distances of the alternative from both solutions, that is, the best solution and the worst solution are considered. This is due to this reason that the most preferred alternative obtained using Ye’s method may not be always be the best one. Moreover, as mentioned earlier in Sect. 1, that is, ‘Introduction’, a distance measure based TOPSIS method do not consider the correlation between two objects due to which a certain amount of useful information may be lost. Therefore, the output of a such a TOPSIS method may not be reliable. Thus, in MCDM problems where the criteria bears certain correlation with alternatives, the proposed MCDM method may be the best choice. This establishes the efficacy of proposed MCDM method.

5.2 If criteria weights are unknown

Now, we compute the above Example for the case when criteria weights are unknown to us. The computational steps are as follows:

Step 1. The computed values of symptoms weights using (41) are given by

$$\begin{aligned} w_1=.1526; w_2=.4811; w_3=.2979; w_4=.0685. \end{aligned}$$
(54)

Step 2. The computed values of  \(\curlyvee ^\star \) and  \(\curlyvee _\star \) are same as given in (50).

Step 3. The computed values of weighted correlation coefficients using (44) are given by

$$\begin{aligned} \sqsubset (\curlyvee _1, \curlyvee ^\star )=&.9695, \sqsubset (\curlyvee _2, \curlyvee ^\star )=.9621, \nonumber \\ \sqsubset (\curlyvee _3, \curlyvee ^\star )=&.9458, \sqsubset (\curlyvee _4, \curlyvee ^\star )=.7673, \sqsubset (\curlyvee _5, \curlyvee ^\star )=.7995;\nonumber \\ \sqsubset (\curlyvee _1, \curlyvee _\star )=&.6315, \sqsubset (\curlyvee _2, \curlyvee _\star )=.7120, \sqsubset (\curlyvee _3, \curlyvee _\star )=.6213,\nonumber \\ \sqsubset (\curlyvee _4, \curlyvee _\star )=&.5196, \sqsubset (\curlyvee _5, \curlyvee _\star )=.5358. \end{aligned}$$
(55)

Step 4. The calculated values of  \(\chi _i\)’s\((i=1, 2, \ldots , m)\) are given by

$$\begin{aligned} \chi _1=.3944, \chi _2=.4253, \chi _3=.3965, \chi _4=.4038, \chi _5=.4013. \end{aligned}$$
(56)

Step 5. Applying Step 7 of proposed method, the preferential sequence so obtained is given by

$$\begin{aligned} \curlyvee _2\succ \curlyvee _4\succ \curlyvee _5\succ \curlyvee _3\succ \curlyvee _1. \end{aligned}$$
(57)

Thus, this is the \(\curlyvee _2\) section in which fault is located.

Moreover, the proposed decision-making model may also be applied in other real-world applications, for example, to select the most appropriate supplier for a company, to select the most suitable site for project installation, for purchasing a suitable flat meeting almost all the requirements of a common man, to invest a sum of money to have good returns and any other problem involving conflicting criteria and fuzziness in decision-making.

6 Conclusions

In this paper, we have successfully introduced a new fuzzy information measure as an extension of the fuzzy information measure studied by Gupta et al. (2014). A new modified TOPSIS method based on weighted correlation coefficients is proposed. Estimating the importance of criteria weights, two methods of determining the criteria weights are discussed: for partially known criteria weights and for unknown criteria weights. Further, the proposed information measure and proposed MCDM method have been utilized in locating the fault in a machine.

Limitations and future scope: One of the major limitations of the proposed work is that it is based on the complete probability distribution, that is, \((\hbar _1, \hbar _2, \ldots , \hbar _n)\) such that \(\sum _{i=1}^n \hbar _i=1\). In practice, this may not always the case, that is, sum of probabilities may exceed one or lesser than one, for example D-numbers studied by Deng (2012). Another limitation of this study is that the proposed decision-making method is based on fuzzy information only whereas in practice there exists several generalizations of fuzzy sets, for example, intuitionistic fuzzy sets studied by Atanassov (1986), hesitant fuzzy set introduced by Torra and Narukawa (2009), neutrosophic set Smarandache (2006) etc. Therefore, there is need to develop more general decision-making methods. The work is under consideration and will be reported somewhere else.