1 Introduction

The fuzzy sets [1] allow an element under consideration to have a partial membership degree of belongingness to a set instead of the classical cases of “belongs to” or “not belongs to”. This revolutionary idea of fuzzy sets which can accommodate uncertainty naturally has been studied and applied extensively by researchers across different communities. Later, it was realized that, like the partial grade of membership for belongingness, the non-membership grade is equally important. Thereby, intuitionistic fuzzy sets (IFSs) were introduced and studied by Atanossov [2,3,4,5] which is a generalization of fuzzy set. IFSs consider degrees of membership (\(\mu \)) as well as non-membership (\(\nu \)), with their sum restricted to a value less than equals to one, for an element in the universe of discourse. Obviously, for fuzzy sets the non-membership grade is always fuzzy complement of the membership grade. Considering \(\mu \) as abscissa and \(\nu \) as ordinate, the area under the line \(\mu + \nu = 1\) in the first quadrant geometrically represents the admissible region for IFSs. There are instances where the sum of membership and non-membership grades exceeds one, viz. \(\mu = \nu = \frac{1}{\sqrt{2}}\), and these can’t be represented by IFSs. It necessitates a further generalization of IFSs to encompass such cases. Subsequently, Pythagorean fuzzy sets (PFSs) were introduced by Yager [6, 7] which allow a larger admissible area with sum of the squares of \(\mu \) and \(\nu \) is less than equals to one. It is evident that, every fuzzy set is an IFS, each IFS is a PFS, but not conversely. PFSs allow more accessible area than IFSs, so it can efficiently and accurately handle uncertainty. Because of that, PFSs have drawn the attention of researchers and are being applied in myriad fields like pattern recognition [8,9,10,11,12], decision making [13,14,15,16,17,18,19], medical diagnosis [20,21,22,23,24] and others [25,26,27,28,29,30,31], resulting in improved outcomes, since inception.

To estimate the similarity or difference between PFSs, various distance and similarity measures have been introduced and studied in the literature. These measures are successfully employed to different regime of applications; few among them are delineated here. On the basis of the Minkowski distance measure, Chen [32] introduced a distance for PFSs and applied it to problems of Internet stock and R&D project investment along with some other real-world problems. In [33], the authors proposed several distance measures for PFSs and Pythagorean fuzzy numbers and demonstrated their usefulness via examples. Wei & Wei [23] presented ten different types of similarity measures for PFSs based on cosine function and illustrated their serviceability in medical diagnosis & pattern recognition problems. Ejegwa [34] extended the distance measures for IFSs, viz, Hamming, Euclidean, normalized Hamming, and normalized Euclidean distances, and similarities to PFSs and applied them to multi-criteria decision-making problems and multi-attribute decision-making problems. Taking advantage of the Jensen–Shannon divergence, a new divergence measure for PFSs was introduced in [21]. This measure was further established as a superior tool than the existing ones and was used to solve medical diagnosis problems. Generalized Pythagorean fuzzy normalized Hamming distance, the generalized Pythagorean fuzzy normalized Hausdorff distance and the generalized hybrid Pythagorean fuzzy normalized distance were introduced in [27]. Peng [35] came up with new Pythagorean distance and similarity measures [36] to overcome the deficiencies of existing such measures. In [9], the author modified the distance function introduced by Zhang & Xu in [37] by normalizing it and validated the axiomatic definition of a metric for modified version which was missing in [37].

Summarizing the existing literature, we observe that most of the distance functions for PFSs are introduced as a generalization of IFS counterparts. Few among them are not even normalized and thereby were reintroduced with suitable normalization. On the other hand, the complex mathematical form of some distance measures hinders their applicability and popularity as a handy mathematical tool. In addition to that, none among the existing measures could distinctly distinguish highly uncertain PFSs, i.e. PFSs with very small values of membership and non-membership grades. One encounters such PFSs when minimal knowledge or information is available about a system. So, there is a need for a suitable distance measure to deal with such PFSs. To fill the above-stated lacunae, we introduce a new distance function with simple mathematical form for PFSs which can effectively handle highly uncertain PFSs and is equally efficient in other cases also. Therefore, the main contributions of this work are

  • introduction of a new distance measure for PFSs.

  • construction of simple mathematical form of it to avoid calculation complicacy.

  • ability of handling effectively PFSs with higher degree of uncertainty and equal efficiency in other cases also.

  • applicability of the introduced definition in the fields of multi-attribute decision making (MADM), pattern recognition & medical diagnosis.

The paper is organized as follows. Section 2 recalls the mathematical background of related concepts. The new definition of distance measure for PFSs is proposed and some properties are studied in Sect. 3. Different characteristics of the proposed measure are verified via various numerical examples in Sect. 4. Applicability of the prescribed distance function in various fields is elaborated in Sect. 5 followed by the conclusions in the last section.

2 Mathematical background

In this section, we quickly recall some of the basic definitions and results required for our further studies.

Definition 2.1

[2] Let X be a finite universe of discourse. An IFS K in X is defined as \(K = \{\langle \mu _K(x), \nu _K(x)\rangle | x \in X\}, \, \mu _K:X \rightarrow [0, 1]\) and \( \nu _K: X \rightarrow [0, 1]\) are, respectively, the membership and non-membership functions, such that \(0 \le \mu _K(x) + \nu _K(x) \le 1, \, \forall x \in X\). The degree of hesitancy is given by \(\pi _K(x) = 1 - \mu _K(x) - \nu _K(x)\).

Definition 2.2

[6, 13] Let X be a finite universe of discourse. A Pythagorean fuzzy set (PFS) K in X is defined as \(K = \{\langle K_Y(x), K_N(x)\rangle \,|\, x \in X\}, \, K_Y:X \rightarrow [0, 1]\) and \( K_N: X \rightarrow [0, 1]\), respectively, indicate the guaranteed membership and non-membership functions, such that \(0 \le K_Y^2(x) + K^2_N(x) \le 1, \, \forall x \in X\). The degree of hesitancy is given by \(K_H(x)={\rm Hes}(x) = \sqrt{1 - K_Y^2(x) - K_N^2(x)}\).

It is clear from the definitions that the admissible area under Pythagorean membership grades is greater than that of intuitionistic membership grades.

Unless mentioned otherwise, from now onward any PFS, say \(K = \{\langle K_Y(x), K_N(x)\rangle \,|\, x \in X\}\) is denoted simply as K for brevity. The guaranteed membership \(K_Y\), non-membership \(K_N\) and the degree of hesitancy \(K_H\) are always associated with the PFS K in this notation.

2.1 Operations

Let \(\mathcal {PFS}(X)\) denote the class of all Pythagorean fuzzy sets defined on the finite universe of discourse X. The following operations [6, 13] are valid for all members \(K, L \in \mathcal {PFS}(X)\),

$$ \begin{aligned}&K=L \Leftrightarrow K_Y(x) = L_Y(x)\, \& \, K_N(x) = L_N(x) \,\forall x \in X, \end{aligned}$$
(1)
$$ \begin{aligned}&K \subseteq L \Leftrightarrow K_Y(x) \le L_Y(x)\, \& \, K_N(x) \ge L_N(x) \,\forall x \in X,\, {\rm we~ say}\, K \subset L \Leftrightarrow K\subseteq L, K \ne L, \end{aligned}$$
(2)
$$\begin{aligned}&{\rm the\,~ complement\,~ of}\, K \,{\rm is\, ~defined~\, as}\, K^c = \{\langle K_N(x), K_Y(x)\rangle \,|\, x \in X\},\end{aligned}$$
(3)
$$\begin{aligned}&K \cup L = \{\langle {\rm max} \left[ K_Y(x), L_Y(x) \right] , {\rm min} \left[ K_N(x), L_N(x) \right] \rangle \,|\, x \in X\}, \end{aligned}$$
(4)
$$\begin{aligned}&K \cap L = \{\langle {\rm min} \left[ K_Y(x), L_Y(x) \right] , {\rm max} \left[ K_N(x), L_N(x) \right] \rangle \,|\, x \in X\}. \end{aligned}$$
(5)

2.2 Distance measures

Below we list some of the popularly used definitions of distance measure for Pythagorean fuzzy sets. The following normalized versions of distance functions between two PFSs KL were introduced in [21].

  1. 1.

    Normalized Hamming distance

    $$\begin{aligned} D_{HmN} (K,L) = \dfrac{1}{2n} \sum \limits _{i=1}^n\left( |K_Y^2(x_i) - L_Y^2(x_i)| + |K_N^2(x_i) - L_N^2(x_i)| +|K_H^2(x_i) - L_H^2(x_i)| \right) ; \end{aligned}$$
    (6)
  2. 2.

    The Euclidean distance

    $$\begin{aligned} D_{EN} (K,L) = \sqrt{\dfrac{1}{2n}\sum \limits _{i=1}^n \left( (K_Y^2(x_i) - L_Y^2(x_i))^2 + (K_N^2(x_i) - L_N^2(x_i))^2+(K_H^2(x_i) - L_H^2(x_i))^2 \right) }; \end{aligned}$$
    (7)
  3. 3.

    The normalized Chen’s distance

    $$\begin{aligned} D_{CN} (K,L) = \left[ \dfrac{1}{2n}\sum \limits _{i=1}^n \left( |K_Y^2(x) - L_Y^2(x)|^\beta + |K_N^2(x) - L_N^2(x)|^\beta +|K_H^2(x) - L_H^2(x)|^\beta \right) \right] ^{\frac{1}{\beta }}; \end{aligned}$$
    (8)

    where \(\beta \ge 1\) is a distance parameter. For \(\beta =1\) we get the Hamming distance and \(\beta =2\) gives the Euclidean distance.

  4. 4.

    The normalized PFSJS distance

    $$\begin{aligned} D_{FW}(K,L) = \dfrac{1}{n}\sqrt{\dfrac{1}{2} \left[ {\sum _{p} }K^2_p(x_i) \log \left( \dfrac{2K^2_p(x_i)}{K^2_p(x_i)+L^2_p(x_i)}\right) + {\sum _{p} } L^2_p(x_i) \log \left( \dfrac{2L^2_p(x_i)}{K^2_p(x_i)+L^2_p(x_i)}\right) \right] }, \end{aligned}$$
    (9)

    where \(K^2_H(x)= 1- K^2_Y(x)- K^2_N(x)\) and \(p \in \{Y,N,H\}\).

  5. 5.

    Normalized Hausdorff distance [38]

    $$\begin{aligned} D_{HN} (K,L) = \dfrac{1}{n} \sum \limits _{i=1}^n {\rm max}\left[ |K_Y^2(x_i)-L_Y^2(x_i)|,|K_N^2(x_i)-L_N^2(x_i)|\right] ; \end{aligned}$$
    (10)
  6. 6.

    Zhang and Xu distance [37]

    $$\begin{aligned} D_{ZX} (K,L) = \dfrac{1}{2} \sum \limits _{i=1}^n\left( |K_Y^2(x_i) - L_Y^2(x_i)| + |K_N^2(x_i) - L_N^2(x_i)| +|K_H^2(x_i) - L_H^2(x_i)| \right) ; \end{aligned}$$
    (11)
  7. 7.

    Modified Zhang and Xu distance [9]

    $$\begin{aligned} D_{MZX} (K,L) = \dfrac{1}{n} \,D_{ZX} (K,L), \end{aligned}$$
    (12)

    where \(D_{ZX} (K,L)\) denotes Zhang and Xu distance as mentioned in Eq. (11).

This list is by no means exhaustive and few of the popular distance measures are mentioned above. We have compared our results with those obtained through some of the aforesaid distances.

3 Proposed definition with mathematical properties

Here we introduce a novel distance measure for PFSs and study some of its mathematical properties.

Definition 3.1

Given a finite universe \(X = \{x_i\,|\, i= 1, 2, \ldots , n\}\), distance between any two PFSs \( K \, \& \, L\) is defined as

$$\begin{aligned} D(K, L)= \frac{1}{n} \sum _{i=1}^n \frac{|K_Y^2(x_i) - L_Y^2(x_i)| + |K_N^2(x_i) - L_N^2(x_i)| }{K_Y^2(x_i) + L_Y^2(x_i) + K_N^2(x_i) + L_N^2(x_i) }, \end{aligned}$$
(13)

where \( K = \{\langle K_Y(x_i), K_N(x_i)\rangle \,|\, x_i \in X\}\) and \(L = \{\langle L_Y(x_i), L_N(x_i)\rangle \,|\, x_i \in X\}\).

Then we have the following theorem.

Theorem 3.2

Let \(K_1, K_2\) and \(K_3\) be three PFS defined in the universe of discourse X. The introduced distance measure D satisfies the following properties:

  1. (i)

    Boundedness: \(0 \le D(K_1, K_2) \le 1, \, \forall K_i \);

  2. (ii)

    Reflexivity: \(D(K_1, K_1) = 0 \, \forall K_i \) ;

  3. (iii)

    Symmetry: \(D(K_1, K_2) = D(K_2, K_1)\);

  4. (iv)

    Separability: \(D(K_1, K_2) = 0 \) iff \(K_1= K_2\);

A distance function satisfying the conditions (ii), (iii) & (iv) of Theorem 3.2 is known as semi-metric in the literature [39].

Proof

  1. (i)

    It is evident that, \({\rm max}\, \left[ \dfrac{|K_Y^2(x_i) - L_Y^2(x_i)| + |K_N^2(x_i) - L_N^2(x_i)| }{K_Y^2(x_i) + L_Y^2(x_i) + K_N^2(x_i) + L_N^2(x_i)}\right] =1\) for each \(x_i\)

    \(\Rightarrow \dfrac{1}{n} \sum \limits _{i=1}^n \dfrac{|K_Y^2(x_i) - L_Y^2(x_i)| + |K_N^2(x_i) - L_N^2(x_i)| }{K_Y^2(x_i) + L_Y^2(x_i) + K_N^2(x_i) + L_N^2(x_i) } \le 1\).

    We have \(K_Y(x_i),\, L_Y(x_i) \in [0,1] \,\, \forall x_i \in X \). So D(KL) is a nonnegative quantity and hence we conclude \(0 \le D(K, L) \le 1\).

  2. (ii)

    Obvious implication from equality of two PFSs.

  3. (iii)

    Follows from the definition of distance function.

  4. (iv)

    Using reflexivity property, we are left to show \(D(K_1, K_2) = 0 \Rightarrow K_1 = K_2\).

    $$ \begin{aligned} {\rm Now} \,&D(K_1, K_2) = 0 \\ \Rightarrow&\frac{1}{n} \sum _{i=1}^n \frac{|K_{1Y}^2(x_i) - K_{2Y}^2(x_i)| + |K_{1N}^2(x_i) - K_{2N}^2(x_i)| }{K_{1Y}^2(x_i) + K_{2Y}^2(x_i) + K_{1N}^2(x_i) + K_{2N}^2(x_i) }=0\\ \Rightarrow&|K_{1Y}^2(x_i) - K_{2Y}^2(x_i)| + |K_{1N}^2(x_i) - K_{2N}^2(x_i)| = 0\\ \Rightarrow&K_{1Y}^2(x_i) - K_{2Y}^2(x_i) =0 \, \& \, K_{1N}^2(x_i) - K_{2N}^2(x_i) =0\\ \Rightarrow&K_{1Y}(x_i) = K_{2Y}(x_i) \, \& \, K_{1N}(x_i) = K_{2N}(x_i) \quad \forall x_i\\ \Rightarrow&K_1 = K_2. \end{aligned}$$

4 Numerical examples and characteristics

4.1 Boundedness and nonlinearity

\(\diamond \) :

Consider two PFSs K and L defined on \(X= \{x\}\), where \(K=\langle a, b \rangle ,\, L= \langle b,a \rangle \). As K and L are PFSs, we must have \( a^2 + b^2 \le 1\). Distance between such K and L using Eq. (13), is portrayed in Fig. 1 (left) and it is verified that, the distance is 0 iff \(a = b \), i.e. when \(A=B\). We adhere to the fact that, the PFS L is the complement of K and a mirror reflection about the \(45^{\circ }\) line will take PFS L to PFS K. This geometrical interpretation infers that, the PFSs lying on \(a=b\) line are self-complement and thereby result in null distances as evident from Fig. 1 (left). It further verifies that the proposed distance function is bounded between 0 and 1. Additionally, for a selected value of b, say 0.2, Fig. 1 (right) portrays the nonlinear behaviour of the distance function. In a similar manner, for a given value of a, the proposed distance yields similar characteristics against b.

\(\diamond \) :

Let us take a PFS, \(K=\langle a,b \rangle \) defined on \(X= \{x\}\). Being the membership & non-membership values of a PFS, \(a,\, b \in [0,1]\) and \(a^2+b^2 \le 1\). Considering \(a^2\), \(b^2\) as the membership & non-membership grades, respectively, we construct another PFS, say \(L = \langle a^2, b^2 \rangle \). Now, we calculate the distance between PFSs K & L using Eq. (13) and plot its variation in Fig. 2, for all allowed values of a and b. Boundedness and nonlinearity of the proposed distance are apparent here.

\(\diamond \) :

Next, consider the PFSs, \(K = \langle a, b \rangle \) and \(L = \langle \beta , 1-\beta \rangle \) defined on \(X= \{x\}\). For such a choice of L, the distance between them now varies against a,  b and \(\beta \). Without loss of generality, following four choices of K are considered

$$\begin{aligned} (i)\, K = \langle 1, 0 \rangle , \, (ii)\, K = \langle 0, 1 \rangle , \, (iii)\, K = \langle 0.4, 0.8 \rangle \,{\rm and} \, (iv)\, K = \left\langle \frac{1}{\sqrt{2}}, \frac{1}{\sqrt{2}} \right\rangle . \end{aligned}$$

and variations of D(KL) with respect to \(\beta \) are portrayed in Fig. 3. It is evident from the figure that the distance function D is bounded for all allowed values of \(\beta \) and its variation shows nonlinearity in all the cases.

Fig. 1
figure 1

Variation of distance function D(KL) with respect to (left) a and b. (right) a for a fixed \(b=0.2\)

Fig. 2
figure 2

Variation of D(KL) against all feasible values of a & b

Fig. 3
figure 3

Variation of D(KL) with respect to \(\beta \) for some selected K

4.2 Superiority

To demonstrate the superiority of the introduced distance function, a comparison among results obtained by our method and few popular distance functions is executed. For this purpose, let us take three PFSs, KLM defined on \(X = \{x \}\), where \(K = \langle .01, .02 \rangle ,\, L = \langle .02, .02 \rangle ,\, M = \langle .1, .2 \rangle \). As mentioned earlier, the proposed function is sensitive to highly uncertain PFSs, thereby, without loss of generality, aforesaid three PFSs are chosen as exemplary. Then, we calculate the distances D(MK) and D(ML) using Eq. (13) and few other distance functions (using Eqs. (6)–(10)). The results are displayed in Table 1. It is evident that the distance measures \( D_{HmN},\, D_{EN}\, \& \, D_{HN} \) couldn’t differentiate \( K \, \& \,L\) from M . However, \(D_{CN}\) and \(D_{FW}\) could distinguish, but they fall short when the results are rounded off to second decimal places only. On the other hand, the proposed function could clearly identify them as distinct sets even after rounding off to second decimal place. This affirms the superiority of our proposed definition.

Table 1 Comparison of different distance measures

Next example demonstrates that the applicability of the introduced definition is not limited to highly uncertain PFSs only.

4.3 Counter-intuitive case

We consider pairs of PFSs \(A_1=\{ \langle .10, .10 \rangle , \langle .10, .10 \rangle \} = A_2\); \(A_3= \{ \langle .714, .640 \rangle , \langle .640, .557 \rangle \} = A_4\).

\(B_1 =\{ \langle .02, .02 \rangle , \langle .02, .01 \rangle \}\); \(B_2= \{ \langle .01, .01 \rangle , \langle .02, .02 \rangle \}\); \(B_3 = \{ \langle .640, .640 \rangle , \langle .714, .640 \rangle \}\); \(B_4= \{ \langle .781, .557 \rangle , \langle .557, .458 \rangle \}\), such that \(A_1 \ne A_3\) and all \(B_i\)’s are distinct. Distances between such pairs of PFSs are evaluated using the proposed definition (Eq. (13)) & some popular distance measures (using Eqs. (6)–(10)) with a comparison given in Table 2.

Table 2 Comparison of distance measures for different pairs of PFSs A & B

It is evident from Table 2 that, our distance function can clearly distinguish such pairs by comparing the values in columns 1 & 2 and 3 & 4, respectively, while others fail to do so. Results obtained via the proposed distance are marked in bold faces in the table.

5 Applications of the proposed definition

5.1 Multi-attribute decision-making (MADM) problems

In an MADM problem, one tries to make the best possible decision out of finite alternatives, by taking into account a collection of clashing attributes with preferred weights. Here all the alternatives against each attribute are represented by PFSs and the concept of similarity measure is used to obtain a decision. Summarizing the scores of evaluators for each attribute concerning every alternative the representation is obtained. Weights for the attributes have to be determined a priori. A decision is achieved by implementing the following algorithm.

  1. Step 1:

    Write the PFS representation of the alternatives, say, \(M_i = \{\langle M^{(i)}_Y(x_j), M^{(i)}_N(x_j)\rangle | x_j \in X \}\), popularly known as decision matrix. Here, \( M^{(i)}_Y\) and \(M^{(i)}_N\) represent the guaranteed membership and non-membership grades, respectively, for the \(i^{\rm th}\) alternative, \(M_i\).

  2. Step 2:

    Construct positive ideal PFS using Eq. (4) from the alternatives as

    $$\begin{aligned} M^+ = \underset{i}{\bigcup } M_i. \end{aligned}$$
    (14)
  3. Step 3:

    Similarly construct negative ideal PFS using Eq. (5) as

    $$\begin{aligned} M^- =\underset{i}{\bigcap } M_i. \end{aligned}$$
    (15)
  4. Step 4:

    Calculate the distances \(D(M_i, M^+)\) and \(D(M_i, M^-)\). For a given weight vector \(w = \{w_i, 1 \le i \le n\}\) the distance measure should be interpreted as the weighted distance given by

    $$\begin{aligned} D(K, L)= \frac{1}{n} \sum _{i=1}^n w_i \frac{|K_Y^2(x_i) - L_Y^2(x_i)| + |K_N^2(x_i) - L_N^2(x_i)| }{K_Y^2(x_i) + L_Y^2(x_i) + K_N^2(x_i) + L_N^2(x_i) }. \end{aligned}$$
    (16)
  5. Step 5:

    Measure the similarity S, using the distance calculated in the step 4. Here various functional forms of similarity measure can be explored.

  6. Step 6:

    Evaluate relative similarity \(S_r\) for \(M_i\) through the formula

    $$\begin{aligned} S_r(M_i) = \dfrac{S(D_+^{(i)}) }{S(D_+^{(i)} )+S(D_-^{(i)})}, \end{aligned}$$
    (17)

    \({\rm where} \, D_+^{(i)} =D(M_i,M^+) \,{\rm and} \, D_-^{(i)} =D(M_i,M^-)\) are estimated using Eq. (16).

  7. Step 7:

    Rank the alternatives \(M_i\) in ascending order of relative similarity values.

5.1.1 Face mask selection

In the COVID-19 pandemic, wearing a face-mask has become an integral part of the life. There are plenty of face-masks available in the market, few among them are disposable medical masks (\(M_1\)), medical surgical masks (\(M_2\)), particulate respirators (N95) (\(M_3\)), ordinary non-medical masks (\(M_4\)), medical protective masks (\(M_5\)) and gas masks (\(M_6\)). The set of clashing attributes for each mask type are leakage rate (\(A_1\)), reutilizability (\(A_2\)), quality of raw material (\(A_3\)) and filtration capability (\(A_4\)). To buy a face-mask a person will put his/her preferred weight on the four attributes which leads to an MADM problem. The PFS representation of each attribute for every mask type is given in Table 3 (for more details please refer to [40]).

Table 3 PFS representation of different mask types

From Table 3, positive and negative ideal PFSs are calculated using Eqs. (14) & (15), respectively, as

$$\begin{aligned}&M^{+}=\{\langle 0.73,0.25\rangle ,\langle 0.80,0.12\rangle ,\langle 0.73,0.31\rangle ,\langle 0.73, 0.42\rangle \};\\&M^{-}=\{\langle 0.10,0.64\rangle ,\langle 0.32,0.66\rangle ,\langle 0.27,0.68\rangle ,\langle 0.23,0.68\rangle \}. \end{aligned}$$

By putting equal weight on each attribute, i.e. \(w =\{.25, .25, .25, .25\}\), we calculate the distances between \(M_i\) and \(M^{+}/M^{-}\) which are listed in Table 4. Using the distance calculated, measure the similarity between \(M_i\) and \(M^{+}/M^{-}\) via the operator \(S(D) = 1-D\). Further the relative similarity \(S_r(M_i)\) is obtained using Eq. (17) and results are displayed in Table 4.

Table 4 Distances of \(M_i\) from \(M^{+}/M^{-}\) and relative similarity measures

Finally, using Step 7, the ranking of face-mask is obtained as follows

$$\begin{aligned} M_1 \succ M_5 \succ M_3 \succ M_6 \succ M_4 \succ M_2. \end{aligned}$$

Therefore, the best choice for the face-mask turned out to be \(M_1\), i.e. disposable medical masks which is consistent with the existing result [40]. A buyer may put more stress on a particular attribute compared to rest, say the attribute - filtration capability (\(A_4\)) to select a face-mask. So, the weight vector modifies to, \(w =\{0.2, 0.1,0.1,0.6\}\) for the attributes, \(A = \{A_1,A_2, A_3, A_4\}\). Following the steps of the algorithm, the best choice of face-mask is \(M_3\) i.e. particulate respirators (N95) mask. Above discussion demonstrates the effect of individual preference on attributes in the outcome of decision-making process.

5.1.2 School selection [38]

Nowadays in spite of having good public sector schools, the trend among parents is to send their children to private sector. Of late private sector schools just spring up like mushrooms. So selecting a suitable private school has become a daunting task for parents. Let us consider a collection of five private sector schools, designated as \(S_i\) where \(i = 1, 2, \ldots 5\), in a locality. Considering educationalists and experts opinion the following attributes of the schools are found to be the most influencing and significant [38]: personalized attention (\(A_1\)), academic accolades (\(A_2\)), skilful and up-to-date teachers (\(A_3\)), parents and community engagement (\(A_4\)) and technical facilities (\(A_5\)). The PFS representation of each school against the attributes is detailed in Table 5.

Table 5 PFS representation of different schools

Following the algorithm, we obtain

$$\begin{aligned}&S^+= \{\langle 0.8, 0.4\rangle , \langle 0.9, 0.2\rangle , \langle 0.9, 0.1\rangle , \langle 0.7, 0.3\rangle , \langle 0.8, 0.2\rangle \} \\&S^-= \{\langle 0.6, 0.6\rangle , \langle 0.7, 0.5\rangle , \langle 0.6, 0.5\rangle , \langle 0.4, 0.6\rangle , \langle 0.5, 0.6\rangle \} \end{aligned}$$

Choices of weightage to the attributes are subjective. Here, we consider two distinct weight vectors, say \(w_1=\{0.18, 0.32, 0.23, 0.13, 0.14\}\), a case of unequal weightage and \(w_2 = \{0.2,0.2,0.2,0.2,0.2\}\), an example of equal emphasis to each attribute. Using \( w_1 \, \& \,w_2\), we evaluate the weighted distance between \(S_i\) and \(S^{+}/S^{-}\) using Eq. (16). These weighted distances are further used to obtain similarity measures and subsequently relative similarity measures. We have employed three distinct functional forms of similarity measure (S(D)) as listed in Table 6.

Table 6 Relative similarity measures corresponding to \( w_1 \, \& \,w_2\)

Higher the value of relative similarity, better is the rank. On that note,

\(\diamond \):

ranking corresponding to \(w_1\) is \(S_4 \succ S_2 \succ S_5 \succ S_1 \succ S_3\). The best choice of the private sector school turns out to be \(S_4\) which is matching with the existing literature [38].

\(\diamond \):

Ranking corresponding to \(w_2\) is \(S_4 \succ S_3 \succ S_5 \succ S_2 \succ S_1\). Interestingly, here also the best option is found to be \(S_4\); however positions of the rest have altered.

It is to be observed that, different functional forms of similarity measures have provided same ranking irrespective of the weight vector.

5.2 Pattern recognition

A pattern recognition problem is addressed using the proposed distance measure. The problem is consisting of a test sample \(\mathcal {S} = \{ \langle 0.7,\, 0.7\rangle , \langle 0.3,\, 0.8\rangle , \langle 0.7, \,0.2\rangle \}\), which is to be classified as one among three given patterns \(\mathcal {P}_1 = \{\langle 0.5,\, 0.6\rangle , \langle 0.7,\, 0.3\rangle , \langle 0.8, \,0.5\rangle \},\, \mathcal {P}_2 = \{\langle 0.6,\, 0.6\rangle , \langle 0.4,\, 0.8\rangle , \langle 0.6, \,0.1\rangle \}\) and \(\mathcal {P}_3 = \{ \langle 0.6,\, 0.7\rangle , \langle 0.4,\, 0.9\rangle , \langle 0.6, \,0.4\rangle \}\). The steps to be followed in the process are:

\(\diamond \):

Step 1: Obtain the PFS representations of the patterns as well as the sample.

\(\diamond \):

Step 2: Calculate the distance of \(\mathcal {S}\) from each of \(\mathcal {P}_1, \mathcal {P}_2 \) and \(\mathcal {P}_3\) via Eq. (13).

\(\diamond \):

Step 3: Classify the test sample as the pattern from which minimum value is obtained in step 2.

\(\diamond \):

Step 4: Calculate degree of confidence (DoC) [41] using the formula :

$$\begin{aligned} DoC = \sum \limits _{j=1, j \ne j_0}^n \left| D(\mathcal {P}_{j},\mathcal {S}) - D (\mathcal {P}_{j_0},\mathcal {S})\right| , \end{aligned}$$
(18)

where \(\mathcal {P}_{j_0}\) is the classified pattern for the test sample \(\mathcal {S}\). Greater value of DoC guarantees the higher confidence level of the prediction. It is to be noted that, the process can be readily applied for a problem with n numbers of patterns.

Figure 4 includes the distances of \(\mathcal {S}\) from the three given patterns using our proposed definition and DoC against this prediction. From Fig. 4, it is evident that the test sample \(\mathcal {S}\) has minimum distance from pattern \(\mathcal {P}_2\) and hence is classified as \(\mathcal {P}_2\). For comparison step 2 is repeated using \( D_{HmN}, D_{EN}, D_{CN}, D_{HN} \, \& \, D_{FW}\) and the results are displayed in Fig. 4. Except \(D_{CN}\), others have predicted the test sample as pattern \(\mathcal {P}_2\), whereas the distance formula \(D_{CN} \) with \(\beta =4\) has classified it as \(\mathcal {P}_3\). Our result coincides with the conclusions drawn through majority of the distance measures. Besides that, the proposed distance has the highest value of DoC which assures the confidence level of the classification.

Fig. 4
figure 4

PFS distances of a test sample from patterns

5.3 Medical diagnosis [24, 42,43,44,45,46,47,48,49,50]

Suppose there are four patients P= \(\{\mathrm {Ali, Baruah, Chatterjee, Deka}\}\) having the symptoms \(S =\) {\(S_1\): Temperature, \(S_2\): Headache, \(S_3\): Stomach pain, \(S_4\): Cough, \(S_5\): Chest pain\(\}\). The symptoms of the patients represented by PFS are listed in Table 7. The set of possible diagnosis \(D = \{ D_1:\) Viral fever, \(D_2:\) Malaria, \(D_3:\) Typhoid, \(D_4:\) Stomach problem, \(D_5:\) Chest problem \(\}\) is represented in Table 8.

Table 7 PFS representation of symptoms
Table 8 Symptom characteristics for diagnosis

Now we calculate the distance between S and D for each of the patients using the proposed definition (Eq. (13)) and have obtained the results as displayed in Fig. 5. In Table 9 we provide a comparison of our prediction with that of few existing results obtained via either IFS- or PFS-based distance measures and thereby confirming our predictions with the majority of existing ones.

Fig. 5
figure 5

PFS distances of each symptom from all possible diagnosis

Table 9 Comparison of prediction

6 Conclusions

Here, we have introduced a distance measure for PFSs whose simplicity in the mathematical form is an extra boon to it. The salient feature of the distance measure is its efficacy in distinguishing PFSs with high hesitancy. Supremacy of the distance function is established via suitable numerical examples and reasonability is tested by beating the counter-intuitive cases. Nonlinearity & bounded characteristics are proved and verified through proper examples. Applicability of the distance measure is established in various fields like pattern recognition, medical diagnosis and multi-attribute decision-making problems. We believe, the proposed distance function will find its serviceability in new avenues of application.