1 Introduction

Since the theory of fuzzy sets (FSs) is introduced by Zadeh [1], it has been successfully used in various fields. Later, several extension forms are proposed such as interval-valued fuzzy sets [2], Atanassov’s intuitionistic fuzzy sets (IFSs) [3] and vague sets [4]. In 1996, Bustince and Burillo [5] showed the notion of vague sets coincides with that of IFSs. As an extension of FSs, IFSs are characterized by a membership degree, a non-membership degree and a hesitancy degree. So vagueness in real applications [613].

As two important information measures in the theory of fuzzy sets, entropy and similarity measure of IFSs have been widely investigated by many researchers from different point of views. Burillo and Bustince [14] introduced the notion of entropy of IFSs to measure the degree of intuitionism of an IFS. Bustince et al. [15] presented fuzzy entropy of IFSs by using the fuzzy implication operators. Szmidt and Kacprzyk [16] extended the axioms given by De Luca and Termini [17] and proposed another axiomatic definition for the entropy of IFSs. Szmidt and Kacprzyk [16], Wang and Lei [18] and Huang and Liu [19], respectively, gave an entropy of IFSs from different point of views, which are shown equivalently by Wei et al. [20]. On the other hand, the similarity measures of IFSs are also studied by many researchers [10, 2125], whilst the application of similarity measures in digital image processing is considered in the literature [26, 27].

However, all the above researches are based on the assumption that the elements in a set are independent, and each intuitionistic fuzzy value (IFV) has the same importance. However, in many practical situations, the elements in a set are usually correlative [28, 29]. The fuzzy measure, introduced by Sugeno [30], has been shown a very effective tool for modeling the correlative characteristics among elements [3135], and has been successfully used to deal with decision problems [3641]. As far as we know, however, there is less investigation on entropy and similarity measure of Atanassov’s intuitionistic fuzzy sets by using fuzzy measures. In the study we first introduce an entropy of IFSs, and give a construction method to obtain the similarity measure of IFSs. To overall reflect the interactive characteristics between features, we further define three Shapley-weighted similarity measures. If the information about the weights of features is partly known, models for the optimal fuzzy measure on feature set are constructed. Since the fuzzy measure is defined on the power set, it makes the problem exponentially complex. Thus, it is not easy to obtain the fuzzy measure of each combination in a set when it is large. The \(\lambda\)-fuzzy measure proposed by Sugeno [30] seems to well deal with this issue, which only needs n variables to determine a \(\lambda\)-fuzzy measure on a set with n elements. For this reason, we further research Atanassov’s intuitionistic fuzzy pattern recognition problems under \(\lambda\)-fuzzy measures.

This paper is organized as follows: in Sect. 2, we review some basic concepts about IFSs, entropy and similarity measure. In Sect. 3, we propose a new entropy of IFSs, and present a construction method to obtain the similarity measure of IFSs by using entropy. Based on the Shapley function with respect to fuzzy measures, we propose three Shapley-weighted similarity measures, which reflect the interactive between elements. In Sect. 4, models for the optimal fuzzy measure on feature set are established, and an approach to pattern recognition under Atanassov’s intuitionistic fuzzy environment is developed. To simplify the complexity of solving a fuzzy measure, we further study Atanassov’s intuitionistic fuzzy pattern recognition under \(\lambda\)-fuzzy measures. Moreover, the corresponding examples are given to illustrate the developed procedure. In Sect. 5, the conclusions are made.

2 Preliminaries

2.1 Some basic concepts

By extending Zadeh’s fuzzy sets, Atanassov [3] introduced the concept of Atanassov’s intuitionistic fuzzy sets (IFSs) as follows:

Definition 1

[3] Let \(X\) be a no empty finite set. An IFS A in X is expressed as

$$A = \left\{ {\left\langle {x,u_{A} (x),v_{A} (x)} \right\rangle |x \in X} \right\},$$

where \(u_{A} (x) \in [0,1]\) and \(v_{A} (x) \in [0,1]\), respectively, denote the degrees of membership and non-membership of element \(x \in X\) with the condition \(u_{A} (x) + v_{A} (x) \le 1\). The hesitancy degree is denoted by \(\pi_{A} (x) = 1 - u_{A} (x) - v_{A} (x)\).

When \(u_{A} (x) = 1 - v_{A} (x)\) for each \(x \in X\), we get a fuzzy set, expressed by \(A = \left\{ {\left\langle {x,[u_{A} (x),1 - u_{A} (x),]} \right\rangle |x \in X} \right\}\). The set of all IFSs in X is denoted by IFS(X).

Definition 2

Let \(A = \left\{ {\left\langle {x,u_{A} (x),v_{A} (x)} \right\rangle |x \in X} \right\}\) and \(B =\) \(\left. {\left\{ {\left\langle {x,u_{B} (x),v_{B} (x)} \right\rangle |} \right.x \in X} \right\}\) be two IFSs in X, then

  1. (1)

    \(A \subseteq B\) if and only if \(u_{A} (x) \le u_{B} (x),v_{A} (x) \ge v_{B} (x)\),

  2. (2)

    \(A = B\) if and only if \(A \subseteq B\) and \(A \supseteq B\),

  3. (3)

    \(A^{C} = \left\{ {\left\langle {x,v_{A} (x),u_{A} (x)} \right\rangle |x \in X} \right\}.\)

Definition 3

[16] A real-valued function E: IFS(X) \(\to\) [0,1] is called an entropy measure of IFSs, if it satisfies the following axiomatic requirements:

(E1) E(A) = 0 if and only if A is a crisp set;

(E2) E(A) = 1 if and only if \(u_{A} (x) = v_{A} (x)\) for each \(x \in X\);

(E3) E(A) = E(A C);

(E4) E(A) ≤ E(B) if \(A \subseteq B\) with \(u_{B} (x) \le v_{B} (x)\) for each \(x \in X\), or \(A \supseteq B\) with \(u_{B} (x) \ge v_{B} (x)\) for each \(x \in X\).

Definition 4

[23] A real-valued function S: IFS(X) \(\times\) IFS(X) \(\to\) [0,1] is called a similarity measure of IFS(X), if it satisfies the following conditions:

(S1) 0 ≤ S(A,B) ≤ 1;

(S2) S(A,B) = 1 if and only if A = B;

(S3) S(A,B) = S(B,A);

(S4) If \(A \subseteq B \subseteq C\), then \(S(A,C) \le S(A,B)\) and \(S(A,C) \le S(B,C)\).

In the rest parts, without special explanation, we always assume that the universe X is a finite set, denoted by {x 1, x 2,…, x n }.

2.2 Several entropy of IFSs

Entropy, as an information measure, plays an important role in uncertain theory. Burillo and Bustince [5] defined the distance measure between Atanassov’s intuitionistic fuzzy sets and gave an axiomatic definition of Atanassov’s intuitionistic fuzzy entropy and a theorem which characterizes it. Furthermore, Burillo and Bustince [5] proposed the entropy that measures the distance from the considered set to IFSs rather than crisp sets, denoted by.

$$E_{BB} (A) = \sum\limits_{i = 1}^{n} {\left( {1 - \varPhi (u_{A} (x),v_{A} (x))} \right)} ,$$
(1)

where the function \(\varPhi\):[0,1] × [0,1] → [0,1] satisfies the properties (1) \(\varPhi (x,y) = 1\) iff x + y = 1, and \(\varPhi (x,y) = 0\) iff x = y = 0; (2) \(\varPhi (x,y) = \varPhi (y,x)\); (3) \(\varPhi (x,y) \le \varPhi (x',y')\) for \(x \le x',y \le y'\).

Burillo and Bustince [5] gave us a good example of how to define an entropy measure from a theoretical point of view. The special cases given by Burillo and Bustince [5] did not consider hesitancy information.

Later, Szmidt and Kacprzyk [16] introduced an entropy measure of IFSs, which is based on the biggest cardinality (max-sigma-count) of IFSs, denoted by.

$$E_{SK} (A) = \frac{1}{n}\sum\limits_{i = 1}^{n} {\frac{{\hbox{max} {\kern 1pt} {\kern 1pt} {\text{Count}}(A_{i} \cap A_{i}^{C} )}}{{\hbox{max} {\kern 1pt} {\kern 1pt} {\text{Count}}(A_{i} \cup A_{i}^{C} )}}} ,$$
(2)

where for each i, A i denotes the single-element IFS corresponding to the element \(x_{i} \in X\), described as \(A_{i} = \left\langle {x_{i} ,u_{A} (x_{i} ),v_{A} (x_{i} )} \right\rangle\), and \(\hbox{max} {\kern 1pt} {\kern 1pt} {\text{Count}}(A_{i} \cap A_{i}^{C} ) = \left\langle {x_{i} ,} \right.\hbox{min} \left. {\left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\},\hbox{max} \left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\}} \right\rangle\), \(\hbox{max} {\text{Count}}(A_{i} \cup A_{i}^{C} )\)

$$= \left\langle {x_{i} ,\hbox{max} \left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\},\hbox{min} \left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\}} \right\rangle .$$

Later, Huang and Liu [19] proposed an entropy measure of vague sets. Since the vague sets are IFSs [5]. When we apply it in the setting of IFSs, for any \(A \in\) IFS(X), it can be transformed as follows [20]:

$$E_{HL} (A) = \frac{1}{n}\sum\limits_{i = 1}^{n} {\frac{{1 - |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} )}}{{1 + |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} )}}} .$$
(3)

Recently, Wang and Lei [18] gave another entropy measure of IFSs, for any \(A \in\) IFS(X), defined by.

$$E_{WL} (A) = \frac{1}{n}\sum\limits_{i = 1}^{n} {\frac{{\hbox{min} \left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\} + \pi_{A} (x_{i} )}}{{\hbox{max} \left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\} + \pi_{A} (x_{i} )}}} .$$
(4)

Wei et al. [20] pointed out that the entropy measures defined by Szmidt and Kacprzyk [16], Wang and Lei [18] and Huang and Liu [19] are equivalent, which can be seen as the arithmetic mean of the ratio of each corresponding item.

Different from the above-mentioned entropy, Li et al. [42] gave another entropy measure of IFSs, expressed by:

$$E_{L} (A) = \frac{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \wedge v_{A} (x_{i} )} \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \vee v_{A} (x_{i} )} \right)} }}.$$
(5)

E L is the ratio of the sum of different corresponding items, but it does not consider the hesitancy information.

2.3 Similarity measures of IFSs

Similarity measure, as another important information measure, is applied to denote the similarity degree of fuzzy sets and has received considerable attention. Li and Cheng [23] introduced a concept of similarity measure of IFSs and applied it to pattern recognition. Later, Liang and Shi [24] pointed out that Li and Cheng’s method is not always reasonable in some examples, and gave an improved entropy of IFSs, defined by:

$$S_{LS} (A,B) = {1 - \root{p}\of{{\frac{{\sum\nolimits_{i = 1}^{n} {(f_{{u_{AB} }} (i) + f_{{v_{AB} }} (i))^{p} } }}{n}}}},$$
(6)

where \(f_{{v_{AB} }} (i) = |(1 - v_{A} (x_{i} ))/2 - (1 - v_{B} (x_{i} ))/2|\), \(f_{{u_{AB} }} (i) = {{|u_{A} (x_{i} ) - u_{B} (x_{i} )|} \mathord{\left/ {\vphantom {{|u_{A} (x_{i} ) - u_{B} (x_{i} )|} 2}} \right. \kern-0pt} 2}\), and A,B ∈ IFS(X).

Mitchell [10] adopted a statistical approach and interpreted IFSs as ensembles of ordered fuzzy sets to modify Li and Cheng’s similarity measure. Let \(S_{u} (A,B)\) and \(S_{v} (A,B)\) denote the similarity measures between the “low” membership functions u A and u B and between the “high” membership functions 1 − v A and 1 − v B , respectively, as follows:

$$S_{u} (A,B) = 1 - \root{p}\of{{\frac{{\sum\nolimits_{i = 1}^{n} {|u_{A} (x_{i} ) - u_{B} (x_{i} )|^{p} } }}{n}}},$$
$$S_{v} (A,B) = 1 - \root{p}\of{{\frac{{\sum\nolimits_{i = 1}^{n} {|v_{A} (x_{i} ) - v_{B} (x_{i} )|^{p} } }}{n}}}.$$

They then defined the following modified similarity measure of IFSs.

$$S_{M} (A,B) = \frac{1}{2}\left( {S_{u} (A,B) + S_{v} (A,B)} \right)$$
(7)

for IFSs A,B ∈ IFS(X).

Based on the Hausdorff distance, Huang and Yang [43] defined three similarity measures of IFSs, denoted by

$$S_{HY}^{1} (A,B) = 1 - d_{H} (A,B),$$
(8)
$$S_{HY}^{2} (A,B) = \frac{{e^{{ - d_{H} (A,B)}} - e^{ - 1} }}{{1 - e^{ - 1} }},$$
(9)
$$S_{HY}^{3} (A,B) = \frac{{1 - d_{H} (A,B)}}{{1 + d_{H} (A,B)}},$$
(10)

where \(d_{H} (A,B) = \frac{1}{n}\sum\limits_{i = 1}^{n} {\hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }\) and A,B ∈ IFS(X).

3 New entropy and similarity measure of IFSs

Based on analysis above, this section gives a new entropy and similarity measure of IFSs.

3.1 A new entropy of IFSs

Based on the definitions of IFSs and entropy, we introduce the following entropy of IFSs:

$$GE(A) = \frac{{\sum\nolimits_{i = 1}^{n} {f_{1} (u_{A} (x_{i} ),v_{A} (x_{i} ))} }}{{\sum\nolimits_{i = `}^{n} {f_{2} (u_{A} (x_{i} ),v_{A} (x_{i} ))} }},$$

where \(A \in\) IFS(X), the functions f 1:[0,1] × [0,1] → [0,1] and f 2:[0,1] × [0,1] → [0.5,1] satisfy the properties (i) \(f_{1} (x,y) = 0\) and \(f_{2} (x,y) = 1\) iff x = 0, y = 1, or x = 1, y = 0; (ii) \(f_{1} (x,y) = f_{2} (x,y)\) iff x = y, otherwise, \(f_{1} (x,y) < f_{2} (x,y)\); (iii) \(f_{k} (x,y) = f_{k} (y,x)\), k = 1, 2; (iv) \(\frac{{f_{1} (x,y)}}{{f_{2} (x,y)}} \le \frac{{f_{1} (x',y')}}{{f_{2} (x',y')}}\) for \(x \le x' \le y' \le y\), or \(y \le y' \le x' \le x\).

Theorem 1

The mapping E: IFS(X) \(\to\) [0,1], defined by GE, is an entropy of IFSs.

Proof

To prove GE is an entropy of IFSs, it only needs to show that GE satisfies (E1)–(E4) given in Definition 3.

(E1): If A is a crisp set, then we have \(u_{A} (x_{i} ) = 0\), \(v_{A} (x_{i} ) = 1\) or \(u_{A} (x_{i} ) = 1\), \(v_{A} (x_{i} ) = 0\). It gets \(f_{1} (u_{A} (x_{i} ),v_{A} (x_{i} )) = 0\) and \(f_{2} (u_{A} (x_{i} ),v_{A} (x_{i} )) = 1\) for each \(x_{i} \in X\). Thus, GE(A) = 0. On the other hand, if GE (A) = 0, by f 1:[0,1] × [0,1] → [0,1] and f 2:[0,1] × [0,1] → [0.5,1], it has \(f_{1} (u_{A} (x_{i} ),v_{A} (x_{i} )) = 0\) for each \(x_{i} \in X\). Namely, \(u_{A} (x_{i} ) = 0\), \(v_{A} (x_{i} ) = 1\) or \(u_{A} (x_{i} ) = 1\), \(v_{A} (x_{i} ) = 0\). Thus, A is a crisp set.

(E2): When \(u_{A} (x_{i} ) = v_{A} (x_{i} )\) for each \(x_{i} \in X\), we have \(f_{1} (u_{A} (x_{i} ),v_{A} (x_{i} )) = f_{2} (u_{A} (x_{i} ),v_{A} (x_{i} ))\). Then, GE(A) = 1. On the other hand, suppose that GE(A) = 1. Since \(f_{1} (u_{A} (x_{i} ),v_{A} (x_{i} )) \le f_{2} (u_{A} (x_{i} ),v_{A} (x_{i} ))\) for each \(x_{i} \in X\), we get \(f_{1} (u_{A} (x_{i} ),v_{A} (x_{i} )) = f_{2} (u_{A} (x_{i} ),v_{A} (x_{i} ))\). Namely, \(u_{A} (x_{i} ) = v_{A} (x_{i} )\), \(x_{i} \in X\).

(E3): From \(A^{C} = \left\{ {\left\langle {x_{i} ,v_{A} (x_{i} ),u_{A} (x_{i} )} \right\rangle |x_{i} \in X} \right\}\) and \(f_{k} (x,y) =\) \(f_{k} (y,x)\), k = 1, 2, one easily gets GE(A) = GE(A C).

(E4): When \(A \subseteq B\) and \(u_{B} (x_{i} ) \le v_{B} (x_{i} )\) for each \(x_{i} \in X\), we have

$$u_{A} (x_{i} ) \le u_{B} (x_{i} ) \le v_{B} (x_{i} ) \le v_{A} (x_{i} )$$

for each \(x_{i} \in X\).

Thus,

$$\frac{{f_{1} (u_{A} (x_{i} ),v_{A} (x_{i} ))}}{{f_{2} (u_{A} (x_{i} ),v_{A} (x_{i} ))}} \le \frac{{f_{1} (u_{B} (x_{i} ),v_{B} (x_{i} ))}}{{f_{2} (u_{B} (x_{i} ),v_{B} (x_{i} ))}}\,\forall x_{i} \in X.$$

By GE, it gets GE(A) ≤ GE(B).

Similarly, when \(A \supseteq B\) with \(u_{B} (x_{i} ) \ge v_{B} (x_{i} )\) for each \(x_{i} \in X\), one can also prove GE(A) ≤ GE (B).

Next, let us pay more attention to a special case. To get more information on IFSs, combining the entropy given by Szmidt and Kacprzyk [16], Wang and Lei [18], Huang and Liu [19] and Li et al. [42], we define the following entropy of IFSs:

$$E_{N} (A) = \frac{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \wedge v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \vee v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }} ,$$
(11)

where \(A \in\) IFS(X).

In E N , when we delete the hesitancy information of each element, it reduces to be the entropy given by Li et al. [42].

Corollary 1

The mapping E: IFS(X)\(\to\)[0,1], defined by E N , is an entropy of IFSs.

Theorem 2

Let E N : IFS(X) \(\to\) [0,1] be an entropy measure given as ( 11 ), then E N can be equivalently expressed by.

$$E'(A) = \frac{{\sum\nolimits_{i = 1}^{n} {\hbox{max} {\kern 1pt} {\kern 1pt} {\text{Count}}(A_{i} \cap A_{i}^{C} )} }}{{\sum\nolimits_{i = 1}^{n} {\hbox{max} {\kern 1pt} {\kern 1pt} {\text{Count}}(A_{i} \cup A_{i}^{C} )} }}$$
(12)

and

$$E''(A) = \frac{{\sum\nolimits_{i = 1}^{n} {\left( {1 - |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} )} \right)} }}{{\sum\limits_{i = 1}^{n} {\left( {1 + |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} )} \right)} }}$$
(13)

for any \(A \in\) IFS(X).

Proof

We first show (11) can be equivalently expressed by (12). For (12): By.

$$\hbox{max} {\kern 1pt} {\kern 1pt} Count(A_{i} ) = u_{A} (x_{i} ) + \pi_{A} (x_{i} )$$

for any i = 1,2,…,n, it has.

$$E_{N} (A) = \frac{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \wedge v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \vee v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }}$$
$$= \frac{{\sum\nolimits_{i = 1}^{n} {\left( {\hbox{min} \left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\} + \pi_{A} (x_{i} )} \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {\hbox{max} \left\{ {u_{A} (x_{i} ),v_{A} (x_{i} )} \right\} + \pi_{A} (x_{i} )} \right)} }}$$
$$= \frac{{\sum\nolimits_{i = 1}^{n} {\hbox{max} {\text{Count}}(A_{i} \cap A_{i}^{C} )} }}{{\sum\nolimits_{i = 1}^{n1} {\hbox{max} {\text{Count}}(A_{i} \cup A_{i}^{C} )} }}$$

For (13): suppose that \(u_{A} (x_{i} ) \ge v_{A} (x_{i} )\) for some x i , then.

$$1 - |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} ) = 2\left( {v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)$$
$$= 2\left( {u_{A} (x_{i} ) \wedge v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right),$$
$$1 + |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} ) = 2\left( {u_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)$$
$$= 2\left( {u_{A} (x_{i} ) \vee v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right).$$

Similarly, when \(u_{A} (x_{i} ) \le v_{A} (x_{i} )\) for some x i , then.

$$1 - |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} ) = 2\left( {u_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)$$
$$= 2\left( {u_{A} (x_{i} ) \wedge v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right),$$
$$1 + |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} ) = 2\left( {v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)$$
$$= 2\left( {u_{A} (x_{i} ) \vee v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right).$$

From (13), it gets.

$$\begin{aligned} E''(A) &= \frac{{\sum\nolimits_{i = 1}^{n} {\left( {1 - |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} )} \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {1 + |u_{A} (x_{i} ) - v_{A} (x_{i} )| + \pi_{A} (x_{i} )} \right)} }}\\ &= \frac{{\sum\nolimits_{i = 1}^{n} {2\left( {u_{A} (x_{i} ) \wedge v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }}{{\sum\nolimits_{i = 1}^{n} {2\left( {u_{A} (x_{i} ) \vee v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }}\\&= \frac{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \wedge v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {u_{A} (x_{i} ) \vee v_{A} (x_{i} ) + \pi_{A} (x_{i} )} \right)} }} . \end{aligned}$$

That’s \(E_{N} = E{\prime \prime }\).

3.2 A new similarity measure of IFSs

Different from the similarity measures of IFSs introduced in Sect. 2.3, we here give a construction approach to get the similarity measure of IFSs by using entropy.

Let A, B \(\in\) IFS(X). For each \(x_{i} \in X\), define

$$u_{AB} (x_{i} ) = \frac{{1 + \hbox{min} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }}{2},$$
(14)
$$v_{AB} (x_{i} ) = \frac{{1 - \hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }}{2}.$$
(15)

Let

$$G(A,B) = \left\{ {\left\langle {x_{i} ,\left( {u_{AB} (x_{i} ),v_{AB} (x_{i} )} \right)} \right\rangle |x_{i} \in X} \right\}.$$

It is not difficult to get G(A,B) \(\in\) IFS(X).

Theorem 3

Let E be an entropy of IFS(X), then the mapping S: IFS(X) × IFS(X) \(\to\) [0,1], defined by S(A,B) = E(G(A,B)) for each pair of IFSs A and B, is a similarity measure of IFSs.

Proof

(S1): Since \(E(A) \in [0,1]\) for any A \(\in\) IFS(X), and G(A,B) is an IFS in X, we have \(E(G(A,B)) \in [0,1]\).

(S2): From the definition of the entropy measure of IFSs, we have \(E(G(A,B))\) = 1 if and only if \(u_{AB} (x_{i} ) = v_{AB} (x_{i} )\) for each \(x_{i} \in X\). From (14) and (15), we know \(u_{A} (x_{i} ) = u_{B} (x_{i} ) = v_{A} (x_{i} ) = v_{B} (x_{i} ) = 0\) for each \(x_{i} \in X\). Thus, \(A = B\).

(S3): from the construction of G(A,B), it is obvious that G(A,B) = G(B,A). Thus, \(E(G(A,B)) = E(G(B,A))\).

(S4): when \(A \subseteq B \subseteq C\), it has \(u_{A} (x_{i} ) \le u_{B} (x_{i} ) \le u_{C} (x_{i} )\) and \(v_{A} (x_{i} ) \ge v_{B} (x_{i} ) \ge v_{C} (x_{i} )\) for each \(x_{i} \in X\). Namely, \(|u_{A} (x_{i} ) - u_{C} (x_{i} )| \ge |u_{A} (x_{i} ) - u_{B} (x_{i} )|\) and \(|v_{A} (x_{i} ) - v_{C} (x_{i} )| \ge |v_{A} (x_{i} ) - v_{B} (x_{i} )|\) for each \(x_{i} \in X\). Thus, \(G(A,B) \subseteq G(A,C)\).

Since \(u_{AB} (x_{i} ) \ge \frac{1}{2} \ge v_{AB} (x_{i} )\) for each \(x_{i} \in X\), it follows from Definition 3 that \(E(G(A,B)) \ge E(G(A,C))\).

Similarly, one can also prove\(E(G(B,C)) \ge E(G(A,C))\).

Corollary 2

Let E N be an entropy measure of IFSs defined by (11), then the mapping \(S_{N}\), given in Theorem 3, i.e., \(S_{N} (A,B) = E_{N} (G(A,B))\) for each pair of IFSs A and B, is a similarity measure and can be denoted by

$$S_{N} (A,B) = \frac{{\sum\nolimits_{i = 1}^{n} {\left( {1 - \hbox{min} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {1 + \hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}.$$

Corollary 3

Let E L be an entropy measure of IFSs defined by (5), then the mapping \(S_{L}\), given in Theorem 3, i.e., \(S_{L} (A,B) = E_{L} (G(A,B))\) for each pair of IFSs A and B, is a similarity measure and can be denoted by

$$S_{L} (A,B) = \frac{{\sum\nolimits_{i = 1}^{n} {\left( {1 - \hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}{{\sum\nolimits_{i = 1}^{n} {\left( {1 + \hbox{min} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}.$$

Corollary 4

Let E SK be an entropy measure of IFSs defined by (2), then the mapping \(S_{SK}\), given in Theorem 3, i.e., \(S_{SK} (A,B) = E_{SK} (G(A,B))\) for each pair of IFSs A and B, is a similarity measure and can be denoted by

$$S_{SK} (A,B) = \frac{1}{n}\sum\limits_{i = 1}^{n} {\frac{{1 - \hbox{min} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }}{{1 + \hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }}} .$$

From Corollaries 2–4, we know they are more easily calculated than some existing similarity measures [10, 24], as well as consider more information than some existing similarity measures [43].

Example 1

Assume that there are four kinds of minerals A = {A 1, A 2, A 3, A 4}, and a recognized sample \(\varepsilon\), which are represented by IFSs in the feature space C = {c 1, c 2, c 3}. Suppose we have the following data.

$$A_{1} = \left\{ {\left\langle {c_{1} ,0.4,0.4} \right\rangle ,\left\langle {c_{2} ,0.3,0.5} \right\rangle ,\left\langle {c_{3} ,0.6,0.3} \right\rangle } \right\},$$
$$A_{2} = \left\{ {\left\langle {c_{1} ,0.2,0.6} \right\rangle ,\left\langle {c_{2} ,0.3,0.5} \right\rangle ,\left\langle {c_{3} ,0.4,0.6} \right\rangle } \right\},$$
$$A_{3} = \left\{ {\left\langle {c_{1} ,0.2,0.4,} \right\rangle ,\left\langle {c_{2} ,0.4,0.5} \right\rangle ,\left\langle {c_{3} ,0.3,0.5} \right\rangle } \right\},$$
$$A_{4} = \left\{ {\left\langle {c_{1} ,0.2,0.5} \right\rangle ,\left\langle {c_{2} ,0.4,0.4,} \right\rangle ,\left\langle {c_{3} ,0.4,0.3} \right\rangle } \right\},$$
$$\varepsilon = \left\{ {\left\langle {c_{1} ,0.2,0.6} \right\rangle ,\left\langle {c_{2} ,0.2,0.5} \right\rangle ,\left\langle {c_{3} ,0.4,0.3} \right\rangle } \right\}.$$

Calculate the similarity measure between A i (i = 1,2,3,4) and \(\varepsilon\), the results are presented in Table 1.

Table 1 The results with respect to the different similarity measures (p = 1)

From the boldface letters in Table 1, it shows that the sample \(\varepsilon\) belongs to the pattern A 4 according to the similarity measures (8)–(10) and S L . Furthermore, the sample \(\varepsilon\) belongs to the pattern A 2 according to the similarity measures S N and S SK . However, the similarity measures S LS and S M cannot classify this sample.

3.3 The Shapley-weighted similarity measures of IFSs

Although there are many similarity measures of IFSs, they are all based on the assumptions that the elements in a set are independent, and each feature has the same importance. In most situations, these assumptions do not hold, for example, we give the following classical example: “we are to evaluate a set of different brands of cars in relation to three subjects: {security, service, price}, we want to give more importance to security than to service or price, but on the other hand we want to give some advantage to cars that are good in security and in any of service and price”. In this situation, it is not suitable to endow their weights by using additive measures. Fuzzy measures [30] as a powerful tool for modeling the interaction among elements can well deal with this issue.

Definition 5

[30] A fuzzy measure on a finite set N = {1,2,…,n} is a set function μ: P(N) → [0, 1] satisfyingμ (ϕ) = 0, μ (N) = 1,If A, BP(N) and A ⊆ B then μ (A) ≤ μ (B),

where P(N) denotes the power set of N.

In the pattern recognition, \(\mu (A)\) can be viewed as the importance degree of feature set A. Especially, if \(A = \{ i\}\), then \(\mu (i)\) is the importance degree of the feature i. When we have \(\mu (A) = \sum\nolimits_{i \in A} {\mu (i)}\) for any \(A \in P(N)\), the fuzzy measure \(\mu\) degenerates to be an additive measure.

When there are inter-dependent or interactive phenomena among features, the importance of each feature is not only determined by itself, but also receives the influence from other features. In order to overall reflect the interaction between features, we shall use their Shapley values as their weights. The Shapley function [44] as one of the most important payoff indices has been deeply researched in game theory, which satisfies several reasonable axioms, denoted by

$$\varphi_{i} (\mu ,N )= \sum\limits_{S \subseteq N\backslash i} {\frac{(n - s - 1)!s!}{n!}} (\mu (S \cup i) - \mu (S)),\;\forall i \in N$$
(16)

where \(\mu\) is a fuzzy measure as given in Definition 5, s and n denote the cardinalities of S and N, respectively.

From Definition 5 and the Shapley function, it is not difficult to get \(\varphi_{i} (\mu ,N )\ge 0\) for each \(i \in N\) and \(\sum\nolimits_{i = 1}^{n} {\varphi_{i} (\mu ,N )} = \mu (N)\). Thus, \(\left\{ {\varphi_{i} (\mu ,N )} \right\}_{i \in N}\) is a weight vector. Further, if the fuzzy measure \(\mu\) is an additive measure, namely, there is no interaction between features; their Shapley values are equal to the importance of themselves. That’s \(\varphi_{i} (\mu ,N )= \mu (i)\) for any \(i \in N\). Based on the introduced similarity measures in Sect. 3.2, we give the following Shapley-weighted similarity measures of IFSs.

$$S_{SN} (A,B) = \frac{{\sum\nolimits_{i = 1}^{n} {\varphi_{i} (\mu ,N )\left( {1 - \hbox{min} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}{{\sum\nolimits_{i = 1}^{n} {\varphi_{i} (\mu ,N )\left( {1 + \hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}$$
(17)
$$S_{SL} (A,B) = \frac{{\sum\nolimits_{i = 1}^{n} {\varphi_{i} (\mu ,N )\left( {1 - \hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}{{\sum\nolimits_{i = 1}^{n} {\varphi_{i} (\mu ,N )\left( {1 + \hbox{min} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} } \right)} }}$$
(18)
$$S_{SSK} (A,B) = \frac{1}{n}\sum\limits_{i = 1}^{n} {\varphi_{i} (\mu ,N )\frac{{1 - \hbox{min} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }}{{1 + \hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }}}$$
(19)

where \(\varphi_{i} (\mu ,N )\) as given (16).

Similarly, we have the following Shapley-weighted similarity measures of IFSs.

(1) The Shapley-weighted similarity measure [24]

$$S_{SLS} (A,B) = 1 - {\root{p}\of{{\frac{{\sum\nolimits_{i = 1}^{n} {\left( {\varphi_{i} (\mu ,N )(f_{{u_{AB} }} (i) + f_{{v_{AB} }} (i))} \right)^{p} } }}{n},}}}$$
(20)

where \(f_{{u_{AB} }} (i)\) and \(f_{{v_{AB} }} (i)\) as given in (6).

(2) The Shapley-weighted similarity measure [10]

$$S_{SM} (A,B) = \frac{1}{2}\left( {S_{Su} (A,B) + S_{Sv} (A,B)} \right) ,$$
(21)

where

$$S_{u} (A,B) = 1 - \root{p}\of{{\frac{{\sum\nolimits_{i = 1}^{n} {\left( {\varphi_{i} (\mu ,N )|u_{A} (x_{i} ) - u_{B} (x_{i} )|} \right)^{p} } }}{n}}},$$
$$S_{v} (A,B) = 1 - {\root{p}\of{{\frac{{\sum\nolimits_{i = 1}^{n} {\left( {\varphi_{i} (\mu ,N )|v_{A} (x_{i} ) - v_{B} (x_{i} )|} \right)^{p} } }}{n}}}}.$$

(3) The Shapley-weighted similarity measures [43]

$$S_{SHY1} (A,B) = 1 - d_{SH} (A,B),$$
(22)
$$S_{SHY2} (A,B) = \frac{{e^{{ - d_{SH} (A,B)}} - e^{ - 1} }}{{1 - e^{ - 1} }},$$
(23)
$$S_{SHY3} (A,B) = \frac{{1 - d_{SH} (A,B)}}{{1 + d_{SH} (A,B)}},$$
(24)

where \(d_{SH} (A,B) = \frac{1}{n}\sum\limits_{i = 1}^{n} {\varphi_{i} (\mu ,N )\hbox{max} \{ |u_{A} (x_{i} ) - u_{B} (x_{i} )|,|v_{A} (x_{i} ) - v_{B} (x_{i} )|\} }\) and A,B ∈ IFS(X).

If there is no interaction among elements in N, then we get their corresponding weighted similarity measures.

4 Approaches to pattern recognition based on fuzzy measures

4.1 A general case

If the fuzzy measure of each combination in feature set is given, then we give the following decision procedure to pattern recognition under Atanassov’s intuitionistic fuzzy environment.

Step 1: Suppose that there exist m patterns A = {A 1, A 2,…,A m } and n features C = {c 1,c 2,…,c n }. The evaluation of each pattern A i w.r.t. each feature c j is an Atanassov’s IFV

$$A_{i} = \left\{ {\left\langle {c_{j} ,a_{ij} ,b_{ij} } \right\rangle |j\; = \;1,2, \ldots ,n} \right\}\quad i = 1,2, \ldots ,m.$$

Furthermore, assume that there is a sample \(\varepsilon\) to be recognized, which is represented by an IFS \(\varepsilon = \left\{ {\left\langle {c_{j} ,e_{j} ,f_{j} } \right\rangle |j\; = \;1,2, \ldots ,n} \right\}.\)

Step 2: Calculate the Shapley value of each feature using (16).

Step 3: Calculate the Shapley-weighted similarity measure between A i (i = 1,2,…,m) and \(\varepsilon\) using (17) or (18, 19), and then select the best one.

Step 4: End.

According to the entropy theory, if the entropy value for a feature is small across patterns, it can provide decision-makers with useful information. Therefore, the feature should be assigned a bigger weight; otherwise, such a feature will be judged unimportant by most decision-makers. In other words, such a feature should be evaluated as a very small weight. If the information about the weights of features is incompletely known, the following linear programming model for the optimal \(\lambda\)-fuzzy measure on feature set C is built.

$${ \hbox{min} }\;\sum\limits_{j = 1}^{n} {\sum\limits_{i = 1}^{m} {E(A_{ij} } } )\varphi_{j} (\mu ,C )$$
$${\text{s}} . {\text{t}}.\left\{ {\begin{array}{*{20}l} {\mu \left( \emptyset \right) = 0,\quad \mu \left( C \right) = 1,} \\ {\mu \left( {c_{j} } \right) \in H_{j} ,\quad j = 1,2, \ldots ,n,} \\ {\mu \left( S \right) \le \mu \left( T \right)\;\forall S,\quad T \subseteq C,\;S \subseteq T,} \\ \end{array} } \right.\;$$
(25)

where \(\varphi_{j} (\mu ,C )\) is the Shapley value of the feature c j , and \(H_{j} = [h_{j}^{ - } ,h_{j}^{ + } ]\) is its range.

Since (28) is a linear programming, we can easily get its solution by using Simplex method. If there are no interactive characteristics among elements in a set, then we get the corresponding model for the optimal weight vector.

Example 2

In Example 1, if the importance of features is different, which is, respectively, given by [0.4,0.6], [0.3,0.5] and [0.6,0.8]. Then the main steps are given as follows:

Step 1: Calculate the fuzzy measures of all combinations in feature set C. Let E = E BB and \(\varPhi (u_{A} (x),v_{A} (x)) =\) \(u_{A} (x) + v_{A} (x)\). From the model (25), the following linear programming model is established.

$$\hbox{min} {\kern 1pt} {\kern 1pt} {\kern 1pt} 0.12(\mu^{BB} (c_{1} ) - \mu^{BB} (c_{2} ,c_{3} )) - 0.03(\mu^{BB} (c_{2} ) - \mu^{BB} (c_{1} ,c_{3} ))\; - 0.08(\mu^{BB} (c_{3} ) - \mu^{BB} (c_{1} ,c_{2} )) + 0.73$$
$${\text{s}} . {\text{t}} .\left\{ {\begin{array}{*{20}l} {\mu^{BB} (S) \le \mu^{BB} (T){\kern 1pt} {\kern 1pt} {\kern 1pt} S,T \subseteq \{ c_{1} ,c_{2} ,c_{3} \} ,{\kern 1pt} {\kern 1pt} S \subseteq T} \\ \begin{gathered} \mu^{BB} (c_{1} ) \in [0.4,0.6],\mu^{BB} (c_{2} ) \in [0.3,0.5], \hfill \\ \mu^{BB} (c_{3} ) \in [0.6,0.8] \hfill \\ \end{gathered} \\ \end{array} } \right..$$

Solve above linear programming, it gets

$$\mu^{BB} (c_{1} ) = \mu^{BB} (c_{2} ) = \mu^{BB} (c_{1} ,c_{2} ) = 0.4,\mu^{BB} (c_{3} ) =$$
$$\mu^{BB} (c_{1} ,c_{3} ) = 0.8,\mu^{BB} (c_{2} ,c_{3} ) = \mu^{BB} (c_{1} ,c_{2} ,c_{3} ) = 1;$$

Step 2: By (16), calculate the feature Shapley values w.r.t. µ BB, it has

$$\varphi_{1} (\mu^{BB} ,C )= 0.13,\varphi_{2} (\mu^{BB} ,C )= 0.23,\varphi_{3} (\mu^{BB} ,C )= 0.63;$$

Step 3: Calculate the Shapley-weighted similarity measure between A i (i = 1,2,3,4) and \(\varepsilon\), the results are presented in Table 2.

Table 2 The results for the Shapley-weighted similarity measures (p = 1)

Similarly, when E = E N or E = E L , the results are presented in Table 3.

Table 3 The results for the Shapley-weighted similarity measures (p = 1)

Furthermore, when E = E SK , the results are presented in Table 4.

Table 4 The results for the Shapley-weighted similarity measures (p = 1)

According to Tables 2, 3 and 4, the different ranking results are obtained by using the Shapley-weighted similarity measures (17)–(24). But all ranking results show that the recognized sample \(\varepsilon\) belongs to the fourth kind of minerals (A 4), see the boldface letters in Tables 2, 3 and 4.

Example 3

[45] Let us consider a set of diagnoses Q = {Q 1(Viral fever), Q 2(Malaria), Q 3(Typhoid)}, and a set of symptoms S = {s 1(Temperature), s 2(Headache), s 3(Cough)}. Suppose a patient, with respect to all the symptoms, can be represented by the following IFS:

P (patient) = {<s 1, 0.6, 0.1>, <s 2, 0.5, 0.4>, <s 3, 0.7, 0.2>};

And assume each diagnosis Q i (i = 1,2,3) can also be viewed as an IFS with respect to all the symptoms as follows:

Q 1 (Viral fever) = {<s 1, 0.4, 0>, <s 2, 0.3, 0.5>, <s 3, 0.4, 0.3>};

Q 2 (Malaria) = {<s 1, 0.7, 0>, <s 2, 0.2, 0.6>, <s 3, 0.7, 0>};

Q 3 (Typhoid) = {<s 1, 0.3, 0.3>, <s 2, 0.6, 0.1> , <s 3, 0.2, 0.6>}.

The importance of symptoms is different, which is, respectively, given by [0.4,0.5], [0.3,0.4] and [0.2,0.3]. Our aim is to classify the patient P to one of the diagnoses Q 1, Q 2 and Q 3. Similar to Example 2, if E = E BB , E = E N or E = E SK the results are presented in Table 5.

Table 5 The results for the Shapley-weighted similarity measures (p = 1)

When E = E L , the results are presented in Table 6.

Table 6 The results for the Shapley-weighted similarity measures (p = 1)

According to Table 5, the same ranking results are obtained, and from Table 6, we get the different ranking results. However, all ranking results show that the patient belongs to the diagnosis Q 2 (Malaria), see the boldface letters in Tables 5 and 6.

4.2 A special case

As we know, the fuzzy measure is defined on the power set, which makes the problem exponentially complex. Thus, it is not easy to get the fuzzy measure of each combination in a set when it is large. For this reason, we further research pattern recognition under \(\lambda\)-fuzzy measures, which will largely simplify the complexity of solving a fuzzy measure.

Definition 6

[30] Let \(g_{\lambda }\): P(N) → [0, 1] be a fuzzy measure.\(g_{\lambda }\) is called a \(\lambda\)-fuzzy measure if

$$g_{\lambda } (A \cup B) = g_{\lambda } (A) + g_{\lambda } (B) + \lambda g_{\lambda } (A)g_{\lambda } (B)$$

for any \(A,B \subseteq N\) with \(A \cap B = \emptyset\), where \(\lambda > - 1\).

It is evident if \(\lambda = 0\), then \(g_{\lambda }\) is an additive measure, which means there is no interaction between subsets A and B. If \(\lambda > 0\), then \(g_{\lambda }\) is a superadditive measure, which indicates there exists complementary interaction between subsets A and B. If \(- 1 < \lambda < 0\), then \(g_{\lambda }\)is a subadditive measure, which shows there exists redundancy interaction between subsets A and B.

For a finite set N, the \(\lambda\)-fuzzy measure \(g_{\lambda }\) can be equivalently expressed by

$$g_{\lambda } (A) = \left\{ {\begin{array}{*{20}c} {\frac{1}{\lambda }\left( {\mathop \varPi \limits_{i \in A} [1 + \lambda g_{\lambda } (i)] - 1} \right){\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\text{if}}{\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} \lambda \ne 0} \\ {{\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} \sum\limits_{i \in A} {g_{\lambda } (i)} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} {\text{if}}{\kern 1pt} {\kern 1pt} {\kern 1pt} {\kern 1pt} \lambda = 0{\kern 1pt} } \\ \end{array} } \right..$$
(26)

Since \(\mu (N) = 1\), we know \(\lambda\) is determined by \(\mathop \varPi \limits_{i \in N} [1 + \lambda g_{\lambda } (i)] = 1 + \lambda\). So when each \(g_{\lambda } (i)\) is given, one can get the value of \(\lambda\). From (26), for a set with n elements it only needs n values to get a \(\lambda -\) fuzzy measure.

When the information about the weights of features is partly known, the following linear programming model for the optimal \(\lambda -\) fuzzy measure on feature set C is built.

$${ \hbox{min} }\;\sum\limits_{j = 1}^{n} {\sum\limits_{i = 1}^{m} {E(A_{ij} } } )\varphi_{j} (g_{\lambda } ,C )$$
$${\text{s}} . {\text{t}} .\left\{ {\begin{array}{*{20}l} {g_{y} \left( \emptyset \right) = 0,g_{y} \left( C \right) = 1,} \\ {g_{y} \left( {c_{j} } \right) \in H_{j} \left( {j = 1,2, \ldots ,n} \right)} \\ {\lambda > - 1} \\ \end{array} } \right.$$
(27)

where \(\varphi_{j} (\mu ,N )\) is the Shapley value of the feature c j (j = 1, 2,…,n), and \(H_{j} = [h_{j}^{ - } ,h_{j}^{ + } ]\) is its range.

Similar to the decision procedure given in Sect. 4.1, the main steps to Atanassov’s intuitionistic fuzzy pattern recognition under \(\lambda\)-fuzzy measures are given as follows:

Step 1′: See Step 1.

Step 2′: When the fuzzy measure of each combination in feature set is given, calculate their Shapley values by (16); otherwise, use the model (27) to get the optimal \(\lambda\)-fuzzy measure on feature set C. Then, calculate their Shapley values using (16).

Step 3′: Calculate the Shapley-weighted similarity measures between A i (i = 1,2,…,m) and \(\varepsilon\) using (17) [or (18), (19)], and then select the best one.

Step 4′: End.

Example 4

In Example 2, if we make a decision by using the \(\lambda\)-fuzzy measures, then the main steps are given as follows:

Step I-1′: Calculate the optimal \(\lambda\)-fuzzy measure on feature set C, it gets the optimal \(\lambda\)-fuzzy measure on feature set C is \(\left( {g_{\lambda } (c_{j} )} \right)_{{j \in \{ 1,2,3\} }} = \left( {0.4,0.3,0.6} \right)\), where \(\lambda = - 0.9439\).

Step II-2′: By (16), it gets the feature Shapley values

$$\varphi_{1} (g_{\lambda } ,C )= 0.3032,\varphi_{2} (g_{\lambda } ,C )= 0.2409,\varphi_{3} (g_{\lambda } ,C )= 0.456.$$

Step II-3′: Calculate the Shapley-weighted similarity measure between A i (i = 1,2,3,4) and \(\varepsilon\), the results are presented in Table 7.

Table 7 The results for the Shapley-weighted similarity measures (p = 1)

In Example 4, the same ranking results are obtaining by using the Shapley-weighted similarity measures (17)–(24), and all ranking results show that the recognized sample \(\varepsilon\) belongs to the fourth kind of minerals (A 4), see the boldface letters in Table 7.

Example 5

In Example 3, if we make a decision by using \(\lambda\)-fuzzy measures, Similar to Example 4, if E = E BB , E = E N or E = E SK ; the results are presented in Table 8.

Table 8 The results for the Shapley-weighted similarity measures (p = 1)

Similarly, if E = E L ; the results are presented in Table 9.

Table 9 The results for the Shapley-weighted similarity measures (p = 1)

According to Tables 8 and 9, the same ranking results are obtained, and the ranking results all show that the patient to the diagnosis Q 2 (malaria), see the boldface letters in Tables 8 and 9.

5 Conclusions

We have researched entropy and similarity measures of IFSs and proposed a new entropy of IFSs. Furthermore, we give a construction method to get the similarity measure of IFSs by using entropy. To deal with the interactive characteristics among features, we further introduce three Shapley-weighted similarity measures of IFSs, which can be seen an extension of some weighted similarity measures. When the information about the weights of features is incompletely known, the model for the optimal fuzzy measure on feature set is built. As we know, the fuzzy measure is defined on the power set, which makes the problem exponentially complex. In order to simplify the complexity of solving a fuzzy measure and reflect the interaction among features, we further research the pattern recognition problems under \(\lambda\)-fuzzy measures. Based on the introduced entropy, similarity measure and the model for the optimal fuzzy measure on feature set, we develop an approach to pattern recognition problems under Atanassov’s intuitionistic fuzzy environment.

Similar to the application of the Shapley-weighted similarity measures in pattern recognition, we can also apply them in some other fields, such as digital image processing, clustering analysis and decision-making.