1 Introduction

Rough set theory, proposed by Pawlak Pawlak (1982), provides a useful method for dealing with incomplete and inconsistent knowledge. The theory is widely used in various fields, such as knowledge recognition, data mining, image processing and so on.

We know that classical rough set theory is based on an equivalence relation. Objects in the same equivalence class are indiscernible. Pawlak’s rough sets are usually used to deal with data sets described with nominal features. However, in many practical situations, the data sets are not suitable for being handled by Pawlak’s rough sets (Bonikowski et al. 1998; Baszczyński et al. 2011; Cattaneo 1998; Chen et al. 2006, 2007, 2011; Diker and Ugur 2012). As a result, many generalized rough set models have been developed in terms of different requirements. Then, the neighborhood-based rough sets (Wang et al. 2016, 2017; Yao 2011), similarity relation rough sets (Slowinski and Vanderpooten 2000; Yao 1998), tolerance relation rough sets (Skowron and Stepaniuk 1996; Xu et al. 2013), and binary relation rough sets (Wang et al. 2017) were constructed, respectively. In particular, some objets have multiple attribute values for a multi-valued attribute. So, we need a covering of the universe instead of a partition. In 1983, Zakowski firstly proposed the covering rough set model Zakowski (1983). After Zakowski, many authors studied the properties of covering rough sets (Bonikowski et al. 1998; Chen et al. 2007; Ge and Li 2011; Kong and Xu 2018a, b; Kong and Wei 2015; Liu and Wang 2011; Liu et al. 2014; Liu and Zhu 2008; Liu and Sai 2009; Shi and Gong 2010). Especially, more and more scholars are working on the attribute reducts of covering rough sets or covering information systems. For example, a pioneering work related to the reducts of covering rough sets was constructed, where the concept of reducts of covering was introduced and the procedure to find a reduct for a covering was shown Zhu and Wang (2003). Meanwhile, the approach to attribute reducts of consistent and inconsistent covering decision systems are firstly introduced by Chen et al. (2007). At the same time, Chen et al. (2007) originally proposed the discernibility to design algorithms that compute all the reducts of consistent and inconsistent covering decision systems. Compared with the attribute reduct method presented in Chen et al. (2007), Wang et al. constructed a new discernibility matrix, which greatly reduced the computational complexity (Wang et al. 2014). In Wang et al. (2015), Wang et al. also provided a new method for constructing simpler discernibility matrix with covering rough sets, and improved some characterizations of attribute reduct provided by Tsang et al. (2008). Moreover, Tan et al. introduced matrix-based methods for computing set approximations and reducts of a covering information system (Tan et al. 2015).

From the perspective of granular computing, an equivalence relation on the universe can be regarded as a granularity, and the corresponding partition can be regarded as a granular structure. Hence, Pawlak’s rough set theory is based on a single granularity. In Qian et al. (2010), Qian and Liang firstly extended the single granulation rough sets to the multiple granulation rough sets, where the set approximations were defined by using multiple equivalence relations on the universe. At present, more and more attention has been paid to extending the multi-granulation rough set theory (Li et al. 2016, 2017; Lin et al. 2013; Liu and Wang 2011; Zhang and Kong 2016). Xu et al. developed the multi-granulation rough set model in ordered information systems (Xu et al. 2012). At the same time, based on the multi-granulation rough set theory and fuzzy set theory, Xu et al. proposed a multi-granulation fuzzy rough set model and a multi-granulation fuzzy rough sets in a fuzzy tolerance approximation space (Xu et al. 2014). Meanwhile, Yang also generalized the multi-granulation rough sets into fuzzy rough sets, and discussed the corresponding properties in incomplete information systems Yang et al. (2011). Kong et al. studied the operation and algebraic properties of multi-granulation rough sets and multi-granulation covering rough sets, respectively (Kong et al. 2018; Kong and Wei 2017). She et al. explored the topological structures and obtained many excellent conclusions (She and he 2003). Recently, Xu et al. proposed a generalized multi-granulation rough set model by introducing a support characteristic function and an information level (Xu et al. 2011). However, it is still an open problem regarding the attribute reducts of multi-granulation information systems. Therefore, the main objective of this paper is to study the attribute reduct theory of multi-granulation information system.

The rest of this paper is organized as follows. In Sect. 2, we briefly review some basic concepts of Pawlak’s rough sets, multi-granulation rough sets and covering. In Sect. 3, we firstly introduce the concept of the arbitrary union reduct, and discuss some basic properties of this reduct. Furthermore, we develop an algorithm to compute the arbitrary union reduct in an information system. In Sect. 4, we propose the concept of neighborhood union reduct, and study some important properties of neighborhood union reduct. In order to deal with real-life cases, we develop an algorithm to compute the neighborhood union reduct in an information system. In Sect. 5, we define the neighborhood intersection reduct. Many meaningful results of this reduct are explored. In Sect. 6, the interrelationships among the several types of reducts are discussed in detail. In Sect. 7, an illustrate example is given to show how to select the optimal reduct to compute the multi-granulation lower and upper approximations more efficiently. Finally, Sect. 8 concludes this study.

2 Preliminaries

In this section, we review some basic concepts and notions of Pawlak’s rough sets, multi-granulation rough sets and covering. More details can be seen in references Chen et al. (2007), Qian et al. (2010), Qian and Liang (2006), Zhu and Wang (2003).

Let be an information system, where U is a nonempty finite set, called a universe; A is a nonempty attribute set; \(V=\cup _{a\in A}V_{a}\), \(V_{a}\) is a set of its values; \(f: U\times A\rightarrow V\) is an information function with \(f(x,a)\in V_{a}\) for each \(a\in A\) and \(x\in U\). The family of attribute subsets is denoted by , where \(A_{i}\subseteq A,~ i=1,2,\ldots ,m.\) The equivalence class of an object x with respect to is defined by: \([x]_{A_{i}}=\{y\in U|f(x,a)=f(y,a), a\in A_{i}\}\). Let \(U/A_{i}=\{[x_{1_{i}}]_{A_{i}},[x_{2_{i}}]_{A_{i}},\ldots ,[x_{n_{i}}]_{A_{i}}\}\) is a partition of U and \([x_{1_{m}}]_{A_{m}},[x_{2_{m}}]_{A_{m}},\ldots ,[x_{n_{m}}]_{A_{m}}\}\). Then, for each \(X\subseteq U\), the lower and upper approximations of X with respect to \(A_{i}\) are defined as follows:

$$\begin{aligned} \underline{A_{i}}(X)=\{x\in U|[x]_{A_{i}}\subseteq X\},~~~~ \overline{A_{i}}(X)=\{x\in U|[x]_{A_{i}}\cap X\ne \emptyset \}. \end{aligned}$$

Definition 2.1

Qian et al. (2010) Let be an information system, \(X\subseteq U\), and . The optimistic multi-granulation lower and upper approximations of X with respect to are defined as follows:

$$\begin{aligned} \underline{OM_{\sum _{i=1}^m A_{i}}}(X)=\{x\mid \vee _{i=1}^m ([x]_{A_{i}}\subseteq X)\}, ~~~~\overline{OM_{\sum _{i=1}^m A_{i}}} (X)=\{x\mid \wedge _{i=1}^m ([x]_{A_{i}}\cap X\ne \emptyset )\}. \end{aligned}$$

where “\(\vee \)” means the logical operator “or”, which represents that the alternative conditions are satisfied, and “\(\wedge \)” means the logical operator “and”, which represents that all of the conditions are satisfied. Here, the word “optimistic” means that just one granular structure is needed to satisfy with the inclusion between an equivalence class and a target concept when multiple independent granular structures are available in problem processing.

Definition 2.2

Qian and Liang (2006) Let be an information system, \(X\subseteq U\), and . The pessimistic multi-granulation lower and upper approximations of X with respect to are defined as follows:

$$\begin{aligned} \underline{PM_{\sum _{i=1}^m A_{i}}}(X)=\{x\mid \wedge _{i=1}^m ([x]_{A_{i}}\subseteq X)\}, ~~~~\overline{PM_{\sum _{i=1}^m A_{i}}} (X)=\{x\mid \vee _{i=1}^m ([x]_{A_{i}}\cap X\ne \emptyset )\}. \end{aligned}$$

Here, the word “pessimistic” means that all granular structures are needed to satisfy with the inclusion between an equivalence class and a target concept when multiple independent granular structures are available.

Definition 2.3

Zakowski (1983) Let be a family of nonempty subsets of U. is called a covering of U if . The ordered pair is called a covering approximation space.

Definition 2.4

Zhu and Wang (2003) Let be a covering approximation space and . If K is a union of some sets in , we say that K is a union reducible element of ; Otherwise, we say that K is a union irreducible element of . Meanwhile, for a covering of U, the new union irreducible covering through above reduction is called a union reduct of , and denoted by .

Notice that the union reduct of a covering can be computed by deleting all the union reducible elements. Therefore, the union reduct is the minimum covering by deleting redundancy. Meanwhile, for each subset of the universe, the union reduct of a covering can induce the same lower and upper approximations (Zhu and Wang 2003). However, for the union reduct of a covering, it may contain other redundant elements which are not the union reducible elements. To address this issue, the intersection reduct of a covering is proposed by Chen et al. (2015).

Definition 2.5

Chen et al. (2015) Let be a covering approximation space and . If K is an intersection of some sets in , we say that K is an intersection reducible element of ; Otherwise, we say that K is an intersection irreducible element of . Meanwhile, for a covering of U, the new intersection irreducible covering through above reduction is called an intersection reduct of , and denoted by .

In Chen et al. (2015), the properties of the intersect reduct of a covering are examined. Particularly, the intersection reduct is investigated from the viewpoint of concept lattice theory.

3 Attribute reduction with respect to arbitrary union

In this section, we will introduce the concept of the arbitrary union reduct, and then discuss some interesting properties of this reduct. Meanwhile, we develop an algorithm to compute the arbitrary union reduct.

Definition 3.1

Let be an information system, and the family of attribute subsets. is called an arbitrary union reducible element of , if for each \(x\in U\), there exist \(\Gamma _{x}\subseteq \{1,2,\ldots ,i-1,i+1,\ldots ,m\}\) and \(V_{x}\subseteq U\) such that \([x]_{A_{i}}=\cup _{j\in \Gamma _{x}} \cup _{y\in V_{x}} [y]_{A_{j}}\); Otherwise, \(A_{i}\) is called an arbitrary union irreducible element of . If every element in is irreducible, we say that is irreducible; Otherwise, is reducible.

Definition 3.2

Let be an information system, and the family of attribute subsets. The new family of attribute subsets through the above reduct is called the arbitrary union reduct of , and denoted by .

Notice that the covering of a universe U has only one reduct. i.e., is unique (Zhu and Wang 2003). Here, we raise a question: is unique? In the following, we will employ an example to answer the question.

Example 3.1

Let be an information system, where \(U=\{x_{1},x_{2},\ldots ,x_{20}\}\), .

$$\begin{aligned} U/A_{1}= & {} \{\{x_{1}\},\{x_{2}\},\{x_{3}\},\{x_{4}\},\{x_{5}\},\{x_{6}\}, \{x_{7},x_{8},x_{9}\},\{x_{10},x_{11},x_{12}\},\\&\{x_{13},x_{14},x_{15}, x_{16}\},\{x_{17},x_{18}\},\{x_{19},x_{20}\}\};\\ U/A_{2}= & {} \{\{x_{1},x_{2},x_{3}\},\{x_{4},x_{5},x_{6}\},\{x_{7}\}, \{x_{8}\},\{x_{9}\},\{x_{10}\},\{x_{11}\},\{x_{12}\},\\&\{x_{13},x_{14},x_{17}\}, \{x_{15}\},\{x_{16},x_{19},x_{20}\},\{x_{18}\}\};\\ U/A_{3}= & {} \{\{x_{1},x_{2}\},\{x_{3},x_{4}\},\{x_{5},x_{6}\}, \{x_{7},x_{8}\},\{x_{9},x_{10},x_{11},x_{12}\},\\&\{x_{13},x_{14},x_{15}\}, \{x_{16},x_{17},x_{18}\},\{x_{19}\},\{x_{20}\}\};\\ U/A_{4}= & {} \{\{x_{1},x_{2},x_{3},x_{4}\},\{x_{5},x_{6},x_{7}\},\{x_{8},x_{9},x_{10}\},\{x_{11},x_{12}\},\\&\{x_{13},x_{14},x_{15}\}, \{x_{16},x_{17},x_{18}\},\{x_{19},x_{20}\}\}. \end{aligned}$$

It is clear that or \(\{A_{1},A_{2},A_{4}\}\). Therefore, is not unique.

If is a union reducible element of and , then \(K_{1}\) is a union reducible element of if and only if \(K_{1}\) is a union reducible element of Zhu and Wang (2003). Similarly, we have the following result.

Proposition 3.1

Let be an information system, and . Suppose is an arbitrary union reducible element of , and . If for each \(x\in U\), we have \([x]_{A_{i}}\ne [x]_{A_{j}}\), then \(A_{j}\) is an arbitrary union reducible element of if and only if \(A_{j}\) is an arbitrary union reducible element of .

Proof

\((\Leftarrow )\) It is immediate.

\((\Rightarrow )\) Suppose \(A_{j}\) is an arbitrary union reducible element of . For each \(x\in U\), there exist \(\Gamma _{x}\subseteq \{1,2,\ldots ,j-1,j+1,\ldots ,m\}\) and \(V_{x}\subseteq U\) such that \([x]_{A_{j}}=\cup _{t\in \Gamma _{x}} \cup _{y\in V_{x}} [y]_{A_{t}}\). Case1: If \(i~\overline{\in } ~\Gamma _{x}\), it is clear that \(A_{j}\) is an arbitrary union reducible element of ; Case2: If \(i\in \Gamma _{x}\), without lost of generality, let \([x]_{A_{j}}=[z]_{A_{i}}\cup (\cup _{t\in \Gamma _{x}/\{i\}} \cup _{y\in V_{x}/\{z\}} [y]_{A_{t}})\), where \(z\in V_{x}\). On the one hand, if \([z]_{A_{i}}=[x]_{A_{i}}\), by the assumption in this proposition, we have \([z]_{A_{i}}\subset [x]_{A_{j}}\). Since \(A_{i}\) is an arbitrary union reducible element of , there exist \(\Gamma _{x}^{'}\subseteq \{1,2,\ldots ,i-1,i+1,\ldots ,m\}\) and \(V_{x}^{'}\subseteq U\) such that \([z]_{A_{i}}=\cup _{s\in \Gamma _{x}^{'}} \cup _{w\in V_{x}^{'}} [w]_{A_{s}}\). Because \(U/A_{i}\) is a partition of U, then \(j~\overline{\in }~ \Gamma _{x}^{'}\). Therefore, \([x]_{A_{j}}=(\cup _{s\in \Gamma _{x}^{'}} \cup _{w\in V_{x}^{'}} [w]_{A_{s}})\cup (\cup _{t\in \Gamma _{x}/\{i\}} \cup _{y\in V_{x}/\{z\}} [y]_{A_{t}})\). Denote \(\widetilde{\Gamma _{x}}=(\Gamma _{x}^{'}\cup \Gamma _{x})/\{i\},~\widetilde{V_{x}}=(V_{x}^{'}\cup V_{x})/\{z\}\), then \([x]_{A_{j}}=\cup _{k\in \widetilde{\Gamma _{x}}} \cup _{u\in \widetilde{V_{x}}} [u]_{A_{k}}\). So, \(A_{j}\) is an arbitrary union reducible element of . On the other hand, if \([z]_{A_{i}}\ne [x]_{A_{i}}\), then we have \([z]_{A_{i}}\subset [x]_{A_{j}}\). According to the above discussion, \(A_{j}\) is an arbitrary union reducible element of . \(\square \)

In the following, we can develop an algorithm to compute the arbitrary union reduct .

figure a

In Zhu and Wang (2003), Zhu and Wang indicated that if K is a union reducible element of and \(X\subseteq U\), then the covering lower approximations of X with respect to and , respectively, are same. Here, we also raise a similar question: if is an arbitrary union reducible element of , and \(X\subseteq U\), are the optimistic multi-granulation lower approximations of X with respect to and , respectively, same? To solve this problem, we have the following result.

Proposition 3.2

Let be an information system, , and \(X\subseteq U\). If is an arbitrary union reducible element of , then the optimistic multi-granulation lower approximations of X with respect to and , respectively, are same.

Proof

By Definition 2.1, it is obvious that \(\underline{OM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\subseteq \underline{OM_{\sum _{i=1}^m A_{i}}}(X)\). Then, for each \(x\in \underline{OM_{\sum _{i=1}^m A_{i}}}(X)\), there exists such that \([x]_{A_{k}}\subseteq X\). If \(A_{k}\ne A_{j}\), then \(x\in \underline{OM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\). If \(A_{k}=A_{j}\), since \(A_{j}\) is a reducible element of , there exists such that \([x]_{A_{s}}\subseteq [x]_{A_{k}}\subseteq X\). According to Definition 2.1, we have \(x\in \underline{OM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\). So, \(\underline{OM_{\sum _{i=1}^m A_{i}}}(X)\subseteq \underline{OM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\). Therefore, \(\underline{OM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)= \underline{OM_{\sum _{i=1}^m A_{i}}}(X)\). \(\square \)

Similarly, the result with respect to the optimistic multi-granulation upper approximation is shown as follows.

Proposition 3.3

Let be an information system, , and \(X\subseteq U\). If is an arbitrary union reducible element of , then the optimistic multi-granulation upper approximations of X with respect to and , respectively, are same.

According to Propositions 3.2 and 3.3, we have the following result.

Theorem 3.1

Let be an information system, , and \(X\subseteq U\). Then the optimistic multi-granulation lower and upper approximations of X with respect to and , respectively, are same.

If is an arbitrary union reducible element of , and \(X\subseteq U\), then, are the pessimistic multi-granulation lower approximations of X with respect to and , respectively, same? In order to answer the question, a counterexample is given as follows:

Example 3.2

Let be an information system, where \(U=\{x_{1},x_{2},\ldots ,x_{9}\}\), .

$$\begin{aligned} U/A_{1}= & {} \{\{x_{1},x_{2}\},\{x_{3},x_{4},x_{5}\},\{x_{6}\},\{x_{7},x_{8}\},\{x_{9}\}\};\\ U/A_{2}= & {} \{\{x_{1}\},\{x_{2},x_{3},x_{4},x_{5},x_{6}\},\{x_{7}\},\{x_{8},x_{9}\}\};\\ U/A_{3}= & {} \{\{x_{1},x_{2},x_{3},x_{4},x_{5}\},\{x_{6},x_{7}\},\{x_{8},x_{9}\}\}; \end{aligned}$$

Clearly, \(A_{3}\) is an arbitrary union reducible element of . For \(X=\{x_{7},x_{8},x_{9}\}\), we have \(\underline{PM_{\sum _{i=1}^3 A_{i}}}(X)=\{x_{8},x_{9}\}\). However, \(\underline{PM_{\sum _{i=1}^2 A_{i}}}(X)=\{x_{7},x_{8},x_{9}\}\). Therefore, \(\underline{PM_{\sum _{i=1}^3 A_{i}}}(X)\ne \underline{PM_{\sum _{i=1}^2 A_{i}}}(X)\).

Similarly, it is not difficult to find that the pessimistic multi-granulation upper approximations of X with respect to and , respectively, are different.

Proposition 3.4

Let be an information system, and , two families of attribute subsets. If for each \(X\subseteq U\), the optimistic multi-granulation lower approximations of X with respect to and , respectively, are same. Then we have that .

Proof

Suppose that . For each , we have \(\underline{OM_{\sum _{k=1}^{m_{1}} A_{1k}}}(K)=K\). Since the optimistic multi-granulation lower or upper approximations of K with respect to and , respectively, are same. Then we have \(\underline{OM_{\sum _{l=1}^{m_{2}} A_{2l}}}(K)=K\). Therefore, there exist such that \(K=\cup _{i=1}^{s}K_{i}\). Similar to the above proof, there exist such that \(K_{i}=\cup _{j=1}^{n_{i}}K_{ij}\). So, \(K=\cup _{i=1}^{s}\cup _{j=1}^{n_{i}}K_{ij}\). By the definition of the arbitrary union reduct, we have \(K_{ij}=K\) for all ij. Therefore, for all i, \(K_{i}=K\). Hence, K is an element of .

On the other hand, we can similarly prove that any element of is an element of . Therefore, we have . \(\square \)

Theorem 3.2

Let be an information system, and , two families of attribute subsets. For each \(X\subseteq U\), the optimistic multi-granulation lower approximations of X with respect to and , respectively, are same if and only if .

Proof

(\(\Rightarrow \)) It is immediate by Proposition 3.4.

(\(\Leftarrow \)) Since . For each \(X\subseteq U\), we have that the optimistic multi-granulation lower approximations of X with respect to and , respectively, are same. By Theorem 3.1, we have that the optimistic multi-granulation lower approximations of X with respect to and , respectively, are same. \(\square \)

Similar to Proposition 3.4 and Theorem 3.2, we have the following results.

Proposition 3.5

Let be an information system, and , two families of attribute subsets. For each \(X\subseteq U\), the optimistic multi-granulation upper approximations of X with respect to and , respectively, are same. Then we have that .

Theorem 3.3

Let be an information system, and , two families of attribute subsets. For each \(X\subseteq U\), the optimistic multi-granulation upper approximations of X with respect to and , respectively, are same if and only if .

4 Attribute reduction with respect to neighborhood union

In this section, we firstly propose the concept of neighborhood union reduct. Then we study some important properties of neighborhood union reduct. Finally, we develop an algorithm to compute the neighborhood union reduct in an information system.

Definition 4.1

Let be an information system, and the family of attribute subsets. is called a neighborhood union reducible element of , if each \(x\in U\), there exists \(\Gamma _{x}\subseteq \{1,2,\ldots ,i-1,i+1,\ldots ,m\}\) such that \([x]_{A_{i}}=\cup _{j\in \Gamma _{x}} [x]_{A_{j}}\); Otherwise, \(A_{i}\) is called a neighborhood union irreducible element of . If every element in is irreducible, we say that is irreducible; Otherwise, is reducible.

Definition 4.2

Let be an information system, and the family of attribute subsets. The new family of attribute subsets through the above reduction is called the neighborhood union reduct of , and denoted by .

Example 3.1 shows that the arbitrary union reduct is not unique. Now, we investigate that whether the neighborhood union reduct is unique. In the following, an example is employed to answer the question.

Example 4.1

Let be an information system, where \(U=\{x_{1},x_{2},\ldots ,x_{9}\}\), .

$$\begin{aligned} U/A_{1}= & {} \{\{x_{1},x_{2},x_{3}\},\{x_{4},x_{5}\},\{x_{6},x_{7}\},\{x_{8},x_{9}\}\};\\ U/A_{2}= & {} \{\{x_{1},x_{2},x_{4},x_{5}\},\{x_{3}\},\{x_{6},x_{7},x_{9}\},\{x_{8}\}\};\\ U/A_{3}= & {} \{\{x_{1},x_{2},x_{3},x_{4},x_{5}\},\{x_{6},x_{7},x_{8}\},\{x_{9}\}\};\\ U/A_{4}= & {} \{\{x_{1},x_{2}\},\{x_{3},x_{4},x_{5}\},\{x_{6},x_{7},x_{8},x_{9}\}\};\\ U/A_{5}= & {} \{\{x_{1},x_{2},x_{4},x_{5}\},\{x_{3}\},\{x_{6},x_{7},x_{8}\},\{x_{9}\}\}. \end{aligned}$$

Then, we can find that or \(\{A_{1},A_{2},A_{4},A_{5}\}\). So is not unique.

Compared with Proposition 3.1, the similar result with respect to the neighborhood union reduct is shown as follows:

Proposition 4.1

Let be an information system, and . Suppose is a neighborhood union reducible element of , and . If for each \(x\in U\), we have \([x]_{A_{i}}\ne [x]_{A_{j}}\), then \(A_{j}\) is a neighborhood union reducible element of if and only if \(A_{j}\) is a neighborhood union reducible element of .

In the following, we give an algorithm to find the neighborhood union reduct .

figure b

Similarly, we have the following results.

Proposition 4.2

Let be an information system, , and \(X\subseteq U\). If is a neighborhood union reducible element of , then the optimistic multi-granulation lower approximations of X with respect to and , respectively, are same.

Proof

The proof is similar to that of Proposition 3.2. \(\square \)

Proposition 4.3

Let be an information system, , and \(X\subseteq U\). If is a neighborhood union reducible element of , then the optimistic multi-granulation upper approximations of X with respect to and , respectively, are same.

Proof

The proof is similar to that of Proposition 3.3. \(\square \)

According to Propositions 4.2 and 4.3, we have the following result.

Theorem 4.1

Let be an information system, , and \(X\subseteq U\). Then the optimistic multi-granulation lower and upper approximations of X with respect to and , respectively, are same.

If is a neighborhood union reducible element of , and \(X\subseteq U\), then, are the pessimistic multi-granulation lower approximations of X with respect to and , respectively, same?

Proposition 4.4

Let be an information system, , and \(X\subseteq U\). If is a neighborhood union reducible element of , then the pessimistic multi-granulation lower approximations of X with respect to and , respectively, are same.

Proof

Based on Definition 2.2, it is obvious that \(\underline{PM_{\sum _{i=1}^m A_{i}}}(X)\subseteq \underline{PM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\). On the other hand, for each \(x\in \underline{PM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\), we have that \(x\in [x]_{A_{i}},~i=1,2,\ldots ,j-1,j+1,\ldots ,m.\) Because is a neighborhood union reducible element of , then there exists \(\Gamma _{x}\subseteq \{1,2,\ldots ,j-1,j+1,\ldots ,m\}\) such that \([x]_{A_{j}}=\cup _{t\in \Gamma _{x}} [x]_{A_{t}}\). So \(x\in [x]_{A_{j}}\). i.e., \(x\in [x]_{A_{i}},~i=1,2,\ldots ,m.\) Therefore, \(\underline{PM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\subseteq \underline{PM_{\sum _{i=1}^m A_{i}}}(X)\). In a word, we have that \(\underline{PM_{\sum _{i=1}^m A_{i}}}(X)=\underline{PM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\). \(\square \)

If is a neighborhood union reducible element of , and \(X\subseteq U\), then, are the pessimistic multi-granulation upper approximations of X with respect to and , respectively, still same?

Proposition 4.5

Let be an information system, , and \(X\subseteq U\). If is a neighborhood union reducible element of , then the pessimistic multi-granulation upper approximations of X with respect to and , respectively, are same.

Proof

The proof is similar to that of Proposition 4.4. \(\square \)

From Propositions 4.4 and 4.5, we have the following result.

Theorem 4.2

Let be an information system, , and \(X\subseteq U\). Then the pessimistic multi-granulation lower and upper approximations of X with respect to and , respectively, are same.

5 Attribute reduction with respect to neighborhood intersection

In this section, inspired by the concept of intersection reduct proposed by Chen et al. (2015), the concept of neighborhood intersection reduct are introduced. Meanwhile, some meaningful properties of neighborhood intersection reduct are studied.

Definition 5.1

Let be an information system, and the family of attribute subsets. is called a neighborhood intersection reducible element of , if for each \(x\in U\), there exists \(\Gamma _{x}\subseteq \{1,2,\ldots ,i-1,i+1,\ldots ,m\}\) such that \([x]_{A_{i}}=\cap _{j\in \Gamma _{x}} [x]_{A_{j}}\); Otherwise, \(A_{i}\) is called a neighborhood intersection irreducible element of . If every element in is irreducible, we say that is irreducible; Otherwise, is reducible.

Definition 5.2

Let be an information system, and the family of attribute subsets. The new family of attribute subsets through the above reduction is called the neighborhood intersection reduct of , and denoted by .

Examples 3.1 and 4.1 show that the arbitrary union reduct and the neighborhood union reduct are both not unique. Next, we wonder if the neighborhood intersection reduct is unique. To address this issue, an example is shown as follows.

Example 5.1

Let be an information system, where \(U=\{x_{1},x_{2},\ldots ,x_{9}\}\), .

$$\begin{aligned} U/A_{1}= & {} \{\{x_{1},x_{2},x_{4}\},\{x_{3},x_{5},x_{6}\},\{x_{7},x_{8}\},\{x_{9}\}\};\\ U/A_{2}= & {} \{\{x_{1},x_{2}\},\{x_{3}\},\{x_{4}\},\{x_{5},x_{6}\},\{x_{7},x_{8},x_{9}\}\};\\ U/A_{3}= & {} \{\{x_{1},x_{2},x_{3}\},\{x_{4},x_{5},x_{6}\},\{x_{7},x_{8}\},\{x_{9}\}\};\\ U/A_{4}= & {} \{\{x_{1}\},\{x_{2},x_{3},x_{4}\},\{x_{5},x_{6}\},\{x_{7}\},\{x_{8},x_{9}\}\};\\ U/A_{5}= & {} \{\{x_{1}\},\{x_{2}\},\{x_{3}\},\{x_{4}\},\{x_{5},x_{6}\},\{x_{7},x_{8},x_{9}\}\}. \end{aligned}$$

Then, we can find that or \(\{A_{1},A_{3},A_{4},A_{5}\}\). Therefore, is not unique.

In the following, we will construct an algorithm for finding the neighborhood intersection reduct .

figure c

Proposition 5.1

Let be an information system, , and \(X\subseteq U\). If is a neighborhood intersection reducible element of , then the pessimistic multi-granulation lower approximations of X with respect to and , respectively, are same.

Proof

It is clear that \(\underline{PM_{\sum _{i=1}^m A_{i}}}(X)\subseteq \underline{PM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\). Meanwhile, for each \(x\in \underline{PM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\), we have \([x]_{A_{s}}\subseteq X,~s\in \{1,2,\ldots ,j-1,j+1,\ldots ,m\}\). Because \(A_{j}\) is a reducible element of , then there exists \(\Gamma _{x}\subseteq \{1,2,\ldots ,j-1,j+1,\ldots ,m\}\) such that \([x]_{A_{j}}=\cap _{t\in \Gamma _{x}} [x]_{A_{t}}\). Then, \([x]_{A_{j}}\subseteq [x]_{A_{t}},~t\in \Gamma _{x}\). i.e., \([x]_{A_{j}}\subseteq X\). So, \(x\in \underline{PM_{\sum _{i=1}^m A_{i}}}(X)\). Therefore, \(\underline{PM_{\sum _{i=1}^m A_{i}}}(X)=\underline{PM_{\sum _{i=1,i\ne j}^m A_{i}}}(X)\). \(\square \)

Proposition 5.2

Let be an information system, , and \(X\subseteq U\). If is a neighborhood intersection reducible element of , then the pessimistic multi-granulation upper approximations of X with respect to and , respectively, are same.

According to Propositions 5.1 and 5.2, the following result holds.

Theorem 5.1

Let be an information system, , and \(X\subseteq U\). Then the pessimistic multi-granulation lower and upper approximations of X with respect to and , respectively, are same.

If is a neighborhood intersection reducible element of , and \(X\subseteq U\), are the optimistic multi-granulation lower approximations of X with respect to and , respectively, same? Then, a counterexample is given as follows:

Example 5.2

(Continued from Example 5.1) According to Example 5.1, we know that \(A_{5}\) is a neighborhood intersection reducible element of . For \(X=\{x_{2},x_{4},x_{6},x_{7}\}\), we have that \(\underline{OM_{\sum _{i=1}^5 A_{i}}}(X)=\{x_{2},x_{4},x_{7}\}, \underline{OM_{\sum _{i=1}^4 A_{i}}}(X)=\{x_{4},x_{7}\}\). Therefore, \(\underline{OM_{\sum _{i=1}^5 A_{i}}}(X)\ne \underline{OM_{\sum _{i=1}^4 A_{i}}}(X)\).

Similarly, if \(A_{i}\) is a neighborhood intersection reducible element of , then the optimistic multi-granulation upper approximations of X with respect to and , respectively, are different.

6 Interrelationships among the three types of attribute reductions

In this section, we will deeply explore the interrelationships among several types of attribute reducts.

Proposition 6.1

Let be an information system, and . If is a neighborhood union reducible element of , then \(A_{j}\) is an arbitrary union reducible element of .

Proof

It is clear by Definitions 3.1 and 4.1.

Conversely, if is an arbitrary union reducible element of , is \(A_{j}\) a neighborhood union reducible element of ? A counterexample is given as follows: \(\square \)

Example 6.1

(Continued from Example 3.2) It can be found that \(A_{3}\) is an arbitrary union reducible element of . However, \(A_{3}\) is not a neighborhood union reducible element of .

Proposition 6.2

Let be an information system, , then and are both the arbitrary union reducts of .

Proof

It is clear by Definitions 3.1 and 4.1.\(\square \)

Theorem 6.1

Let be an information system, , and \(X\subseteq U\), then the optimistic multi-granulation lower and upper approximations of X with respect to , and , respectively, are same.

Theorem 6.2

Let be an information system, , and \(X\subseteq U\), then the pessimistic multi-granulation lower and upper approximations of X with respect to , and , respectively, are same.

Example 6.2

Let be an information system, where \(U=\{x_{1},x_{2},\ldots ,x_{10}\}\), . Let

$$\begin{aligned} U/A_{1}= & {} \{\{x_{1},x_{2}\},\{x_{3},x_{4}\},\{x_{5},x_{6}\},\{x_{7},x_{8}\},\{x_{9},x_{10}\}\};\\ U/A_{2}= & {} \{\{x_{1},x_{2},x_{3},x_{4}\},\{x_{5},x_{6}\},\{x_{7}\},\{x_{8}\},\{x_{9},x_{10}\}\};\\ U/A_{3}= & {} \{\{x_{1},x_{2},x_{3},x_{4},x_{5},x_{6}\},\{x_{7}\},\{x_{8}\},\{x_{9},x_{10}\}\};\\ U/A_{4}= & {} \{\{x_{1},x_{2},x_{7},x_{8},x_{9},x_{10}\},\{x_{3},x_{4},x_{5},x_{6}\}\};\\ U/A_{5}= & {} \{\{x_{1},x_{2},x_{5},x_{6}\},\{x_{3},x_{4},x_{7}\},\{x_{8},x_{9},x_{10}\}\};\\ U/A_{6}= & {} \{\{x_{1},x_{2},x_{3},x_{4},x_{5},x_{6}\},\{x_{7}\},\{x_{8},x_{9},x_{10}\}\}. \end{aligned}$$

Then, It can be found that or \(\{A_{1},A_{3}\}\), . Furthermore, we have that .

Proposition 6.3

Let be an information system, and . Then we have that .

Proof

It is clear by Definitions 3.1 and 4.1. \(\square \)

According to Proposition 6.3, we have that . Then, does the equation hold? A counterexample is given as follows:

Example 6.3

(Continued from Example 6.2) Let . However, . Therefore, . Furthermore, by Proposition 6.4, it can be easily found that .

Example 6.4

Let be an information system, where \(U=\{x_{1},x_{2},\ldots ,x_{10}\}\), . Let

$$\begin{aligned} U/A_{1}= & {} \{\{x_{1},x_{2}\},\{x_{3},x_{4}\},\{x_{5},x_{6}\},\{x_{7},x_{8}\},\{x_{9},x_{10}\}\};\\ U/A_{2}= & {} \{\{x_{1},x_{2},x_{3},x_{4},x_{5}\},\{x_{6},x_{7},x_{8},x_{9},x_{10}\}\};\\ U/A_{3}= & {} \{\{x_{1},x_{2},x_{3},x_{4}\},\{x_{5},x_{6},x_{7},x_{8}\},\{x_{9},x_{10}\}\};\\ U/A_{4}= & {} \{\{x_{1},x_{2},x_{5},x_{6}\},\{x_{3},x_{4}\},\{x_{7},x_{8},x_{9},x_{10}\}\};\\ U/A_{5}= & {} \{\{x_{1},x_{2}\},\{x_{3},x_{4},x_{5}\},\{x_{6},x_{7},x_{8},x_{9},x_{10}\}\};\\ U/A_{6}= & {} \{\{x_{1},x_{2},x_{4},x_{5}\},\{x_{3}\},\{x_{6},x_{7},x_{8},x_{9},x_{10}\}\}. \end{aligned}$$

It can be found that , , and .

Then, we have , and . Similarly, it can be found that , and .

Here, we will raise a question: does the equation hold? In the following, a counterexample is employed to answer the question.

Example 6.5

Let be an information system, where \(U=\{x_{1},x_{2},\ldots ,x_{12}\}\), and . Let

$$\begin{aligned} U/A_{1}= & {} \{\{x_{1},x_{2},x_{3}\},\{x_{4},x_{5},x_{6}\},\{x_{7},x_{8},x_{9}\},\{x_{10},x_{11},x_{12}\}\};\\ U/A_{2}= & {} \{\{x_{1},x_{2},x_{3},x_{4},x_{5},x_{6}\},\{x_{7},x_{8},x_{9},x_{10},x_{11},x_{12}\}\};\\ U/A_{3}= & {} \{\{x_{1},x_{2},x_{3},x_{7},x_{8},x_{9}\},\{x_{4},x_{5},x_{6},x_{10},x_{11},x_{12}\}\};\\ U/A_{4}= & {} \{\{x_{1},x_{2},x_{7},x_{9}\},\{x_{3},x_{6},x_{8},x_{11}\},\{x_{4},x_{5},x_{10},x_{12}\}\};\\ U/A_{5}= & {} \{\{x_{1},x_{3},x_{7},x_{9}\},\{x_{2},x_{5},x_{9},x_{12}\},\{x_{4},x_{6},x_{10},x_{11}\}\};\\ U/A_{6}= & {} \{\{x_{1},x_{3},x_{8},x_{9}\},\{x_{2},x_{5},x_{7},x_{10}\},\{x_{4},x_{6},x_{11},x_{12}\}\};\\ U/A_{7}= & {} \{\{x_{1},x_{4},x_{7},x_{10}\},\{x_{2},x_{3},x_{8},x_{9}\},\{x_{5},x_{6},x_{11},x_{12}\}\}. \end{aligned}$$

Then we have , and . Therefore, .

Next, the main results of this paper are shown in the following tables.

The main results with respect to optimistic multi-granulation lower and upper approximations

Number

Optimistic multi-granulation lower approximations

Optimistic multi-granulation upper approximations

1

2

3

4

View full size image

View full size image

The main results with respect to pessimistic multi-granulation lower and upper approximations

Number

Pessimistic multi-granulation lower approximations

Pessimistic multi-granulation upper approximations

1

2

3

4

View full size image

View full size image

Remark

In the above two tables, symbols mean the optimistic multi-granulation lower approximations of \(X\subseteq U\) with respect to , respectively. Other symbols in tables indicate similar meanings.

7 An example

As an application of several types of attribute reducts proposed in the paper, a real-world example of a multi-granulation information system with respect to exam scores is employed. Then, as many researchers are working on various optimization theories Xu et al. (2017) and algorithms (Abualigah et al. 2018a, b, c, d, 2017a, b, c; Abualigah and Khader 2017d; Abualigah et al. 2017e; Abualigah and Hanandeh 2015; Al-Betar and Abualigah 2017), we will explain how to choose the optimal attribute reduct so as to compute the multi-granulation lower and upper approximations more efficiently.

Example 7.1

The following table indicates a multi-granulation information system with respect to exam scores, \(U=\{x_{1},x_{2},\ldots ,x_{30}\}\) is a universe including thirty students of mechanical engineering in Jimei University, is a family of attribute sets, where \(A_{1}=\{a_{1}^{1},a_{2}^{1},a_{3}^{1}\}\), and \(a_{1}^{1},a_{2}^{1},a_{3}^{1}\) stands for three courses: Advanced Mathematics, Linear Algebra, Probability and Statistics, respectively; \(A_{2}=\{a_{1}^{2},a_{2}^{2}\}\), and \(a_{1}^{2},a_{2}^{2}\) stands for two courses: College English 1, College English 2, respectively; \(A_{3}=\{a_{1}^{3},a_{2}^{3},a_{3}^{3}\}\), and \(a_{1}^{3},a_{2}^{3},a_{3}^{3}\) stands for three courses: Engineering Drawing, Engineering Mechanics, Feedback Control of Dynamic Systems, respectively; \(A_{4}=\{a_{1}^{4},a_{2}^{4}\}\), and \(a_{1}^{4},a_{2}^{4}\) stands for two courses: Programming in C, Computer Practice, respectively; \(A_{5}=\{a_{1}^{5}\}\), and \(a_{1}^{5}\) stands for College Physics; \(A_{6}=\{a_{1}^{6},a_{2}^{6}\}\), and \(a_{1}^{6},a_{2}^{6}\) stands for two courses: College PE 1, College PE 2, respectively; \(A_{7}=\{a_{1}^{7},a_{2}^{7}\}\), and \(a_{1}^{7},a_{2}^{7}\) stands for two courses: Experiment 1, Experiment 2, respectively. At the same time, the exam scores are divided into four grades: Good, Medium, Pass and Fail, denoted by 1, 2, 3 and 4, respectively.

An information system with respect to exam scores

U

\(~a_{1}^{1}\)

\(a_{2}^{1}\)

\(a_{3}^{1}\)

\(a_{1}^{2}\)

\(a_{2}^{2}\)

\(a_{1}^{3}\)

\(a_{2}^{3}\)

\(a_{3}^{3}\)

\(a_{1}^{4}\)

\(a_{2}^{4}\)

\(a_{1}^{5}\)

\(a_{1}^{6}\)

\(a_{2}^{6}\)

\(a_{1}^{7}\)

\(a_{2}^{7}\)

\(x_{1}\)

2

2

2

2

2

2

1

2

2

1

1

1

2

2

2

\(x_{2}\)

2

2

3

2

2

2

2

3

2

2

1

2

3

2

2

\(x_{3}\)

3

2

2

2

3

2

2

3

2

2

2

3

3

2

1

\(x_{4}\)

3

3

3

3

3

3

3

3

3

2

4

3

4

3

2

\(x_{5}\)

1

1

2

1

1

1

1

2

1

1

1

2

2

1

1

\(x_{6}\)

2

1

2

1

2

2

1

2

1

2

1

1

2

1

2

\(x_{7}\)

4

3

4

3

2

3

2

3

3

3

3

3

3

3

3

\(x_{8}\)

3

3

3

3

3

3

3

3

3

2

4

3

4

3

2

\(x_{9}\)

2

2

3

2

2

2

2

3

2

2

1

3

3

2

2

\(x_{10}\)

1

1

1

1

1

1

1

1

1

1

1

1

2

1

1

\(x_{11}\)

2

2

3

2

2

2

2

3

2

2

1

2

3

2

2

\(x_{12}\)

2

3

3

2

3

3

2

3

2

3

2

3

3

2

3

\(x_{13}\)

2

1

1

2

2

1

1

1

1

2

1

2

3

2

2

\(x_{14}\)

2

3

4

3

2

3

3

4

3

3

3

2

2

3

3

\(x_{15}\)

2

3

3

2

3

3

2

3

2

3

2

3

4

2

3

\(x_{16}\)

2

3

2

3

3

2

3

3

2

3

3

3

3

2

3

\(x_{17}\)

1

1

2

1

2

1

1

2

1

1

1

2

2

1

1

\(x_{18}\)

2

1

2

1

2

2

1

2

1

2

1

1

2

1

2

\(x_{19}\)

3

2

2

2

3

2

2

3

2

2

2

3

3

2

1

\(x_{20}\)

2

2

2

2

2

2

1

2

2

1

1

2

3

2

2

\(x_{21}\)

1

1

1

1

1

1

1

1

1

1

1

1

2

1

1

\(x_{22}\)

2

3

2

3

3

2

3

3

2

3

3

3

4

2

3

\(x_{23}\)

1

1

2

1

2

1

1

2

1

1

1

2

2

1

1

\(x_{24}\)

2

2

3

2

2

2

2

3

2

2

1

2

3

2

2

\(x_{25}\)

2

3

4

3

2

3

3

4

3

3

3

1

2

3

3

\(x_{26}\)

3

2

2

2

3

2

2

3

2

2

2

3

3

2

1

\(x_{27}\)

2

1

2

1

2

2

1

2

1

2

1

2

3

1

2

\(x_{28}\)

2

3

3

2

3

3

2

3

2

3

2

3

4

2

3

\(x_{29}\)

2

2

2

2

2

2

1

2

2

1

1

1

2

2

2

\(x_{30}\)

2

2

2

2

2

2

1

2

2

1

1

2

3

2

2

From the above information system, we have that \(U/A_{1}=\{\{x_{1},x_{20},x_{29},x_{30}\},\{x_{2},x_{9},x_{11},x_{24}\},\{x_{3},x_{19},x_{26}\},\{x_{4},x_{8}\}, \{x_{5},x_{17},x_{23}\},\{x_{6},x_{18},x_{27}\},\{x_{7}\}, \{x_{10},x_{21}\},\{x_{12},x_{15},x_{28}\},\{x_{13}\},\{x_{14},x_{25}\},\{x_{16},x_{22}\}\}\). Similarly, \(U/A_{i}, i=2,3,4,5,6,7\) can be easily presented.

According to Algorithms 1, 2 and 3, we have , , . Meanwhile, it is clear that , , and . Therefore, we have seven attribute reducts with respect to . i.e., , , , , , , and .

From the view of granular computing, we will raise a question: which reduct is the best selection to compute the multi-granulation lower and upper approximations? On the one hand, because the optimistic multi-granulation lower and upper approximations with respect to , , , and , respectively, are same, and from above discussion, it can be found that . Therefore, , , and are all the optimal selections to compute the optimistic multi-granulation lower and upper approximations.

On the other hand, we can obtain the following relations: and , although the pessimistic multi-granulation lower and upper approximations with respect to , , and , respectively, are same. Thus is the optimal attribute reduct to compute the pessimistic multi-granulation lower and upper approximations.

8 Conclusion

In this part, we first introduce the main conclusions obtained in the paper. Then, we make further prospects for future research work.

(i) Main conclusions of our paper It is clear that reduct theory plays an important role in pattern recognition and machine learning. Although multi-granulation rough set theory has been widely studied, attribute reduct of multi-granulation rough sets has not yet been explored. Therefore, the main content of this paper is to investigate the attribute reduct theory of multi-granulation information system. Specifically, the following problems are solved in this paper. Based on the definitions of multi-granulation upper and lower approximations, three types of attribute reducts are proposed, which are called arbitrary union reduct, neighborhood union reduct and neighborhood intersection reduct, respectively. Then, many important and interesting properties of these attribute reducts are researched. In order to solve practical problems, three effective algorithms are designed. At the same time, the relationships among these three attribute reducts are deeply explored. At last, a real-world example of multi-granulation information system is employed. The multi-granulation information system consists of exam scores of thirty students of mechanical engineering in Jimei University. According to the given algorithms, all the reducts in the information system are calculated one by one. Meanwhile, how to select the optimal reduct is also discussed in detail.

(ii) Future research work The theoretical results in this paper establish a basis for studying attribute reducts of the multi-granulation information systems. On this basis, it is possible to further study the attribute reducts with respect to consistent and inconsistent multi-granulation information systems. Furthermore, we need to develop better algorithms to deal with practical issues in the future.