Abstract
A fuzzy partition is usually described by a matrix U with elements from the unit interval. For a variety of reasons, U is often approximated by simpler matrices. The most popular are the maximum membership approximation and approximations based on α-level cuts of U. Because they create crisp matrices with membership grades either zero or one, most of the detailed information from U is lost. Two methods of gradual non-crisp approximations of U based on a set of parameters T derived from U are proposed. The first produces T-sharper images of U with gradually decreasing amount of fuzziness. The final image is a convex combination of all possible maximum membership approximations of U. The second method generates T-level cuts of U. They represent a gradual transformation from the lower crisp approximation of U by its core (α-cut at level α = 1) to the upper crisp approximation of U by its support (strong α-cut at level α = 0). Properties of both methods are discussed and illustrated with examples.
Access provided by CONRICYT-eBooks. Download conference paper PDF
Similar content being viewed by others
1 Introduction
The relevance of theory of fuzzy sets to cluster analysis has been well documented by many research works and applications [1, 8, 10–12, 14, 16]. A fuzzy partition U of n objects into c clusters is usually described by a matrix of the size c × n with elements from the unit interval. The use of the unit interval allows to express membership grades of U with infinitely many values from 0 to 1. However, this detailed information is not always needed. A large number of different membership grades complicate interpretation and increase computational cost in further applications. In decision-making, an approximation of U often provides sufficient information. There has been extensive research related to approximation of fuzzy sets [3–5, 9, 13, 15]. However, approximation of fuzzy partitions has not received considerable attention. Some work in this direction can be found in [1, 2, 6–8].
In practice, an approximation of a fuzzy partition U is usually performed by the maximum membership method [1] or by an α-cut [5, 15] of U, where α is a real number from the unit interval. Both methods produce crisp matrices with elements from the set {0, 1} and therefore most of the detailed information provided by the original fuzzy partition is lost. The aim of this contribution is to explore some methods of gradual approximations of U by non-crisp matrices. Two approaches, based on two different interpretations of the notion of α-cut of U, are considered. First, the α-cut can be interpreted as a sharper approximation of U in which the large membership grades (at least as large as α) are elevated to 1, while the small membership grades (smaller than α) are reduced to 0. Second, it may be viewed as a coarser approximation of U in which the total number of distinct membership grades of U is reduced to only two values, namely to 0 and 1. In this case, 0 and 1 are the labels for the image of U in the partition of the unit interval into subintervals [0, α) and [α, 1], respectively.
The maximum membership approximation of U is a generalization of the α-cut of U, where the parameter α is replaced by an n-tuple of parameters T = (t 1, …, t n ) derived from U. Therefore, the concept of α-sharpness can be extended to T-sharpness. Also, the maximum membership approximation of U can be obtained from a sequence of its minimum membership approximations [2]. T-sharper approximations and the minimum membership approximation based on T-sharpness will be explored in Sect. 3. Analogously, when the set T = { α} is extended to the set T = {t 1, …, t m } of parameters from (0, 1), the α-cut of U is generalized to the T-level cut of U, which is a coarser image of U in the partition of the unit interval into at most m + 1 subintervals. T-level cuts will be studied in Sect. 4. T-sharper approximations and T-level cuts of a fuzzy partition U represent a compromise between the original matrix U (too fuzzy) and its crisp approximation (too coarse). The parameters in the n-tuple T or in the set T can be derived from U or can be given by a researcher. The rest of this paper is organized as follows. Section 2 reviews some basic concepts of fuzzy partitions. Section 5 provides some concluding remarks.
2 Preliminaries
We assume that X is a finite universal set with the cardinality | X | = n and 2 ≤ c ≤ n. Let V cn denote the set of all real c × n matrices. Then the hard (crisp) c-partition space for X is the set [1]
the probabilistic fuzzy c-partition space for X is the set
and the possibilistic fuzzzy c-partition space for X is the set
The degenerated partition spaces M hcno , M fcno , and M pcno are supersets of M hcn , M fcn , and M pcn , respectively, obtained by the condition
The notation I will be used for the set of integers {1, …, c} and the notation J will be used for the set of integers {1, …, n}. For U, W ∈ V cn ,
Fuzziness of a fuzzy partition U ∈ M fco can be evaluated by the partition entropy defined by
where h(u ij ) = u ij log 1∕c (u ij ), when u ij > 0 and h(0) = 0. Then
-
1.
0 ≤ H(U) ≤ 1
-
2.
H(U) = 0 if and only if U ∈ M hco
-
3.
H(U) = 1 if and only if U has all elements equal to 1∕c
For α ∈ [0, 1], the α-cut of U ∈ M fco is the matrix U α ∈ V cn with elements
The strong α-cut of U is the matrix \(U_{\alpha ^{+}} \in V _{cn}\) with elements
When α = 1, then U α is the core of U, denoted by cor U, and when α = 0, then \(U_{\alpha ^{+}}\) is the support of U, denoted by supp U.
The maximum membership approximation of U ∈ M fco is the partition MM(U) ∈ M fco defined for all (i, j) ∈ I × J by
Note that U may have several different maximum membership approximations if the maximum membership of a classified object in U is not unique.
3 T-Sharper Approximations of Fuzzy Partitions
Sharpening of membership grades of a fuzzy set is obtained by transformation of large membership grades closer to 1 and small membership grades closer to zero. The notions of large and small are understood with respect to a given threshold α ∈ [0, 1]. The following generalization of α-sharpening can be considered in the case of fuzzy partitions.
Definition 1
Consider U, W ∈ M fco and n-tuple T = (t 1, …, t n ), t j ∈ [0, 1] for all j ∈ J. Then W is T-sharper than U, denoted by W ≺ T U, if the membership grades of W satisfy the following properties: for all (i, j) ∈ I × J,
Further in this paper, the notation T = (t j ) will be used instead of T = (t 1, …, t n ).
Remark 1
If T = (α), α ∈ [0, 1], then T-sharpness is reduced to α-sharpness. In order to obtain a T-sharper approximation of U which is a fuzzy partition different from U, the n-tuple T must contain at least one parameter t j such that
Proposition 1
Given T = (t 1, …, t n ) and U, V, W ∈ M fco , the following properties hold:
-
1.
U ≺ T U
-
2.
If U ≺ T V and V ≺ T U, then U = V
-
3.
If U ≺ T V and V ≺ T W, then U ≺ T W
Corollary 1
M fco is partially ordered by the relation ≺ T .
T-sharpening of fuzzy partitions reduces their fuzziness and therefore it can be considered as a method of partial defuzzification.
Proposition 2
Let U, W ∈ M fco and W ≺ T U. Then H(W) ≤ H(U).
Note that if H(W) ≤ H(U), then W does not need to be a T-sharper approximation of U.
The notion of T-sharpness can be used in characterization of the minimum membership approximation of a fuzzy partition. Recall that the method of the maximum membership generates a crisp approximation of U, where the maximum membership grade in each column of U is replaced by 1. On the contrary, the minimum membership grade approximation should be a fuzzy partition where the minimum membership grade in each column of U is replaced by 0. Since this approximation should reduce the partition entropy of U, we request that the minimum membership approximation of U be T-sharper than U.
Definition 2
Assume U ∈ M fco . For each j ∈ J, let s j = max{u ij |, i ∈ I} and r j = min{u ij | u ij > 0, i ∈ I}. Create matrix W ∈ V cn as follows: for all (i, j) ∈ I × J
where ∑ i = 1 c w ij = 1 for all j ∈ J and W ≺ T U for some n-tuple T. Then W is a minimum membership approximation of U denoted by mm(U).
Note that minimum membership approximation of U is not unique.
Proposition 3
Assume U ∈ M fco . For each j ∈ J, let s j = max{u ij |, i ∈ I}, r j = min{u ij | u ij > 0, i ∈ I}, I j = {i ∈ I: u ij = r j }, and I j ∗ = {i ∈ I: u ij > r j }. Consider mapping φ 1: M fco → M fco which assigns to each u ij ∈ U element φ 1(u ij ) ∈ φ 1(U) as follows:
where γ j = r j | I j | ∕ | I j ∗ | . Then φ 1(U) = mm(U) and φ 1(U) ≺ T U, where T = (r j ).
Another way of obtaining a minimum membership approximation of U is based on linear intensification [2].
Proposition 4
Assume U ∈ M fco . For each j ∈ J, let s j = max{u ij |, i ∈ I}, r j = min{u ij | u ij > 0, i ∈ I}, I j = {i ∈ I | u ij > 0}, and m j = | I j | . Consider the mapping (linear intensification) φ 2: M fco → M fco which assigns to each u ij ∈ U element φ 2(u ij ) ∈ φ 2(U) as follows:
Then φ 2(U) = mm(U) and φ 2(U) ≺ T U, where T = (1∕m j ).
A slight modification of φ 1 from Proposition 3 also generates mm(U).
Proposition 5
Assume U ∈ M fco . For each j ∈ J, r j = min{u ij | u ij > 0, i ∈ I}, I j = {i ∈ I | u ij = r j }, s j = max{u ij , i ∈ I}, and I j ∗∗ = {i ∈ I | u ij = s j }. Consider the mapping φ 3: M fco → M fco which assigns to each u ij ∈ U element φ 3(u ij ) ∈ φ 3(U) as follows:
where γ j = r j | I j | ∕ | I j ∗∗ | . Then φ 3(U) = mm(U) and φ 3(U) ≺ T U, where T = (r j ).
A sequence of gradual approximations of U by mm(U) can be created according to Algorithm 1 below.
Algorithm 1
Let U ∈ M fco and φ: M fco → M fco such that φ(U) = mm(U).
- Step 1: :
-
Put k: = 0 and mm k(U): = U.
- Step 2: :
-
Create mm k+1(U) = φ(mm k(U)).
- Step 3: :
-
If mm k+1(U) = mm k(U) or k = c − 2, stop. Else put k: = k + 1 and go to Step 2.
Proposition 6
Assume a sequence mm k(U), k = 0, …, c − 2 of approximations of U generated by Algorithm 1 . Then \(mm^{k+1}(U) \prec _{T_{k}}mm^{k}(U)\) and T k+1 ≥ T k .
Corollary 2
Assume a sequence mm k(U), k = 0, …, c − 2 of approximations of U generated by Algorithm 1 . Then H(mm k+1(U)) ≤ H(mm k(U)).
Remark 2
When U has in each column only non-repeated positive membership grades, then
-
Fuzzy partition mm k(U), k = 1, …, c − 1 created by Algorithm 1 is the kth minimum membership approximation of U. Consequently, mm c−1(U) = MM(U),
-
If φ 2 from Proposition 4 is used in Algorithm 1, then \(mm^{k+1}(U) \prec _{T_{k}}mm^{k}(U),\;k = 0,\ldots,c - 2\), where T k = (1∕(c − k)) and hence T 0 = (1∕c) and T c−1 = (1∕2).
Each mapping φ: M fco → M fco which creates a minimum membership approximation of a fuzzy partition U reduces the number m j of clusters to which object x j ∈ X is classified (2 ≤ m j ≤ c). Then, in decision-making, the following guidelines can be used for the choice of φ.
-
If there is no other condition placed on mm(U), use the mapping φ 1 from Proposition 6. In this case, membership grades of x j greater than the minimal one are uniformly elevated.
-
If the goal is to create a fuzzy partition mm(U) such that the membership grades of x j distributed to m j clusters of U are sharpened with respect to the threshold 1∕m j , use the mapping φ 2 from Proposition 4.
-
If the goal is to create a fuzzy partition mm(U) with more distinct maximum membership grades, use the mapping φ 3 from Proposition 5. Note that φ 3 can be considered as a combination of the minimum membership and the maximum membership methods, because the minimal membership grade of x j is reduced to zero and only its maximal membership grade is elevated.
Example 1
Let U be the fuzzy partition of elements from the set X = {x 1, x 2, x 3, x 4, x 5, x 6} represented by the matrix
Due to repeated maximum membership grades in column 1 and column 5, there are 8 different maximum membership approximations of U. When Algorithm 1 with the mapping φ 1 from Proposition 6 is applied to U, the following non-crisp approximations are obtained. First,
where \(mm^{1}(U) \prec _{T_{0}}(U)\) and T 0 = (0. 03, 0. 15, 0. 1, 0. 12, 0. 25, 0. 09).Partition entropy H(U) = 0. 821 was reduced to H(mm 1(U)) = 0. 643.The second and the third approximations are
where \(mm^{3}(U) \prec _{T_{2}}mm^{2}(U) \prec _{T_{1}}mm^{1}(U),\) T 1 = (0. 12, 0. 375, 0. 3, 0. 34, 0. 25, 0. 24) and T 2 = (0. 5, 1, 1, 1, 0. 25, 0. 45),H(mm 2(U)) = 0. 333 and H(mm 3(U)) = 0. 250.
When Algorithm 1 with the mapping φ 2 from Proposition 4 is applied to U, the following non-crisp approximations are obtained. First,
where \(mm^{1}(U) \prec _{T_{0}}(U)\) and T 0 = (1∕4, 1∕3, 1∕4, 1∕3, 1∕4, 1∕4).Partition entropy H(U) = 0. 821 was reduced to H(mm 1(U)) = 0. 614.The second and the third approximations are
where \(mm^{3}(U) \prec _{T_{2}}mm^{2}(U) \prec _{T_{1}}mm^{1}(U),\) T 1 = (1∕3, 1∕2, 1∕2, 1∕2, 1∕4, 1∕3) and T 2 = (1∕2, 1, 1, 1, 1∕4, 1∕2),H(mm 2(U)) = 0. 325 and H(mm 3(U)) = 0. 250. When Algorithm 1 with the mapping φ 3 from Proposition 5 is applied to U, the following non-crisp approximations are obtained. First,
where \(mm^{1}(U) \prec _{T_{0}}(U)\) and T 0 = (0. 03, 0. 15, 0. 1, 0. 12, 0. 25, 0. 09).Partition entropy H(U) = 0. 821 was reduced to H(mm 1(U)) = 0. 627.The second and the third approximations are
where \(mm^{3}(U) \prec _{T_{2}}mm^{2}(U) \prec _{T_{1}}mm^{1}(U),\) T 1 = (0. 27, 0. 3, 0. 2, 0. 28, 0. 25, 0. 21) and T 2 = (0. 5, 1, 1, 1, 0. 25, 0. 45),H(mm 2(U)) = 0. 333 and H(mm 3(U)) = 0. 250.
4 T-Level Cuts of Fuzzy Partitions
Another way of approximation of a fuzzy partition U is transformation of U to a matrix with reduced number of distinct membership grades from interval (0, 1). This type of transformation is called coarsening. Given U ∈ M fco , the set
is called the level set of U. Obviously, the level set of a crisp partition is empty.
Definition 3
Consider a matrix W ∈ V cn with elements from [0, 1]. The coarseness of W is evaluated by the coefficient
Proposition 7
Assume a fuzzy partition U ∈ M fco . Then
-
1.
0 ≤ κ(U) ≤ 1
-
2.
κ(U) = 1 if and only if U ∈ M hco (crisp partition)
-
3.
κ(U) = 0 if and only if all membership grades of U are non-repeated values between zero and one
An α-cut of U can be considered as a coarser image of U in the partition of the unit interval into subintervals [0, α) and [α, 1], or simply as the level cut of U based on the set of parameters T = { α}, α ∈ (0, 1). Let {δ 1, δ 2, …, δ k } be the set of all distinct membership grades of U such that 1 > δ 1 > δ 2 > … > δ k > 0. Then the level cuts \(U_{\delta _{1}},U_{\delta _{2}},\ldots,U_{\delta _{k}}\) create a sequence of gradual crisp approximations of U such that
Our goal is to find a sequence of gradual non-crisp coarser approximations of U satisfying inequality (18). Based on the concept of p-level sets [3], the concept of T-level cuts of a fuzzy partition U is introduced in the next definition.
Definition 4
Let U ∈ M fco and T = {t 1, t 2, …, t m } be a set of parameters such that 0 < t 1 ≤ t 2 ≤ … ≤ t m < 1. Then the T-level cut of U is the matrix T(U) ∈ V cn defined for all (i, j) ∈ I × J as follows:
The membership grade T(u ij ) evaluates the proportion of parameters from T which are dominated by u ij . If T includes r distinct parameters from (0, 1), then T(U) is an image of U in the partition of the unit interval into r + 1 subintervals. In general, T(U) is a possibilistic fuzzy partition.
Proposition 8
Let U ∈ M fco . Assume a T -level cut T(U) with the membership grades T(u ij ). Then
-
1.
T(0) = 0 and T(1) = 1
-
2.
T(u ij ) ∈ {0, 1∕m, 2∕m, …, (m − 1)∕m, 1}, where m = | T |
-
3.
If u ij ≤ u pq then T(u ij ) ≤ T(u pq )
-
4.
When T = {t 1, t 2, …, t m }, S = {s 1, s 2, …, s m }, and t j ≤ s j for all j = 1, …, m, then S(U) ≤ T(U)
-
5.
If T = { α}, α ∈ (0, 1), then \(\mathbf{T}(U) = U_{\alpha ^{+}}\)
Corollary 3
Let | Λ U | = m. If the set T has less than m parameters (not necessarily distinct), then T(U) is a coarser approximation of U.
A sequence of gradual approximations of U by coarser T-level cuts of U can be created according to Algorithm 2 below.
Algorithm 2
Consider U ∈ M fco .
- Step 1: :
-
Let k be the number of crisp columns in U (i.e., columns containing elements equal only to zero and one). Put n ∗: = n − k.
- Step 2: :
-
Remove from U all crisp columns and arrange the elements in each remaining column from the largest to the smallest. Denote the resulting matrix by U ∗. Let c ∗ be the total number of distinct rows in U ∗ with all elements from (0, 1).
- Step 3: :
-
For q = 1, …c ∗, arrange the elements in the qth row of U ∗ from the smallest to the largest. Denote the resulting matrix by U ∗∗. Assign the qth-row of U ∗∗ to n ∗-tuple T q and create T q (U).
Proposition 9
Consider the T -level cuts of U ∈ M pc created by Algorithm 2 . Then
-
1.
\(cor\;U < \mathbf{T}_{1}(U) < \mathbf{T}_{2}(U) < \mathbf{T}_{c^{{\ast}}}(U) < supp\;U\)
-
2.
κ(T q (U)) < κ(U) for all q = 1, …, c ∗
-
3.
T q (U) ∈ M pco for q ≥ 2
Remark 3
T 1(U) created by Algorithm 2 contains one column with all elements equal to zero (the column, where the smallest from all n maximal membership grades is located). Therefore, T 1(U) is not a possibilistic fuzzy partition of n objects.
Example 2
Consider the fuzzy partition U from Example 1. Then Algorithm 2 applied to U creates the following matrix U ∗
Then T 1 = {0. 25, 0. 35, 0. 4, 0. 55, 0. 6, 0. 6} and
Coefficient of coarseness κ(U) = 0. 375 is increased to κ(T 1(U)) = 0. 833.The second and the third approximations are
with T 2 = {0. 2, 0. 25, 0. 28, 0. 3, 0. 3, 0. 35} and T 3 = {0. 1, 0. 12, 0. 15, 0. 21, 0. 25, 0. 27}, and κ(T 2(U)) = κ(T 1(U)) = 0. 833. Obviously,
5 Conclusion
Two methods of a gradual non-crisp approximation of a fuzzy partition U based on parameters which may be derived from U were proposed. While the first method focuses on reduction of fuzziness of U, the second method increases coarseness of U. The relation of T-sharpness and the coefficient of coarseness were introduced. Although the work in this paper refers to a fuzzy partition of n objects into c clusters, it can be applied to any situation where information is represented by a matrix with elements from the unit interval. This includes, e.g., tables of relative frequencies or normalized evaluations of n objects by c experts. In future work, approximations of fuzzy partitions based on combination of sharpening and coarsening will be studied.
References
Bezdek, J.C.: Pattern Recognition with Fuzzy Objective Function Algorithms. Plenum Press, New York (1981)
Bodjanova, S.: Linear intensification of probabilistic fuzzy partitions. Fuzzy Sets Syst. 141, 319–332 (2004)
Bodjanova, S., Kalina, M.: Coarsening of fuzzy sets. In: Proceedings of IEEE 11th International Symposium on Intelligent Systems and Informatics, pp. 51–56. IEEE, New York (2013)
Deng, X., Yao, Y.: Decision-theoretic three-way approximations of fuzzy sets. Inf. Sci. 279, 702–715 (2014)
Dubois, D., Prade, H.: Fundamentals of Fuzzy Sets. Kluwer, Dordrecht (2000)
Dumitrescu, D., Pop, H.F.: Degenerate and non-degenerate convex decomposition of finite fuzzy partitions - I. Fuzzy Sets Syst. 73, 365–376 (1995)
Dumitrescu, D., Pop, H.F.: Degenerate and non-degenerate convex decomposition of finite fuzzy partitions - II. Fuzzy Sets Syst. 96, 111–118 (1998)
Dumitrescu, D., Lazzerini, B., Jain, L.C.: Fuzzy Sets and Their Applications to Clustering and Training. CRP Press, Boca Raton (2000)
Li, T.J., Zhang, W.X.: Rough fuzzy approximations on two universes of discourse. Inf. Sci. 178, 892–906 (2008)
Maji, P., Pal, S.: Rough set based generalized fuzzy c-means algorithm and quantitative indices. Trans. Syst. Man Cybern. Part B 37, 1529–1540 (2007)
Mitra, S., Banka, H., Pedrycz, W.: Rough-fuzzy collaborative clustering. Trans. Syst. Man Cybern. Part B 37, 795–805 (2005)
Mitra, S., Pedrycz, W., Barman, B.: Shadowed c-means: Integrating fuzzy and rough clustering. Pattern Recogn. 43, 1282–1291 (2010)
Pedrycz, W., Dong, F., Hirota, K.: Finite cut-based approximation of fuzzy sets and its evolutionary optimization. Fuzzy Sets Syst. 160, 3550–3564 (2009)
Peters, G., Crespo, F., Lingras, P., Weber, R.: Soft clustering-fuzzy and rough approaches and their extensions and derivatives. Int. J. Approx. Reason. 54, 307–322 (2013)
Radecki, T.: Level-fuzzy sets. J. Cybern. 7, 189–198 (1977)
Zhou, J., Pedrycz, W., Miao, D.Q.: Shadowed sets in the characterization of rough-fuzzy clustering. Pattern Recogn. 44, 1738–1749 (2011)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Bodjanova, S. (2017). T-Sharper Images and T-Level Cuts of Fuzzy Partitions. In: Palumbo, F., Montanari, A., Vichi, M. (eds) Data Science . Studies in Classification, Data Analysis, and Knowledge Organization. Springer, Cham. https://doi.org/10.1007/978-3-319-55723-6_5
Download citation
DOI: https://doi.org/10.1007/978-3-319-55723-6_5
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-55722-9
Online ISBN: 978-3-319-55723-6
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)