1 Introduction

As an essential branch of machine learning, clustering analysis aims to gather high similarity data samples into the same group. As an unsupervised learning algorithm, clustering has been widely used in many fields, such as image segmentation [1], evaluation of credit risk prediction [2], and pattern recognition [3]. In various clustering algorithms [4,5,6,7], the fuzzy c-means clustering (FCM) proposed by Bellman et al. [8] can integrate the uncertainty of the actual datasets by combining Zada’s fuzzy theory [9]. The use of fuzzy information is mostly driven by the ability to understand operations in a manner akin to human logical thinking, which can capture more information about actual problems [10]. In FCM clustering, the interaction between different clusters is generated by FCM, which can effectively avoid falling into the local optimal solution [11, 12]. Due to the uncertainty in data collection in practical problems, FCM may experience uncertainty when calculating the membership value of a given sample [13]. In other words, due to the fact that fuzzy theory only obtains uncertain information through membership functions in expressing fuzzy information, this can result in the loss of some fuzzy information [14]. Therefore, FCM has certain limitations in comprehensively obtaining uncertain information [15].

In order to improve the problem of fuzzy sets being unable to obtain more uncertain information, Atanassov extended Zada’s fuzzy theory and proposed the intuitionistic fuzzy set [16]. The IFS uses membership and non-membership functions to describe fuzzy information based on fuzzy sets, to avoid information loss. Scholars also mentioned that IFS allows for correct modeling of the problem based on available data and observation [17]. Given the powerful ability of intuitive fuzziness to capture uncertain information, it has been widely applied in multiple fields, such as fuzzy multi-attribute decision-making problems [18], classification problems [19, 20], forecasting problems [21], and others. Due to the fact that IFSs integrated the concepts of non-membership and hesitation degree in addition to membership in the datasets, which better represents the inherent uncertainty of datasets, IFSs are expanded to FCM [15, 22, 23]. Compared with FCM, IFCM has been proven to converge to better positions and have higher performance in some problems [22, 23]. The IFCM, as a version of FCM, computes the partition matrix or membership matrix by determining the membership value of each data point to join in a cluster, and the cluster centroids are also initialized randomly [15]. Therefore, although IFCM is an improvement for FCM in expressing uncertain information, its initial clustering center still heavily relies on the clustering center [24]. In response to this drawback, the current research mainly focuses on identifying the initial clustering center by obtaining density information on data sample distribution. This idea is first extended to FCM to verify the impact of considering sample distribution density on algorithm performance. Currently, most density-based improved algorithms are implemented by obtaining the initial cluster center by the data points with high density based on cut-off distance [25,26,27,28]. To improve the impact of the initial cluster center on the performance of the IFCM algorithm, Varshney et al. [15] also extended the calculation density based on the cut-off distance to IFCM and proposed the density-based IFCM algorithm. The cut-off distance dc∈[0, 1] is defined as a random constant value, which is chosen experimentally. In addition, when calculating the density, the density rate λ, distance rate σ and other parameters need to be adjusted, which makes it difficult to calculate the cut-off density. The above sample density values are unfortunately overdependent on the cut-off distance and heavily affected by noise points.

Apart from the above mentioned shortcomings, the traditional algorithm converges too slowly because all the samples will always affect all the cluster centers [29]. Therefore, scholars have made great efforts to improve the convergence speed of the traditional algorithm. The current improvements can be divided into two types, including improvements to traditional algorithms from the perspective of data features and algorithm principles. From the perspective of data features, the entropy measure [30], distance measure [25, 31], and probabilistic Euclidean distance [26] are extended to obtain the contributions of different features to the sample. For instance, Cherif et al. [27] proposed the three new interval type-2 fuzzy similarity measures and joined with fuzzy C-means algorithm. An intuitionistic kernelized total Bregman divergence is proposed to measure the difference and the weighted local information is introduced into the objective function [28]. Improvements from the perspective of algorithm principles include using advanced similarity measures instead of Euclidean distance measures when calculating membership, and incorporating heuristic algorithms to avoid falling into local optimization solutions [32, 33]. In addition, some scholars have improved traditional algorithms from two perspectives to accelerate their convergence speed on large datasets. For example, a fuzzy C-means algorithm for optimizing data clustering is proposed by incorporating the typicality function [34]. The random sampling plus extension FCM (rseFCM) obtains the final effective clustering results by taking random samples into the literal FCM (LFCM). Starting from the viewpoint that all samples always affect all clustering centers, Zhou et al. [24] proposed a new membership compression method to achieve fast clustering by scaling membership. For the separation and processing of information, Joaquín Pérez et al. [35] applied the equidistance index (EI) to obtain statistical information about the displacement of centroids at each iteration for reducing the number of calculations needed in the classification step of hard c-means, without significant loss of quality reduction. The algorithm proposed by Joaquín Pérez et al. [35] also verified that the data points can be divided by the EI, and can be achieved better performance. The extension of the EI to FCM is promising research.

According to the above analysis, although IFCM has the ability to express fuzzy information, its clustering performance is unstable because its initial clustering center is randomly selected. It is still a challenge for IFCM to obtain the cluster center in a simple way and ensure the clustering performance. In addition, the convergence of the IFCM algorithm is too slow, because all samples always affect all cluster centers. Therefore, another challenge of IFCM is to reduce the number of iterations while ensuring the clustering performance. In light of this, a novel algorithm named EI-IFCM is proposed to obtain the initial cluster center and reduce the number of iterations effectively from the local density and membership boundary of samples. Both theoretical and empirical studies indicate that EI-IFCM is clear, efficient, and flexible. The main contributions are summarized as follows:

  1. (1)

    A new strategy for obtaining initial cluster centers based on local density is proposed. This strategy fully considers the two characteristics of easy calculation and the influence of noise data on the initial clustering center.

  2. (2)

    The EI condition is applied to the datasets division, which can provide relevant knowledge about whether the data sample center changes. The EI condition provides theoretical support for the division of data samples. To our knowledge, the study is among the earliest work that applies the equidistance index condition to the IFCM algorithm.

  3. (3)

    The boundary value of the membership degree of IFCM is derived. It can increase the membership value of the data samples to the maximum value and terminate the subsequent iteration. The calculation of the membership boundary significantly saves iterations.

  4. (4)

    Extensive validation of the EI-IFCM on numerous real-world datasets has been done to demonstrate the superior performance of the algorithm, and the applicability of the algorithm has also been verified.

The rest of this paper is organized as follows. Section 2 presents some basic concepts and new findings of IFS and IFCM. The novel algorithm based on the local density and membership boundary of IFCM is presented in Section 3. In section 4, experiments and sensitivity analysis on some real-world datasets are given. The conclusion of the paper is shown in Section 5.

2 Preliminaries

In this section, the relevant concepts of IFS, IFCM and the equidistance index are reviewed. Based on these basic concepts, the membership boundary of IFCM is derived.

2.1 Intuitionistic fuzzy C means clustering

Atanassov [16] proposed the intuitionistic fuzzy set (IFS) based on membership μA(x) of the fuzzy set by adding the nonmembership νA(x). An IFS A defined on G is given as follows [16].

$$A=\left\{\langle x,{\mu }_{A}(x),{\nu }_{A}(x)\rangle |x\in G\right\}$$
(1)

where μA(x): G→[0,1], and νA(x): G→[0,1] with the condition 0≤μA(x)+νA(x)≤1. The hesitation degree of A is expressed as πA(x)=1-μA(x)-νA(x)≤1.

The intuitionistic fuzzy complement function is first defined by Bustince et al. [36]. Chaira [22] rewrote the intuitionistic fuzzy complement function as N(x)=(1-xα)1/α, where α>0, N(1)=0 and N(0)=1. Chaira [22] also calculated the non-membership of IFS by using the rewritten intuitionistic fuzzy complement function and gave the transformation form of IFS as follows.

$$A=\left\{x,{\mu }_{A}\left(x\right),{\left(1-{{\mu }_{A}\left(x\right)}^{\alpha }\right)}^{1/\alpha }|x\epsilon G\right\}$$
(2)

And the hesitation degree is:

$${\pi }_{IFS}={1-\mu }_{A}\left(x\right){-\left(1-{{\mu }_{A}\left(x\right)}^{\alpha }\right)}^{1/\alpha }$$
(3)

Clustering is described as the process of obtaining different clusters by calculating the membership of data sample points to multiple cluster centers. The IFCM algorithm is an objective function-based clustering algorithm that may optimize the objective function to determine each data sample’s membership degree with each clustering center, hence attaining the purpose of automatic clustering. Suppose X={x1, x2,…, xn} is the observation sample set and the features of each sample are IFSs in the s-dimension. The objective function of the intuitionistic IFCM algorithm is as follows [23]:

$${J}_{{}^{m}}\left(U,V;X\right)=\sum_{i=1}^{c}\sum_{j=1}^{n}{\left({\mu }_{ij}^{(t)}\right)}^{m}\times {\Vert {x}_{j},{v}_{i}^{(t)}\Vert }^{2}$$
(4)

where c is the number of clusters, m is a fuzzy parameter, \(||\boldsymbol{x}_{j}, \boldsymbol{v}^{(t)}_{i}||^{2}\) is Euclidean distance measure between vi (cluster center) and xj (data points), and μij is the membership value of jth data (xj) in ith cluster.

To minimize the Jm(U,V;X) by using an iterative process based on the following equation.

$${\mu }_{ij}^{(t+1)}\text{=}{\left({\sum }_{k=1}^{c}{\left[\frac{{\Vert {x}_{j},{\nu }_{i}^{(t+1)}\Vert }^{2}}{{\Vert {x}_{j},{\nu }_{k}^{(t+1)}\Vert }^{2}}\right]}^{\left(1/\left(m-1\right)\right)}\right)}^{-1}$$
(5)
$${\left({\mu }_{ij}^{(t+1)}\right)}^{*}\text{=}{\mu }_{ij}^{(t+1)}+{\pi }_{ij}^{(t+1)}$$
(6)

where \(\Big(\mu^{(t+1)}_{ij}\Big)^{*}\) denotes the membership of the jth data sample in ith cluster under t+1 iteration.

Based on the calculation of \(\Big(\mu^{(t+1)}_{ij}\Big)^{*}\), the clustering center can be updated using the following equation.

$${v}_{i}^{*}\text{=}\frac{{\sum }_{j=1}^{n}{\left({\mu }_{ij}^{(t+1)}\right)}^{*m}{x}_{j}}{{\sum }_{j=1}^{n}{\left({\mu }_{ij}^{(t+1)}\right)}^{*m}}$$
(7)

where i=1,2,..,c, j=1,2,…,n.

The cluster center and membership matrix are updated after each iteration, and the algorithm stops when \(\mathrm{max}_{ij}|\boldsymbol{U}^{*new}_{ij}- \boldsymbol{U}^{*prev}_{ij}|< \varepsilon\) is satisfied. ε is tolerance for the solution accuracy, which has already been set before implementing the clustering task.

2.2 Boundary value and equidistance index

Zhou et al. [29] obtained the membership degree bounds of FCM by rearranging the Euclidean distances from the cluster centers to the data samples in descending order. They designed the boundary value of the membership degree by using the following inequality [29] for the data sample xj to the nearest cluster center vi.

$$\frac{1}{c}\le {\mu }_{{I}_{j}^{*}j}^{(t+1)}\le {\left[1+(c-1){\left(\frac{{\left(\Vert {x}_{j},{\nu }_{i}\Vert \right)}^{(1)}}{{\left({d}_{j}^{(t+1)}\right)}^{(c)}}\right)}^{2/(m-1)}\right]}^{-1}$$
(8)

where \(\Big(d^{(t+1)}_{j}\Big)^{(c)}\) denotes the max distance from the jth data point xj to the ith cluster center \({\nu }_{i}^{(t+1)}\) in t+1 iteration, and \({I}_{j}^{*}={\text{arg}}\underset{1<i\le c}{{\text{min}}}\left\{{d}_{ij}^{(t+1)}\right\}\).

According to Eq. (8), the boundary value of the intuitionistic fuzzy membership degree of IFCM can be proposed in Lemma 1.

Lemma 1

For data sample xj, the boundary membership value of iteration t+1 can be calculated by formula (9).

$$0\le {\left({\mu }_{ij}^{(t+1)}\right)}^{*}\le 1-{\left(1-{\left[1+(c-1){\left({\Vert {x}_{j},{\nu }_{i}^{(t+1)}\Vert }^{2}/{\left({d}_{j}^{(t+1)}\right)}^{(c)}\right)}^{2/(m-1)}\right]}^{-\alpha }\right)}^{(1/\alpha )}$$
(9)

Suppose the nearest and the second nearest cluster centers of data object xj are \(\boldsymbol{v}_{1}\) and \(\boldsymbol{v}_{2}\), respectively. (d(t+1) j)(1) and (d(t+1) j)(2) are the Euclidean distance from \(\boldsymbol{x}_{j}\) to \(\boldsymbol{v}_{1}\) and \(\boldsymbol{v}_{2}\), respectively. The equidistance index (EI) can be expressed as the difference between the two distances. Let \(\boldsymbol{v}_{1}\) and \(\boldsymbol{v}_{2}\) be the nearest and the second nearest cluster centers of an object \(\boldsymbol{x}_{j}\), respectively. Then the EI can be defined as follows [35].

$${\alpha }_{j}^{(t+1)}=abs\left({\Vert {x}_{j}-{v}_{1}^{(t+1)}\Vert }^{2}-{\Vert {x}_{j}-{v}_{2}^{(t+1)}\Vert }^{2}\right)$$
(10)

It is worth noting that \(0\le {\alpha}^{(t+1)}_{j}\le || \boldsymbol{v}_{1}^{(t+1)} \hbox{-} \boldsymbol{v}_{2}^{(t+1)}||^{2}\). It is known that the EI of each data object xj will change as the iterative process proceeds. There is a high or low likelihood that the cluster of data object xj will be changed. The equidistance threshold helps to identify the objects with a high likelihood of cluster change. Then the offsets of the clustering center in two iterations, i.e. t and t+1 are \(m_{1}=||\boldsymbol{v}_{1}^{(t+1)}\hbox{-}\boldsymbol{v}_{1}^{t}||\) and \(m_{2}=||\boldsymbol{v}_{2}^{(t+1)} \hbox{-}\boldsymbol{v}_{2}^{t}||\), respectively. Then equidistance threshold β(t+1) is defined as follows [35].

$${\beta }^{(t+1)}={m}_{1}+{m}_{2}$$
(11)

From the above analysis, it can be seen that it is feasible to determine whether the cluster of a data object is going to change in the next iteration by comparing EI with the equidistance threshold. Therefore, the partition EI for an object xj can be summarized as follows:

  1. (1)

    If α(t+1) jβ(t+1), the object xj has a high likelihood of cluster change for the next iteration in clustering;

  2. (2)

    If α(t+1) j>β(t+1), the object xj has a low likelihood of cluster change in the next iteration.

Figure 1 can intuitively illustrate the division requirements of partition of EI condition for the object xm in the process of clustering iteration. As shown in the figure, it is assumed that the objects xm and xn are distributed at the positions, and v1 and v2 are the nearest and second nearest centers of these two objects, respectively. The parameter changes before and after iteration are shown in the figure. It can be seen from the figure that the center of xm is not easy to change, while the center of xn is likely to change.

Fig. 1
figure 1

Schematic diagram of partition of EI condition

3 Equidistance Index Intuitionistic Fuzzy C-Means (EI-IFCM) Clustering Algorithm

In this section, we propose a new algorithm named equidistance index intuitionistic fuzzy c-means (EI-IFCM) clustering algorithm by integrating the IFCM and EI. Figure 2 shows the implementation flow chart of the proposed EI-IFCM clustering algorithm. The proposed algorithm mainly includes two parts, namely, the acquisition of the initial cluster center and the update of the membership matrix. In the first part, the local density is fully considered to obtain the initial cluster center. The equidistance index segmentation is applied to the second part to ensure the fast convergence of the proposed algorithm.

Fig. 2
figure 2

Schematic diagram of the EI-IFCM algorithm implementation process

3.1 Acquisition of initial clustering centers

The traditional IFCM is sensitive to the selection of the initial clustering center. The cluster centers are often distributed in areas with dense data points, that is, the cluster centers have a large local density within the cluster range [37]. The initial cluster centers should satisfy the following conditions.

  1. (1)

    If an object vi is a cluster center, its local density is large;

  2. (2)

    If object vi is the cluster center of a cluster in the data set, the Euclidean distance between the object and the object with a higher local density than it must be large, that is, the object has a larger local density.

Therefore, Algorithm 1 is proposed to select the initial clustering centers in the EI-IFCM algorithm. In Algorithm 1 the global average distance of the dataset is obtained by using E.q (12).

Algorithm 1
figure g

Initial center selection

$$\overline{D }=\frac{1}{n(n-1)}{\sum }_{l=1}^{k}{\Vert {x}_{j},{x}_{l}\Vert }^{2}$$
(12)

And λ is the density factor that can affect the change of local density. The local density can be calculated using E.q (13).

$${\rho }_{i}={e}^{\frac{1}{{\sum }_{l=1}^{k}{\Vert {x}_{j},{x}_{l}\Vert }^{2}}}$$
(13)

where e is an exponential function that is applied to calculate the local density of the data sample point. The exponential function can eliminate the influence of noise data on the calculation of local density. Therefore, the exponential function is used to calculate the local density of the sample.

Algorithm 1 selects data sample points whose local distance is greater than the global average distance through a descending arrangement of local density. The purpose is to select a large local density and the distance corresponding to two large local density points is greater than or equal to the global average distance to satisfy the initial cluster center selection conditions.

Algorithm 2
figure h

EI-IFCM

3.2 Update of membership matrix

The IFCM algorithm updates the membership matrix by calculating the Euclidean distance between the cluster centers and all sample objects. It allows all samples to affect all cluster centers, which will result in some waste of computing resources. When the data samples satisfy the EI division condition, it takes work to shift the cluster in the iteration process. Therefore, the fusion of EI condition and membership boundary values into the IFCM algorithm to update the membership matrix is the main contribution of the proposed algorithm. The pseudo-code of EI-IFCM based on the EI condition and membership boundary is shown in Algorithm 2. In the proposed algorithm, the data samples satisfying the EI should be separated from the process and assigned the boundary value of the membership.

After obtaining the initial cluster center in Algorithm 1, Algorithm 2 starts its learning process from a better cluster center. In the process of updating the membership matrix, the dataset is divided according to the EI division condition. The division rules are shown in Eq. (15). The boundary value of membership can be obtained based on Eq. (14) as follows.

$$M=1-{\left(1-{\left[1+(c-1){\left(\frac{{\left(\Vert {x}_{j},{\nu }_{i}^{(t+1)}\Vert \right)}^{(1)}}{{\left({d}_{j}^{(t+1)}\right)}^{(c)}}\right)}^{2/(m-1)}\right]}^{-\alpha }\right)}^{1/\alpha }$$
(14)

Updating the membership of the separated sample Xout data by the following formula.

$${u}_{ij}^{(t+1)}\text{=}\left\{\begin{array}{l}M, \, {\text{i}}{\text{f}} \, i={I}_{j}^{*}\\ {u}_{ij}^{(t)}\left(\frac{1-M}{1-{u}_{ij}^{(t)}}\right), \, {\text{i}}{\text{f}} \, i\ne {I}_{j}^{*}\end{array}\right.$$
(15)

where \(\frac{1-M}{1-{u}_{ij}^{(t)}}\) is a factor less than 1 to maintain the \({u}_{ij}^{(t)}\) is decreased to \({u}_{ij}^{(t+1)}\) for \(i\ne {I}_{j}^{*}\).

4 Experimental results

4.1 Environment settings and measuring indexes

Nine datasets from the UCI datasets (https://archivr.ics.uci.edu/ml/index.php) are utilized in this experiment to verify the competitiveness and effectiveness of the proposed algorithm. The nine datasets mentioned above are aggregated from different groups in fields such as animals, plants, aviation, healthcare etc. In this section, the applicability of the method is demonstrated by analyzing these actual datasets. A personal computer running Windows 10 and Python 3.8.8 is used for all of the experiments. It has an Intel Core i5-1135G7 processor and a maximum memory capacity of 16 GB for all processes. The brand new algorithm is benchmarked with its counterparts such as FCM [8], LFCM [38], IFCM [23], rseFCM [38], MSFCM [29], FCM clustering by varying the fuzziness parameter (vFCM) [39], Improved FCM (IMFCM) [33], and feature weighted FCM (WFCM) [25]. As a classical clustering algorithm, the traditional FCM algorithm is often used to compare the improved clustering algorithms based on FCM. The proposed clustering algorithm integrates the advantages of the IFCM algorithm, and IFCM is also used to compare with our algorithm. The MSFCM applies the boundary value of membership degree to the clustering method for the first time and is used to compare with EI-IFCM. In addition, other methods such as LFCM, rseFCM, IMFCM, WFCM, and vFCM are also used to compare with our methods because of their wide application in clustering in very big datasets. The basic parameters m and α are set to 2 and 0.85, respectively. In addition, the iteration termination parameter ε is set as 1×10-6.

The Rand index (RI) for evaluating clustering methods quality proposed by Rand in 1971 is widely used to measure the similarity of two clustering partitions [40]. Since RI was proposed, it has been widely used in the performance evaluation of clustering algorithms. The larger the RI, the better the clustering performance. And the RI is extended by Hubert et al., which is adjusted rand index (ARI) [41]. The ARI inherits the advantages of RI and its value is between [-1,1]. The Adjusted Mutual Information (AMI) is an information theoretic measure and has the best properties among all these clustering evaluation measures [42]. In addition, the FMI considers the Fowlkes and Mallows Index (FMI) proposed by Fowlkes and Mallows in 1983 [43]. FMI is described as the geometric mean of accuracy and recall, which is used to comprehensively evaluate clustering performance. The larger the value of the above evaluation indexes, the closer the clustering result is to the true value, that is, the better the clustering performance.

4.2 Results and discussion

Table 1 demonstrates the results of FCM, LFCM, IFCM, rseFCM, MSFCM, WFCM, IMFCM, vFCM, and EI-IFCM over nine datasets coming from UCI datasets. The characteristic dimension (d), data amount (n) and category number (c) of the data set are listed in Table 1. On nine datasets, nine algorithms were tried ten times, and the average and standard deviation of each indicator’s ten tests are taken. The standard deviation is roughly 0 when it is less than or equal to 10-7. Specifically, when using the WFCM algorithm, its parameter t=-1, and the parameters a, b, and K of the vFCM algorithm are 0.95, 0.05, and 20, respectively. The selection of parameters for the above two algorithms comes from the optimal parameters validated in [25, 39].

Table 1 The calculation results of 9 algorithms on 9 real datasets

In this section, we use the above evaluation indicators mentioned in section 4.1 to measure the clustering performance of algorithms. Table 1 lists the detailed content of each indicator, with the best result highlighted in bold. In Table 1, in addition to evaluating the performance of a single indicator on different datasets, a comprehensive evaluation is also provided. The “Mean” in the table displays the comprehensive performance of the six indicators on all datasets, as well as the average ranking under all indicators in all datasets. Through verification, it can be found that the proposed algorithm can be well used in distinguishing different categories of plants, classifying diseases in medical treatment, and recognizing objects in aerial images and others.

From Table 1, it can be seen that different algorithms exhibit different performances on different datasets. For traditional algorithms such as FCM and LFCM, the convergence position of the algorithm is learned by randomly initializing the membership matrix and randomly initializing the clustering centers. Compared to other algorithms, they are not competitive on any dataset. The rseFCM algorithm, which integrates LFCM, has been proposed for clustering problems on large datasets. However, it inherits the inherent drawbacks of LFCM, such as randomization of initial centers and slow convergence speed, and its accuracy has not been significantly improved on all datasets. The IFCM algorithm is based on an improvement of the FCM algorithm in considering more uncertain information, which adds hesitation to the membership function of the FCM. Although the IFCM algorithm is random in obtaining the initial clustering centers, it performs well on the Shuttle, Avila, Abalone, Heart, Wine, and Iris datasets due to its consideration of more uncertain information in the data samples.

Compared with the algorithms mentioned above, the MSFCM algorithm performs better on multiple datasets. This is because the MSFCM algorithm proposes the concept of membership boundary from the perspective that all data objects always affect all clustering centers, and integrates it with the membership update process to obtain better and faster convergence positions. From Table 1, it can be seen that MSFCM exhibits good performance on most datasets, which proves that considering the boundary condition of membership is effective for improving the traditional algorithm FCM. The WFCM algorithm is an algorithm that considers feature weighting and the calculation of feature weights depends on the Euclidean distance between the cluster center and the mean of all cluster centers. It can be seen that the datasets Satimage and Avila are sensitive to feature weighting and exhibit good performance, but the performance of the WFCM algorithm on other datasets is mediocre. The IMFCM algorithm takes into account the disadvantage of using Euclidean distance to calculate the membership matrix of the FCM algorithm, which is sensitive to noisy data. It proposes the Euclidean distance function that ensures a significant weight on normal data while a small weight on noisy data. It can be seen that this improvement has shown good performance on the Cancer dataset.

The vFCM algorithm improves FCM from the perspective of selecting and updating fuzzy parameters. VFCM determines whether to use the function m=am0+b to update fuzzy parameters by determining the remainder of the iteration times t and K. It can be known that when t is a multiple of K, the value of m will decrease. For datasets that require multiple iterations to reach the convergence position, the value of m will decrease to 1 (with an initial value of 2). When the value of the fuzzy parameter is 1, it will cause confusion in the updating of membership due to the limitation of 1/(m-1). Therefore, the vFCM algorithm does not converge on some datasets, such as Satimage, Avila, and Shuttle.

The proposed EI-IFCM algorithm integrates the further consideration of uncertain information in membership calculation in IFCM, as well as the use of membership boundary values in algorithm iterations. In addition, the proposed algorithm also considers the selection of initial clustering centers, which is a gap in the aforementioned algorithms. From Table 1, it can be seen that the EI-IFCM algorithm, which integrates multiple advantages, has shown good performance on most datasets. The “mean” row of the table also fully reflects the overall performance of the proposed algorithm on all datasets, which is competitive.

Additionally, it can be seen that our algorithm consumes less time and iterations. This is because the proposed algorithm first starts its learning from better initial clustering centers, which are obtained based on the density of data samples. The proposed algorithm divides the data samples during the learning process based on the EI condition and assigns boundary values for samples that satisfy the EI condition. By starting the learning process from a better location and assigning membership boundary values that match EI condition, the convergence position is achieved quickly during the algorithm operation process. From Table 1, it can be seen that this viewpoint has been validated on almost all datasets.

To verify whether there are statistical differences between algorithms, we include the p-values obtained by the Wilcoxon test for every indicator of the proposed EI-IFCM algorithm and the other algorithms in Table 2. In addition, Table 2 also summarizes whether different indicators under each algorithm are greater than 0.05 (-) or less than 0.05 (+), and counts their quantities. Specifically, due to the non-convergence of the vFCM algorithm on the Satimage, Shuttle, and Avila datasets, these datasets have been removed when comparing EI-IFCM and vFCM algorithms. Although the proposed algorithm has significant differences in five or four indicators for FCM, IFCM, LFCM, and MSFCM, the p-value is still greater than 0.05 on one or two indicators, indicating that the null hypothesis cannot be rejected. Similarly, compared to the recently proposed new algorithms IMFCM, WFCM, and vFCM, overall, the null hypothesis cannot be rejected. Therefore, we can conclude that there is no statistically significant difference between our algorithm and the most recent algorithms.

Table 2 Wilcoxon test p-values obtained by the Wilcoxon test of all evaluate criteria pairs of the proposed algorithm and other algorithms

Figure 3 shows the visualization results of original samples, IFCM and EI-IFCM algorithms on Wine and Iris data sets. (a) and (d) are 2D figures on original samples. (b) and (c) are the clustering results of the two algorithms on the Wine dataset, respectively. (a) and (c) are the clustering results of the two algorithms on the Iris dataset respectively. It is clear from Fig. 3 that the algorithm proposed in this paper can achieve a better clustering effect than IFCM on Wine and Iris datasets. This shows that the proposed algorithm is competitive. Especially on the Wine dataset, there are more obvious clustering results.

Fig. 3
figure 3

Plots of Iris and Wine Datasets in 2-D by IFCM and EI-IFCM Algorithm

To verify the monotonicity of algorithms, Fig. 4 shows the iterative changes of nine algorithms on nine data sets during 30 iterations. The vertical coordinate represents the sum of the objective function and the initial objective function, i.e. J(U,V)/J(U0,V0), and the horizontal coordinate represents the number of iterations. Due to significant differences in target values among different algorithms, using this ratio of J(U,V) and J(U0,V0) can plot the results of all algorithms on the same graph. The initial objective values of FCM, LFCM, IFCM, rseFCM, MSFCM, FWFCM, IMFCM and vFCM algorithms are determined by randomly selected cluster centers or randomly given membership values. The change in the ratio during the iteration process will be significant when the original objective value is large, otherwise, the change will be small. The initial objective value of the EI-IFCM algorithm is determined based on the selection of cluster centers. In addition, it can be seen from Fig. 4 that all datasets changed dramatically at the beginning of the iteration, then gradually slowed down. Compared with other classic algorithms, the EI-IFCM algorithm can achieve convergence with fewer iterations.

Fig. 4
figure 4

Plots of J(U,V)/J(U0,V0) for the iterations of nine algorithms on nine datasets

4.3 Sensitivity analysis

The stability of the initial clustering center is depicted in Fig. 5. Two representative datasets, Seeds and Satimage, are selected to describe this change. This is because there are significant differences in sample size, feature dimension, and clustering number between the two datasets. The abscissa in Fig. 5 is the number of experiments, and the ordinate is the impact of the initial center on the accuracy of the algorithms. The initial cluster center of IFCM is randomly selected. It can be seen from the figure that the initial centers randomly selected will cause the RI of the algorithm to fluctuate greatly, while the initial clustering centers selected through local density contribute to the stability of the algorithm and ensure the clustering accuracy.

Fig. 5
figure 5

Influence of initial centers selection of IFCM and EI-IFCM Algorithms. (a) Seeds dataset, (b) Satimage dataset

Similar to traditional FCM and IFCM algorithms, the parameters of the EI-IFCM algorithm have an impact on clustering performance. In this section, a detailed analysis is conducted on the impact of each parameter.

To analyze the sensitivity of the proposed algorithm to the fuzzy parameter m, the RI values of FCM and IFCM algorithms under different parameters m in four datasets are also measured. Figure 6 shows the changes in EI-IFCM, FCM and IFCM algorithms on m. It can be seen from the figure that the RI of the three algorithms fluctuates on different m. This shows that different datasets have different sensitivity to the fuzzy parameter m, and it is necessary to select appropriate fuzzy parameters according to different datasets to obtain higher clustering accuracy. On the datasets of Seeds, Avila and Abalone, the derived algorithm has roughly the same fluctuation as FCM and IFCM. When m is greater than 2.4, IFCM and EI-IFCM show a large gap with FCM in volatility but EI-IFCM and IFCM have roughly the same volatility trend on the Satimage dataset. This is because IFCM and EI-IFCM are affected by another hesitation parameter α besides parameter m.

Fig. 6
figure 6

Plot of hesitation parameter m on clustering performance of three algorithms

EI-IFCM and IFCM algorithms are affected not only by fuzzy parameters m but also by the parameter α. Parameter α can change the hesitation value of data samples. Figure 7 shows the impact of IFCM and EI-IFCM algorithms on the RI of datasets when parameters α in [0.15,0.95] are considered. It can be seen from the figure that IFCM and EI-IFCM fluctuate with the change of parameter α. The fluctuation trend of RI obtained by two algorithms in different data sets is basically synchronous and consistent. It can be seen from the figure that when α=0.85, the two algorithms achieve the highest clustering accuracy in the Seeds and Satimage datasets at the same time. However, when the two algorithms achieve the highest clustering accuracy on the Avila and Abalone datasets, they are not under the same parameters. The IFCM achieves the highest clustering precision on two Avila when α=0.25, and EI-IFCM achieves the highest clustering precision on it when α is 0.7. The Abalone dataset achieves the highest clustering precision when α=0.65 in IFCM while the highest clustering precision is achieved in EI-IFCM when α=0.8. It is worth noting that the RI is 0.6131 in EI-IFCM and 0.6121 in IFCM when α=0.65, while RI in EI-IFCM is significantly greater than in IFCM when α=0.8 in the Abalone dataset. The same results can be obtained in the Avila dataset. On the whole, the performance of the EI-IFCM algorithm is better than that of the IFCM algorithm, which is most obvious on the Avila dataset, and the excellent ability of the algorithm has also been further proved.

Fig. 7
figure 7

Plot of hesitation parameter α on clustering performance of two algorithms

In order to verify the impact of both parameters on algorithm performance, Figure 8 depicts the RI values obtained after both parameters are changed simultaneously. Among them, the value range is α in [0.1,1], and the value range of m is in [1.2,3.9]. The x-axis in the figure is α and m, respectively, and the y-axis is the RI value. From the figure, it can be seen that the performance of EI-IFCM and IFCM algorithms fluctuates with changes in parameters. There are differences in the sensitivity of different datasets to parameters. For the Avila dataset, when m=2.1 and α=0.4, the RI of the EI-IFCM algorithm achieves good performance. When m=1.5 and α=0.2, the IFCM algorithm achieves better performance. In this experiment, when m=3.9 and α=1, EI-IFCM and IFCM achieved the optimal RI values on both the Satimage and Seeds datasets. In addition, for the Abalone dataset, EI-IFCM achieved optimal performance at m=2.1 and α=0.4, while IFCM achieved optimal performance at m=3 and α=0.7. When the hesitancy parameter α is 1, the hesitancy of IFS can be obtained by using \({\pi }_{IFS}={1-\mu }_{A}\left(x\right)\), which means that the consideration of hesitant information in the data sample is maximized. This means that the consideration of hesitant information in the data sample reaches its maximum state.

Fig. 8
figure 8

Plot of m and α on clustering performance of two algorithms

From the above analysis, it can be concluded that different datasets and algorithms have different sensitivity to parameters. M represents the fuzzy performance of the dataset, while α represents the hesitant information of the uninterrupted dataset. From the analysis, it can be seen that when both parameters are considered simultaneously, higher clustering performance can be achieved than IFCM on most datasets. Therefore, when solving different practical problems, it is necessary to assign different parameters.

Based on the above analysis, it can be concluded that the proposed algorithm is competitive. Based on the above experiments on algorithms performance and parameter sensitivity analysis, the following conclusions can be obtained. Firstly, the EI-IFCM algorithm is competitive compared to other algorithms. From Table 1, it can be seen that the various indicator values of the proposed algorithm are significantly better than other algorithms. Secondly, compared to other excellent algorithms, the proposed new algorithm uses a data density-based approach to obtain the initial clustering center. Compared with other algorithms that randomly select initial clustering centers, the proposed algorithm starts its learning process from a better center, thereby saving learning resources. At the same time, it avoids the instability of the initial clustering centers caused by random selection. Thirdly, by using EI condition to segment the dataset. It can assign membership boundaries to data objects that satisfy EI condition, which can avoid the disadvantage that all samples always affecting all clustering centers. Moreover, by applying the proposed algorithm to datasets in different fields, it can be concluded that the algorithm has applicability in fields such as image classification, plant category recognition, medical disease classification, and others. Finally, by comparing the parameter sensitivity analysis between the classical algorithm and the proposed algorithm, it can be seen that the proposed algorithm inherits the characteristics of the classical algorithm with different sensitivities on different datasets. The EI-IFCM algorithm also has different sensitivities on different datasets. This needs further improvement in future research.

5 Conclusion

Due to the idea of IFCM fusing hesitant information into membership and the recently proposed MSFCM model fusing membership boundaries, this paper proposes a new IFCM clustering algorithm called EI-IFCM, which integrates local density. In the EI-IFCM algorithm, the initial clustering centers are obtained based on the sample distribution density and are responsible for reducing instability caused by random selection. Therefore, the proposed algorithm can start its learning process from scratch with relatively stable initial clustering centers. On the other hand, in response to the viewpoint that all samples always affect all clustering centers in the classical IFCM algorithm, the algorithm combines the EI condition and the membership boundary derived in the paper. In this contribution, considering that the EI can provide partition condition for the datasets in two adjacent iterations, different membership calculation rules are given to ensure fast convergence of data samples. To verify the applicability of the proposed algorithm, experiments are conducted on 9 real world datasets. The experimental results show that the proposed EI-IFCM model has a competitive overall performance compared to other advanced models. By conducting experiments on these real world datasets, it can be verified that the proposed algorithm can be applied to fields such as medical disease classification, geographic image recognition, plant and animal differentiation, and others. This demonstrates the applicability of the algorithm.

However, the main limitation of the EI-IFCM algorithm is its inability to adaptively select the optimal parameters. Further research can obtain the optimal computational parameters by using heuristic algorithms such as particle swarm optimization or genetic algorithms. Moreover, this study also inherits the limitation of classical algorithms that require presetting the number of clusters. Integrating data distribution density into the algorithm to independently obtain the number of clusters can also serve as an improvement to this model, which is one of the future research directions.