1 Introduction

Risk management and control is always a problem that must be valued by managers (Abeysekara 2020). In order to minimize the impact of an accident, managers should evaluate the risks that exist in the system in advance, identify their priorities, and manage them effectively with limited resources so as to prevent accidents or reduce the harm caused by accidents. The issue of risk evaluation covers many fields. For example, in software programs, managers need to consider factors such as project schedules and the technical capabilities of employees (Tavares etal. 2019). In the medical field, the incidence of disease and mortality need to be considered (Altuntas and Gok 2021; Chauhan etal. 2021). In terms of food safety, technical and institutional risks should be considered (Choirun etal. 2020). In addition, when some major natural disasters occur, such as earthquakes (Trivedi 2018) and floods (Zheng et al. 2022), the causes and distribution areas need to be considered and analyzed, so as to reduce losses.

The existing risk evaluation methods mainly list the potential risks according to historical experience in the initial stage, and then, experts give their evaluations, analyzing the importance of these factors from multiple perspectives. For example, failure mode and effect analysis (FMEA) considers the probability, severity, and detection of risk factors (Zhongyi et al. 2021). Fault tree analysis (FTA) considers the causal relationship between factors (Yazdi et al. 2019). The analytic hierarchy process (AHP) identifies the levels of factors and determines the relationship between these levels (Kokangül et al. 2017). Some studies also use Bayesian networks (George and Renjith 2021) and petri nets (Zhang etal. 2020) to evaluate the risk. However, these methods do not consider the influencing relationship between risk factors. In practical applications, it is necessary to consider the internal relationship between factors; the ability to influence other factors and be affected by other factors can represent the importance of factors in the system to a certain extent. In terms of this issue, the DEMATEL method can analyze the importance of factors with their relationship to each other. The DEMATEL method has been used in many fields, such as the economy (Gang et al. 2021; Abraham et al. 2019), blockchain technology (Yadav and Singh 2020; Kouhizadeh et al. 2021; Kamble et al. 2020), and especially in the last 2 years in the control of coronavirus disease 2019 (COVID-19) (Tanvir et al. 2021; Ocampo and Yamagishi 2020), where it has played a huge role. In addition to making up for the defects of other methods, the DEMATEL method also has the advantage that it can be well combined with other theories and methods, such as gray theory (Raj etal. 2020; Kumar et al. 2021; Amirghodsi etal. 2020), evidence theory (Li et al. 2014; Yuan-Wei and Zhou 2019; Shang et al. 2020), and fuzzy theory (Chuanbo et al. 2020; Asan et al. 2018; Feng and Ma 2020). In order to better represent the fuzziness of expert evaluation language, more and more research has combined fuzzy set theory with DEMATEL. Among them, (Asan et al. 2018) proposed a DEMATEL method based on interval-valued hesitant fuzzy sets, which considers artificial uncertainty in the evaluation process and retains the differences among experts. Pandey et al. (2019) proposed a DEMATEL method that combines interval 2-type fuzzy numbers (IT2FS-DEMATEL), which combines triangular fuzzy numbers and trapezoidal fuzzy numbers. In this way, imprecise and ambiguous evaluations can be significantly avoided. These methods based on fuzzy sets need to defuzzify the expert evaluation, and this process tends to average the differences among the evaluations, which cannot effectively deal with the uncertainty generated in the evaluation process. To solve this problem, there are more and more studies on the improvement of DEMATEL based on evidence theory (Evidential-DEMATEL). Evidential-DEMATEL (Li et al. 2014) method converts the fuzzy evaluations to basic probability assignment(BPA), thus avoiding defuzzification. At the same time, the Dempster–Shafer combination rule is used to integrate group decision-making. On this basis, Yuan-Wei and Zhou (2019) used evidence theory to extract the subjectivity of experts in evaluation and proposed a DEMATEL method based on subjective experience and objective data. Shang et al. (2020) used belief entropy to calculate the reliability of evaluations. A reliability coefficient is added to each fused BPA to make the fused results more reasonable. In addition, Lin et al. (2018) extended DEMATEL to D-number theory to overcome the limitation that language evaluation by experts must be mutually exclusive. Since D-number allows information to be missing and incomplete, the improved D-DEMATEL is more suitable for language evaluation. But this method is less effective when there is too much missing information. Jiang et al. (2020) used Z numbers to represent expert evaluation information more flexibly and accurately and used a similarity measure-based method to cluster experts, which effectively dealt with uncertainty in the evaluation process. However, in practical application, the scale of expert groups would not be too large, and the application of this method would be limited. Another trend in the DEMATEL method is the hybrid application of multiple methods. For example, Mohammed et al. (2021) proposed a hybrid method of DEMATEL and TOPSIS that used DEMATEL to quantify the weight of evaluation criteria and the TOPSIS method to provide optimal decisions for decision-makers. Das et al. (2022) used AHP and DEMATEL to analyze the key factors affecting the global supply chain during the outbreak of COVID-19, aiming to help decision-makers develop a risk mitigation framework. There is no doubt that this hybrid multi-method technique can combine the advantages of each method and make the results closer to reality.

In the real world, in order to get more accurate evaluation information, an expert evaluation team is often set up to develop a joint evaluation plan. Due to the differences in historical experience and subjective opinions, their evaluations are not exactly the same, sometimes even contradictory. Therefore, it is necessary to consider the credibility of each expert when accepting their programs. Yazdi et al. (2020) improved DEMATEL by using the Best-Worst method (BWM) to obtain the evaluation criteria and the weight of experts. However, they ignore the impact of expert opinion differences. Chen et al. (2020) introduced belief entropy to measure the amount of information in evaluations and calculated the weight of experts. In order to improve the existing DEMATEL method more effectively, this paper mainly makes improvements to the following two issues:

  1. 1.

    How to effectively measure the credibility and weight of experts?

  2. 2.

    How to deal with uncertainty arising from the evaluation process?

In response to the two issues, we introduce the two-dimensional correlation coefficient (2-DCC) and the negation of BPA on the basis of fuzzy evidential DEMATEL. As for the first issue, the new method uses 2-DCC and the total uncertainty measure (TU) to calculate expert weights from macroscopic and microscopic perspectives, respectively. The 2-DCC has been able to measure the correlation of two matrices (Dikbaş 2017), and some research about the 2-DCC has been applied in environmental monitoring in recent years (Li et al. 2019; DİKBAŞ and BACANLI 2020; Yasar and Dikbas 2022). Compared with other DEMATEL methods, the new method not only considers the differences between the overall evaluations of experts but also the differences between each piece of evaluation, which can better express the credibility of each expert through the differences and is more in line with the actual situation. As for the second issue, the new method uses negation of BPA for weighted fusion to deal with the uncertainty of evaluation. Negation of BPA describes things from the negative perspective (Yager 2014), which can provide more valuable information. In recent years, many studies have focused on the properties and applications of negation of BPA. Mao and Deng (2022) proposed a calculated method of negation of BPA based on the belief interval and applied it in pattern recognition. Yin et al. (2018) proposed a new negation of BPA calculating method and, on this basis, a new uncertainty of BPA measuring method. Dongdong et al. (2020) proposed a new weighted average evidence calculation method using the negation of BPA and applied it to classification problems. Tang et al. (2022) proposed a method of measuring and managing the uncertainty in the negation information of evidence. In general, negation of BPA is becoming a useful tool to deal with conflicts between evidence. The main contribution of this paper is that we first propose to combine macro and total uncertainty to determine the credibility of expert evaluations. This provides a new idea and method to deal with the uncertainty of expert evaluation in DEMATEL.

The rest of this paper is organized as follows: Sect. 2 reviews the theoretical basis of this paper. In Sect. 3, on the basis of the existing fuzzy evidential DEMATEL method, a DEMATEL method based on 2-DCC and negation of BPA is proposed. Then, in Sect. 4, two case studies are used to verify the application of the proposed method. Finally, Sect. 5 is a summary of the content of this paper.

2 Preliminaries

2.1 Evidence theory

Evidence theory is a generalization of the Bayesian theory proposed by Professor Dempster in 1976 and improved by Shafer. Compared with traditional probability theory, evidence theory can effectively represent uncertain information. The relevant definitions are as follows:

Definition 1

Supposing \(\Omega =\lbrace \theta _1,\theta _2,\ldots ,\theta _n\rbrace \) is a set of mutually exclusive events, the elements of \(\Omega \) correspond to all possible events in a fixed scenario, and their probabilities of occurrence do not affect each other, so \(\Omega \) is called the frame of discernment (FOD). The set of all subsets of \(\Omega \) is called the power set of \(\Omega \), be denoted as \(2^\Omega \).

$$\begin{aligned} 2^\Omega =\lbrace \varnothing ,{\lbrace \theta _1\rbrace },{\lbrace \theta _2\rbrace },\ldots ,{\lbrace \theta _n\rbrace },\ldots ,{\lbrace \theta _1,\theta _2\rbrace },\ldots ,\Omega \rbrace \end{aligned}$$
(1)

where \(\varnothing \) is empty set, The elements in \(2^\Omega \) are called propositions, and every proposition has a meaning.

Definition 2

Mass function represents the mapping relationship between the elements of \(2^\Omega \) and the interval \(\left[ 0,1\right] \), defined as follows:

$$\begin{aligned} m:2^\Omega \rightarrow \left[ 0,1\right] \end{aligned}$$
(2)

This mapping is usually defined as a probability and satisfies the following conditions:

$$\begin{aligned} m(\varnothing )=0,\quad \sum _{A\in \Omega } m(A) = 1 \end{aligned}$$
(3)

if m(A)>0, then A is called the focal element, and it indicates the extent to which the evidence supports A.

Definition 3

A set of mass function of the elements in \(2^\Omega \) is called as body of evidence(BOE), also called as the basic probability assignment (BPA), and BPA is defined as follows:

$$\begin{aligned} (\Re ,m)=\left\{ \langle A,m(A)\rangle :A\in 2^\Omega ,m(A)>0\right\} \end{aligned}$$
(4)

where \(\Re \) is a subset of \(2^\Omega \).

Definition 4

Belief function represents the lower limit of a proposition, For proposition A, the belief function is defined as follows:

$$\begin{aligned} \hbox {Bel}(A)=\sum _{B\subseteq A} m(B) \end{aligned}$$
(5)

Plausibility function represents the upper limit of a proposition, defined as follows:

$$\begin{aligned} \hbox {Pl}(A)=\sum _{A\cap B=\varnothing } m(B) \end{aligned}$$
(6)

Definition 5

Under the frame of evidence theory, two BPAs can be combined by Dempster–Shafer combination rule to obtain a new BPA. Supposing \(m_1\) and \(m_2\) are two BPAs in FOD, B and C are focal elements of \(m_1\) and \(m_2\), respectively. The combination rule is defined as follows:

$$\begin{aligned} m_{1,2}(A)= & {} (m_1 \oplus m_2)(A)\nonumber \\= & {} \frac{1}{1-k}\sum _{B\cap C=A} m_1(B)m_2(C) \end{aligned}$$
(7)

where k is called as conflict coefficient and represents the degree of conflict between two BPAs, defined as follows:

$$\begin{aligned} k=\sum _{B\cap C=\varnothing } m_1(B)m_2(C). \end{aligned}$$
(8)

2.2 Total uncertainty measure

There is always uncertainty in the real world (Zhang et al. 2018, 2017). There are many methods for measuring uncertainty under evidence theory (Wang and Song 2018; Deng 2020; Tang et al. 2023). The total uncertainty measure combines discord and non-specificity and gives rise to many uncertainty measures, such as entropy (Gao et al. 2019; Deng 2016), ambiguity measure (Jousselme et al. 2006), and Hartley measure (Pan et al. 2019). Among them, the most widely used method of measuring total uncertainty by Pal et al. (1992) is defined as follows:

Definition 6

$$\begin{aligned} H(m)=-\sum _{x\in \Omega } m(x)\log _2 \frac{m(x)}{|x |} \end{aligned}$$
(9)

where \(|x |\) represents the element’s cardinality. H(m) can be seen as the extension of Shannon entropy, which considers the cardinality of each element on the basis of Shannon entropy, and allocates the uncertainty contained in multi-element propositions. It also satisfies the properties of subadditivity, additivity, and continuity (Pal et al. 1993).

2.3 Negation of basic probability assignment

Yager proposed the negation of the probability distribution, redistributing the probabilities on the basis of the original probability distribution from the standpoint of negation (Yager 2014).

Definition 7

Supposing \(m=\lbrace m(A_1),m(A_2),\ldots ,m(A_n)\rbrace \) is a BPA, the negation of m is defined as follows:

$$\begin{aligned} m({\bar{A}}_i)= \frac{\sum _{j=1,j\ne i}^m(A_j)}{n-1} = \frac{1-m(A_i)}{n-1} \end{aligned}$$
(10)

where n represents the number of focal elements in m, \(m(A_j)\) is the original probability, and \(m({\bar{A}}_i)\) satisfy the following conditions:

$$\begin{aligned} 0\le m({\bar{A}}_i) \le 1 \quad \hbox {and} \quad \sum _{i=1}^n m({\bar{A}}_i)=1 \end{aligned}$$
(11)

2.4 Two-Dimensional correlation coefficient (2-DCC)

Definition 8

Similar to the Pearson correlation coefficient, 2-DCC is capable of evaluating the correlation of two matrices. By measuring the differences between matrices in the horizontal and vertical directions, it determines the horizontal correlation coefficient rh and the vertical correlation coefficient rv. In the horizontal direction, rh is assessed by computing the differences between the matrix elements and the row average, as in Eq. 12. In the vertical direction, rv is assessed by computing the differences between the matrix elements and the column average, as in Eq. 13 (Dikbaş 2017):

$$\begin{aligned} rh=\frac{\sum _m\sum _n(A_{mn}-{\bar{A}}_m)(B_{mn}-{\bar{B}}_m)}{\sqrt{ \left( \sum _m\sum _n(A_{mn}-{\bar{A}}_m)^2\right) \left( \sum _m\sum _n(B_{mn}-{\bar{B}}_m)^2\right) }}\nonumber \\ \end{aligned}$$
(12)
$$\begin{aligned} rv=\frac{\sum _m\sum _n(A_{mn}-{\bar{A}}_n)(B_{mn}-{\bar{B}}_n)}{\sqrt{ \left( \sum _m\sum _n(A_{mn}-{\bar{A}}_n)^2\right) \left( \sum _m\sum _n(B_{mn}-{\bar{B}}_n)^2\right) }}\nonumber \\ \end{aligned}$$
(13)

where m and n represent the row and column of matrices, respectively. \({\bar{A}}_m\) and \({\bar{B}}_m\) represent the mth row average of A and B, respectively. \({\bar{A}}_n\) and \({\bar{B}}_n\) represent the nth column average of A and B, respectively. rh and rv are in the range of \([-1,1]\). If rh or rv is closer to -1, it indicates a poorer correlation between the two matrices in either the horizontal or vertical direction, and if rh or rv is closer to 1, it indicates a stronger correlation between the two matrices in either direction. In this paper, in order to use the correlation between matrices to computing the support of matrices, to computing the weight of matrices, we average the rh and rv to get the overall correlation r, as in Eq. 14. It is simple to demonstrate that r is in the interval \([-1,1]\), which may affect the calculation of weights when \(r<0\). We have two explanations for this issue:

  1. 1.

    In the DEMATEL method, the expert evaluation matrices, while varied, are not generally the opposite. So the correlation between them is greater than 0 in the vast majority.

  2. 2.

    If we force the range of r to [0, 1], the difference between the correlations of matrices will be changed, and this will make the result inaccurate.

$$\begin{aligned} r=\frac{rh+rv}{2} \end{aligned}$$
(14)

2.5 Intuitionistic fuzzy set

Definition 9

For a given domain of discourse X, the Intuitionistic fuzzy set on X is denoted as A (Atanassov 2016).

$$\begin{aligned} A=\lbrace \langle x,\mu _A(x),\nu _A(x) \rangle |x \in X \rbrace \end{aligned}$$
(15)

where \(\mu _A(x)\) and \(\nu _A(x)\) are all a mapping to \(\left[ 0,1\right] \), which satisfy \(0\le \mu _A(x)+\nu _A(x)\le 1\), \(\mu _A(x)\) represents the degree of membership that x with respect to A, and \(\nu _A(x)\) represents the degree of non-membership that x with respect to A. For convenience, \(m=(\mu _A(x),\nu _A(x))\) is called as intuitionistic fuzzy number (IFN).

Definition 10

The score function can convert IFN into a real number and be used to evaluate the merits and demerits of IFNs (Zeshui 2007). The score function is defined as follows:

$$\begin{aligned} S_A = \mu _A(x)-\nu _A(x) \end{aligned}$$
(16)

where \(S_A\in \left[ -1,1\right] \); if \(S_A\le S_B\), then IFN B is better than IFN A.

2.6 Decision-making trial and evaluation laboratory (DEMATEL)

The Decision Laboratory Analysis (DEMATEL) method was first proposed by A. Gabus and E. Fontela at a conference in Geneva in 1971. This method uses graph theory and matrix theory to analyze the influence relationships of each factor in the system. According to the analysis results, the key factors in the system are defined so as to implement effective risk management and control. The steps of the traditional DEMATEL method are as follows:

Step 1: The direct relation matrix (DRM) is defined according to the influence relationship among the factors.

The influence relationship of each factor in the system can be represented by a directed graph, as shown in Fig. 1. Each vertex in the digraph represents a potential risk factor, directed edge represents the influence relationship between two factors, and the weight of directed edge represents the degree of influence. This digraph is converted into an adjacent matrix, which is called the direct relation matrix (Eq. 17). Due to the fuzziness of the influence relationship between factors, such as \(\lbrace \) no influence, weak influence, general influence, and great influence\(\rbrace \), the weight of edges can usually be expressed by different grades (from 0 to 4). It should be noted that if two vertices are not connected by a directed edge, the corresponding element in the adjacent matrix is 0, which is in contradiction with graph theory.

Fig. 1
figure 1

A weighted digraph representing the relationship of influence between factors

$$\begin{aligned} D= \begin{bmatrix} 0&{}w_1&{}0&{}0&{}0\\ 0&{}0&{}0&{}0&{}w_5\\ w_4&{}0&{}0&{}0&{}0\\ w_3&{}0&{}w_2&{}0&{}0 \end{bmatrix} \end{aligned}$$
(17)

Step 2: Normalized DRM

The elements \(d_{ij}\) in DRM are converted into a mapping from \(d_{ij}\) to [0, 1]. There are many ways to normalize, but all of them are based on a maximum value. The common normalize method is defined as follows:

$$\begin{aligned} N=\frac{D}{\max _i\left( \sum _{j=1}^n d_{ij}\right) } \end{aligned}$$
(18)

Step 3: Calculate the total relation matrix (TRM)

The elements \(N_{ij} \in [0,1]\), so as N multiplies itself, and the result goes to 0 indefinitely.

$$\begin{aligned} \lim _{k\rightarrow \infty }N^k=0 \end{aligned}$$
(19)

Each self-multiplication of N represents to increase the indirect influence of factors. The total relation matrix is the sum of the results of continuous self-multiplication of N, indicating the sum of all indirect effects.

$$\begin{aligned} T=N^1+N^2+\cdots +N^k=N(I-N)^{-1} \end{aligned}$$
(20)

where I represents identity matrix, N represents normalized matrix, and T represents total relation matrix.

Step 4: Analyze the importance of factors

Definition 11

Calculate the sum of the ith row of T, which represents the comprehensive influence value of the ith factor on other factors, denoted as \(R_i\), which is defined as follows:

$$\begin{aligned} R_i = \sum _{j=1}^n T_{ij} \end{aligned}$$
(21)

Calculate the sum of the ith column of T, which represents the comprehensive influence value of other factors on the ith factor, denoted as Ci, which is defined as follows:

$$\begin{aligned} C_i = \sum _{j=1}^n T_{ji} \end{aligned}$$
(22)

The value of \(R+C\) indicates the importance of the factor in the system, which is called centrality. The higher the value of \(R+C\), the stronger the correlation between the factor and other factors, which should be paid more attention to. The value of \(R-C\) is called causality, and it shows how much one factor affects another. If the \(R-C\) value of a factor is less than 0, it indicates that the factor is more affected by other factors, that is, the effect factor. If the value of \(R-C\) is greater than 0, it indicates that this factor will affect other factors, that is, cause factors. Risk factors can be defined by the relationship between \(R+C\) and \(R-C\), as in Fig. 2

Fig. 2
figure 2

The mapping between \(R+C\) and \(R-C\)

3 The improved fuzzy evidential DEMATEL based on 2-DCC and negation of BPA

This section will describe the flow of the new method in detail. The new method combines 2-DCC and total uncertainty to measure the weight, and the uncertainty and conflict will be dealt with by the weight average evidence that is obtained by negation of BPA. The general flow of this method is shown in Fig. 3. The detailed steps are as follows:

Fig. 3
figure 3

The flow of the improved DEMATEL method

Step 1: Determine the risk factors in the system and get evaluations from experts.

Step 1-1: The risk factors will be listed by managers and decision-makers according to historical experience.

Step 1-2: Construct a group of experts to evaluate the risk factors. In practice, the evaluations of experts tend to be fuzzy and cannot be represented by an exact number. The expert evaluations are denoted by IFNs in this paper. The common IFNs are shown in Table 1.

Step 2: Obtain the macro-weight of experts. The result of this step can be regarded as the macro-weights of the experts because it takes the differences between their overall evaluations as its basis.

Step 2-1: The IFNs matrix of experts is converted into a score matrix according to the score function.

Step 2-2: Calculate the correlation between each score matrix and the other score matrices by Eqs. 1214, and the correlation coefficient matrix (CCM) is constructed as follows:

$$\begin{aligned} \hbox {CCM}= \begin{bmatrix} r_{11} &{} \cdots &{} r_{1n}\\ \vdots &{} \ddots &{} \vdots \\ r_{n1} &{} \cdots &{} r_{nn}\\ \end{bmatrix} \end{aligned}$$
(23)

Step 2-3: The support of each expert is calculated based on CCM. The formula is as follows:

$$\begin{aligned} \hbox {Sup}_i=\sum _{j=1}r_{ji} \end{aligned}$$
(24)

Step 2-4: Calculate the macro-weight.

$$\begin{aligned} Wm_i=\frac{\hbox {Sup}_i}{\sum _{j=1}\hbox {Sup}_j} \end{aligned}$$
(25)

Step 3: Calculate the micro-weight of evaluations according to the total uncertainty.

Step 3-1: Convert IFNs into BPAs.

Step 3-2: Calculate the negation of BPA with Eq. 10.

Step 3-3: Calculate the total uncertainty with Eq. 9.

Step 4: Weighted fuse evaluations

Step 4-1: Adjust the total credibility of evaluations on the basis of macro-weight and total uncertainty by Eq. 26. \(Wm_i\) is obtained by the differences of the overall evaluations, so \(Wm_i\) is the base number of Eq. 26.

$$\begin{aligned} \hbox {Cred}_i = Wm_i^{Hm_i} \end{aligned}$$
(26)

Step 4-2: The evaluation credibility is normalized to obtain the comprehensive weight of the evaluation, as shown in Eq. 27.

$$\begin{aligned} Wc_i = \frac{\hbox {Cred}_i}{\sum _{j=1}^n \hbox {Cred}_j} \end{aligned}$$
(27)

Step 4-3: Calculate and fuse the weight average evidence (WAE).

$$\begin{aligned} \hbox {WAE} = \sum _{i=1}^n (Wc_i\times m_i) \end{aligned}$$
(28)

Fuse WAE according to Eq. 7. If there are k experts, then perform \(k-1\) times fusion. The fused evaluation result is obtained.

Step 5: The DRM is constructed according to the fusion results, and the TRM is calculated by using the traditional DEMATEL method.

Step 5-1: Only a pair of comprehensive fusion evaluations among risk factors can be obtained by performing Step 4 once. So it is necessary to repeat Step 3 and Step 4 to calculate the comprehensive fusion evaluations among all factors and construct the DRM.

Step 5-2: Normalize the DRM.

Step 5-3: Calculate the TRM.

Step 5-4: Analyze the importance of risk factors. It should be noted that this method uses the negation of BPA to describe the evaluation from the negative perspective, so take the opposite number of the value of \(R-C\) as the result to determine risk factors.

Table 1 The common IFNs in evaluations of influence relationship
Table 2 The list of risk factors

4 Applications and discussion

4.1 Application in emergency management

In this section, an application in emergency management is used to verify the usability and effectiveness of the proposed method. The experiment defined some influencing factors in the emergency management system (Li et al. 2014), such as the “Well-planned emergency relief supply system” and “Reasonable organizational structure and clear awareness of responsibilities.” These factors have a mutual influence relationship. For example, “Well-planned emergency relief supply system” can promote “Applicable emergency response plan and regulations.” Through the new DEMATEL method to define the key factors in the system, the performance of the emergency management system can be effectively improved. The specific description is as follows:

Step 1: Based on the historical experience, the 10 existing risk factors in the system are listed, as shown in Table 2. And three experts were invited to evaluate the 10 risk factors and construct the IFN matrices such as Table 3. (The rest of the evaluation data could be found in Li et al. 2014.) Take the IFN (0.1, 0.9) from F1 to F3 in Table 3 as an example. This IFN under DEMATEL can be interpreted as the degree of membership of F1 directly related to F3 is 0.1 and the degree of non-membership of F1 directly related to F3 is 0.9.

Table 3 The evaluations of the first expert

Step 2: Turn the IFN matrices into the score matrices according to the score function of IFN. The result is shown in Table 4. According to the 2-DCC and overall correlation coefficient r, the correlation coefficient matrix between expert evaluations is constructed. And calculate the macro-weight of expert evaluations.

Table 4 The score matrix of the first expert

Correlation coefficient matrix:

$$\begin{aligned} \hbox {CCM}= \begin{bmatrix} 1 &{} 0.8640 &{} 0.8579\\ 0.8640 &{} 1 &{} 0.8426\\ 0.8579 &{} 0.8426 &{} 1\\ \end{bmatrix} \end{aligned}$$
(29)

Support degree of experts:

$$\begin{aligned} \hbox {Sup}_1=2.7219 \quad \hbox {Sup}_2=2.7067 \quad \hbox {Sup}_3=2.7005 \end{aligned}$$

The macro-weight of experts:

$$\begin{aligned} Wm_1=0.3348 \quad Wm_2=0.3330 \quad Wm_3=0.3322 \end{aligned}$$

Step 3: Convert the IFNs into BPAs. For example, the BPA construct by IFN (0.1,0.8) is: m(Y) = 0.1, m(N) = 0.8, m(YN) = 0.1, where m(Y) represents the probability that there is a direct relationship between factors, m(N) represents the probability that there is no direct relationship between factors, and m(YN) represents the uncertainty between m(Y) and m(N), and then weighted fuse the BPAs by total uncertainty and negation of BPA. Take the evaluations from F1 to F2 of the three experts as an example to demonstrate the calculation process.

The BPAs are construct as follows:

$$\begin{aligned}{} & {} m_1:m_1(Y)=0.04 \quad m_1(N)=0.9 \quad m_1(Y,N)=0.06\\{} & {} m_2:m_2(Y)=0.02 \quad m_2(N)=0.9 \quad m_2(Y,N)=0.08\\{} & {} m_3:m_3(Y)=0.06 \quad m_3(N)=0.9 \quad m_3(Y,N)=0.04 \end{aligned}$$

The negation of BPA:

$$\begin{aligned}{} & {} {\bar{m}}_1:m_1({\bar{Y}})=0.48 \quad m_1({\bar{N}})=0.05 \quad m_1(\overline{Y,N})=0.47\\{} & {} {\bar{m}}_2:m_2({\bar{Y}})=0.49 \quad m_2({\bar{N}})=0.05 \quad m_2(\overline{Y,N})=0.46\\{} & {} {\bar{m}}_3:m_3({\bar{Y}})=0.47 \quad m_3({\bar{N}})=0.05 \quad m_3(\overline{Y,N})=0.48 \end{aligned}$$

The total uncertainty:

$$\begin{aligned} H(m_1)=1.7063 \quad H(m_2)=1.6857 \quad H(m_3)=1.7163 \end{aligned}$$

Step 4: Adjust the comprehensive weight by macro- and micro-weight, and construct WAE and fuse.

The comprehensive weight of BPA:

$$\begin{aligned} Wc_1=0.3358 \quad Wc_2=0.3365 \quad Wc_3=0.3277 \end{aligned}$$

The weighted average evidence:

$$\begin{aligned}{} & {} \textit{WAE}(m):m(Y)=0.4801 \quad m(N)=0.05 \\{} & {} m(Y,N)=0.4699 \end{aligned}$$

Fuse the WAE twice:

$$\begin{aligned} m(Y)=0.8428 \quad m(N)=0.0411 \quad m(Y,N)=0.1161 \end{aligned}$$

So far, we have obtained the fusion evaluation from F1 to F2 of three experts. Next, repeat Step 3 and Step 4 to obtain the fusion evaluation of all other factors, and construct the DRM. We can construct the DRM from positive (m(Y)) and negative (m(N)) (Tables 5 and 6)

Table 5 The fused DRM in positive side
Table 6 The fused DRM in negative side

Step 5: Calculate the TRM on the positive and negative sides, respectively (Tables 7 and 8), and analyze the importance of factors from the two perspectives. Tables 9 and 10 list the related indicators. Due to the negation of BPA, the value of \(R-C\) is negative.

Table 7 The TRM in positive side
Table 8 The TRM in negative side

“Positive” means that there is a direct relationship between two factors, so the value of \(R-C\) is higher when the risk is higher. As seen from Table 9, \(F9>F2>F6>F10>F5>F4>0\), which means that they affect other factors more than others, and define them as cause factors on the positive side. On the contrary, “negative” means that there is no direct relationship between two factors, so the value of \(R-C\) is smaller, the risk is higher. As seen from Table 10, \(F9<F2<F6<F10<F5<F4<0\), so they will be defined as cause factors on the side of negative. The results indicate that the causal factors defined by the two sides are the same.

4.2 Application in online shopping platform

In order to verify the usability of this method in the context of greater uncertainty and fuzzyness, we tested it with the application in online shopping platform. In online shopping, the customer may have some criteria when purchasing merchandise. These criteria are defined as influence factors that have an impact on the customer’s decision. And these criteria will be changed over time. We can determine which criteria have the greatest influence on customers’ decisions by analyzing the criteria of customers’ purchases in different periods through the new DEMATE. In this experiment, we suppose a customer has 12 criteria (factors) when he purchases a food item, they are calories (F1), fat (F2), sugar content (F3), packaging (F4),taste (F5), volume (F6), brand (F7), positioning (F8), advertising (F9), eating methods (F10), easy to cook or not (F11), nutrition or not (F12). And the relationships among the 12 influence factors were evaluated at three different periods in a day. The new DEMATEL was used to determine which factors had the most influence on the customer’s decision-making. Table 11 shows a portion of the detailed evaluations (Gao et al. 2021). The data are the fuzzy evaluation language in Table 1. It is clear that these data have more fuzzyness than the application in Sect. 4.1 because they are only divided into five levels. In addition, the conflict between the evaluations in application2 is even greater.

Table 9 The values of indicators in positive side
Table 10 The values of indicators in negative side
Table 11 The fuzzy language evaluations in the first period

Due to space constraints, we will not demonstrate the specific calculation process. The results are shown below:

Figure 4 shows the result of the three periods and the fused result of this method. An analysis of the results of the three periods shows that the evaluations of three periods are conflicted (especially for F5, F6, and F10). According to the 2-DCC, the macro-weight of experts is \(Wm_1=0.3690\,Wm_2=0.3280\,Wm_3=0.3030\). So the fused result should be closer to the first expert. It follows that our results are reasonable.

Fig. 4
figure 4

The results of different periods and the fused result

4.3 Discussion

In order to better verify the ability of our method to identify risk factors, this part will discuss the effectiveness and accuracy of this method from several aspects by comparing it with other methods (E-DEMATEL Li et al. 2014, D-DEMATEL Zhou et al. 2017, and EFE-DEMATEL Han and Deng 2018). In order to be able to have a more intuitive comparison with other methods, we take the value of \(-(R-C)\) as the criterion in the positive side, and in negative side, we take the value of \(R-C\) as the criterion.

Figures 5 and 6 show the comparison of the value of \(R-C\) with other methods in its application to emergency management. It is obvious that the values of these methods are very close, which verifies the feasibility of our method. The ordered results are shown in Tables 12 and 13, which are roughly the same as with other methods. The highest risk factor in our method is F9, which is not the same as with other methods. As can be seen from Table 17, this is due to the consideration of macro-weights. However, the difference value between F2 and F9 is much smaller than the other methods, which are 0.0902 in the positive side and 0.216 in the negative side. In practice, managers tend to pay equal attention to factors with very similar \(R-C\) values, so our results are also reasonable. Regarding the determination of key risk factors, F9, F2, F6, F10, F5,  and F4 were defined as key factors (causal factors) on both the positive and negative sides. Similar to the other methods, it shows the effectiveness of our method.

Fig. 5
figure 5

The comparison of \(R-C\) with other methods on the positive side

Fig. 6
figure 6

The comparison of \(R-C\) with other methods on the negative side

Table 12 The comparison of ordered result with other methods on the positive side
Table 13 The comparison of ordered result with other methods on the negative side

The comparison of the application in Sect. 4.2 is shown in Fig. 7. Among them, the classic DEMATEL fused the result by average values. The negation of BPA is a process of redistributing probabilities, which will make the probabilities average while preserving the differences between the BPAs and reducing the conflict between the BPAs. Furthermore, in cases of high conflict, the macro-weight is an important factor to consider. It is precisely because we take this factor into account and because the weight of the first period is the highest that our result curve is close to the result curve of the first period, which indicates the credibility of our result. Other methods ignore the differences in overall evaluations, and their results do not closely follow the curve of any period of time. So their results are not credible.

Fig. 7
figure 7

The comparison with other methods

Table 14 displays the ranking results for application 2. Due to the greater fuzzyness and conflict in the experimental data, the sorting results obtained by various methods are not the same. However, the F5 is defined as the highest risk factor, which is the same as other methods. And because we consider the expert macro-weight, several key factors (F1, F4, and F10) in the ranking will be different from other methods, which explains that our methodology further considers potential uncertainties and redefines new risk factors.

Table 14 The ordered result with other methods

Han and Deng (2018) also proposed MAE to evaluate the processing ability of different methods under uncertainty and fuzzyness. MAE is defined as follows:

$$\begin{aligned} \hbox {MAE}=\frac{1}{N}\sum _{i=1}^N\left|V_Y^i-V_N^i \right|\end{aligned}$$
(30)

where N represents the number of risk factors, \(V_Y^i\) represents the value of \(R-C\) of factor i on the positive side, and \(V_N^i\) represents the value of \(R-C\) of factor i on the negative side. The MAE comparison with other methods is shown in Tables 15 and 16. That shows that the MAE of our proposed method is the largest in the two applications, indicating that our method can effectively deal with the uncertainty and fuzzyness in the evaluation process.

Table 15 The comparison of MAE value in Sect. 4.1
Table 16 The comparison of MAE value in Sect. 4.2

The biggest difference between our method and other methods is that we consider the differences between the overall evaluations of experts, and then, the macro-weight of the evaluation is determined. In order to better illustrate the impact of the difference in the overall evaluations of experts on the ordered results, we compared the results of this method with those of the method excluding macro-weight on the basis of Sect. 4.1. The ordered results are shown in Tables 17 and 18. Both the size of the \(R-C\) value and the order of factors are affected by the macro-weight, indicating that the difference between the overall evaluations of experts will have an impact on the result and should be taken into account in practice.

Table 17 The comparison with excluding macro-weight on the positive side
Table 18 The comparison with excluding macro-weight on the negative side

5 Conclusion

The DEMATEL method is widely used in practice in order to effectively control the potential risk factors in the system under the premise of limited resources so as to reduce the damage caused by accidents. But in practice, the evaluation is often inaccurate because of the fuzzyness and uncertainty of expert evaluation. Based on the existing fuzzy evidential DEMATEL method, this paper proposed an improved fuzzy evidential DEMATEL method based on 2-DCC and negation of BPA. The method evaluates the factors in the system with IFN and converts them into a score matrix using the score function of IFN, and the difference between the overall evaluations of experts will be measured by 2-DCC, which is introduced into the subsequent calculation. Then, BPA was modeled for the IFN, and the weighted average evidence was constructed and fused by combining the negation of BPA and total uncertainty. The DRM is constructed according to the fusion results, and risk factors with an \(R-C\) value greater than 0 are considered key factors in the system. This method can effectively identify the key factors in the system and improve the performance of the system.

Compared with other existing methods, the biggest contribution of this paper is to consider the differences between the overall evaluations of experts. Among them, the IFN score function can effectively measure the superiority and inferiority of IFN, and the 2-DCC can effectively measure the horizontal correlation and vertical correlation between matrices. Taking into account the differences in the overall evaluations of the experts will balance the impact of the differences in the evaluations and make the results more reliable. Secondly, the negation of BPA and total uncertainty are used to deal with the uncertainty of the expert evaluation, in which the total uncertainty can effectively quantify the uncertainty of the evaluation and the weighted average evidence obtained by negation of BPA is effectively applied to deal with the conflict between evaluations. Finally, the feasibility and superiority of this method are verified by two applications.

In future research, this method can also be applied to other uncertain fields, such as cost estimation and fault diagnosis. Considering differences in overall information and balancing their effects can make the results more accurate and realistic. In addition, due to the wide application of fuzzy sets, the two-dimensional correlation coefficient can be considered for the measurement of fuzzy numbers (Li and Wei 2020). In this way, the difference between fuzzy numbers can be measured on the premise of retaining the ambiguity of information, which is more consistent with its authenticity.