1 Introduction

Decision-making is the process of identifying and choosing alternatives based on the values and preferences of the decision-makers. Because of the uncertainty and imprecision of real-life decision problems, a variety of methods have been proposed in the literature, such as multi-criteria decision-making (MCDM) [11, 16, 27, 46] and uncertainty decision-making [4952]. MCDM is an important component of decision theory and method research, which has been studied and applied to many fields, including agriculture [11], medical care [9], economics [41], human resources [30], and investment [60].

In traditional MCDM problems, a decision-maker uses precise numbers to express his or her preferences. However, due to the increasing complexity and uncertainty of MCDM problems, decision-makers are no longer able to reduce decisions to finite, precise numbers. Instead, they can provide a range of numbers, also called interval grey numbers [7], to evaluate alternative decisions based on uncertain information. For example, a decision-maker can use the age range 18–25 to classify a person as a “youth”. In this instance, the interval age range may provide a better representation than would a single year. Nevertheless, in some cases interval grey numbers cannot fully depict the information available in a discrete grey number. For instance, if a decision-maker wants to choose a single value from the choices 18, 20, and 25, the discrete grey number \(\{ 18,20,25\}\) would be appropriate, whereas the interval grey number \([18,25]\) cannot accurately represent this situation.

Liu [25] explained that discrete grey numbers can coexist with interval grey numbers, and Yang [56] proposed the use of extended grey numbers (EGNs), which combine both intervals and discrete sets of numbers and can therefore express uncertainty in a more powerful way. Suppose a company has two channels through which to add investments: in one channel, stockholders add 1000 or 2000 dollars, and in the other channel the employees collect 3000 to 4000 dollars. In this case, the EGN \(\{ 1000,2000\} \cup [3000,4000]\) expresses the total investment, which neither the discrete grey number nor the interval grey number do individually. Essentially, grey numbers—including discrete grey numbers, interval grey numbers, and EGNs—are the most basic concepts of grey systems theory, and they are more flexible and practical in dealing with problems involving uncertain information.

Since Deng [7, 8] first proposed grey systems theory, it has been an excellent tool in researching grey MCDM problems. At present, grey MCDM problems have been widely studied and applied to various fields, such as engineering design, economics, supply chain management, and water source protection. Stanujkic [34] proposed an extended multi-objective optimization on the basis of ratio analysis (MOORA) method, in which the criteria values are interval grey numbers. Wang et al. [47] defined the expected probability degree and proposed a grey stochastic MCDM method, in which criteria weights are incompletely certain and alternatives’ criteria values take the form of interval grey numbers. Su et al. [35] proposed a novel hierarchical grey decision-making trial and evaluation laboratory method to identify and analyze criteria values in incomplete information. Chithambaranathan et al. [6] integrated interval grey numbers with ELimination and Choice Translating REality (ELECTRE) and VIsekriterijumska optimizacija i KOmpromisno Resenje (VIKOR) to evaluate the environmental performance of service supply chains. Kuang et al. [17] integrated continuous grey numbers with linguistic expressions and proposed an approach based on Preference Ranking Organization Method for Enrichment Evaluation II (PROMETHEE II) for handling uncertainty in MCDM. The aforementioned grey MCDM problems generally express criteria values as interval grey numbers. However, limited research has been conducted on grey MCDM problems where criteria values are expressed as EGNs.

In some practical MCDM problems, decision-makers may find that the criteria values are random variables with unknown probability density functions or probability mass functions. These problems are called stochastic MCDM problems, and they have extensive application backgrounds [23, 27, 37, 48, 55]. Liu et al. [23] proposed a method based on prospect theory to solve stochastic MCDM problems with interval probability, in which criteria values take the form of uncertain linguistic variables. Okul et al. [27] proposed a method for addressing stochastic MCDM problems by combining the Stochastic Multi-criteria Acceptability Analysis (SMAA) and the Technique for Order Preference by Similarity to Ideal Solution (TOPSIS) method. Tan et al. [37] proposed another method based on prospect stochastic dominance for solving discrete stochastic MCDM problems with aspiration levels. Most existing studies that solve stochastic MCDM problems use real numbers as the criteria values, while only a few studies use grey numbers. Owing to the increasing complexity and uncertainty of decision problems, MCDM methods in which criteria values are simultaneously random and grey are increasingly used in the practical decision-making process.

Currently available methods for solving MCDM problems can be classified into two categories: methods based on complete rationality and methods based on bounded rationality. Within these two categories, more specific decision-making models and methods are based on the rational choice model using Expected Utility Theory (EUT). These can be further subdivided into three types as follows. (1) The first subtype includes methods based on distance measures, including TOPSIS [13] and VIKOR [28]. The TOPSIS method determines the solution with the shortest distance from the ideal solution and the farthest distance from the negative-ideal solution; it possesses the merits of robust logical structure and simple computation procedure as well as the ability to consider the ideal and negative-ideal solution simultaneously [16]. The VIKOR method provides the maximum group utility of the majority and the minimum of the individual regret of the opponent; its calculations are simple and straightforward. (2) The second subtype of methods includes those based on aggregation operators, such as the arithmetic aggregation operator [22], the geometric aggregation operator [22], the Heronian mean (HM) operator [24, 57], the Prioritized Average (PA) operator [45, 53], the Bonferroni mean operator [3, 38, 39], and the Choquet integral operator [36, 44]. These aggregation operators, which provide a comprehensive evaluation value for each alternative, are widely used in solving MCDM problems. However, the calculations involved are complex, and different operators consistently produce different ranking results. (3) The third subtype of methods is based on outranking relation models such as ELECTRE [32, 42] and PROMETHEE [4]. The ELECTRE method considers the minimum individual regret, while the PROMETHEE method focuses on the maximum group utility.

In the aforementioned MCDM methods based on complete rationality, the decision-makers are fully rational when evaluating the given alternatives. That is to say, their information is complete, their cognition is infinite, and they have sufficient time. However, during the real-life decision-making process, decision-makers’ rationality is limited by the information they are given, their personal cognitive limitations, and the amount of time provided for them to make the decision. As a result, their real-life behaviors usually depart from the predictions using EUT. Furthermore, various contradictions to EUT have been identified, such as the Allias paradox [1] and the Ellsberg paradox [10]. Responding to the fact that people only have limited rationality in a decision-making process, Simon [33] considered and proposed the bounded rationality principle. Ever since Simon’s [33] pioneer work in the field, the concept of bounded rationality has been increasingly emphasized in MCDM methods [14, 43]. The most commonly used decision-making methods that are based on bounded rationality are prospect theory [15, 40] and regret theory [2, 26]. Both of these theories are designed to preserve the basic structure of EUT, but they also account for behaviors not predicted by EUT [18]. For example, in prospect theory, researchers use value and probability weight, respectively, to replace utility and probability in EUT. In regret theory, researchers compare the obtained outcome with the outcome that could have been obtained had the decision-maker chosen a different option to introduce a correction to change the utility assessment.

In prospect theory, alternatives are evaluated by different functions in terms of gains and losses with respect to one or more reference points. The calculation functions contain five parameters, \(\alpha ,\beta ,\theta ,\gamma \, and \, \delta\), which depend on the psychological behavior of the decision-makers and are therefore difficult to determine. \(\alpha\) and \(\beta\) are the concave-convex degree of the region power function of the gains and losses, \(\theta\) shows that the region value power function is steeper for losses than for gains, \(\gamma\) is the risk gain attitude coefficient, and \(\delta\) is the risk loss attitude coefficient. Unlike prospect theory, regret theory does not need to specify a reference point; furthermore, the calculation functions only contain two parameters, the risk aversion coefficient and the regret aversion coefficient, resulting in more extensive applications. The regret reflects the difference in decision-makers’ positions by choosing one of the unselected alternatives instead of the selected alternative. Rejoice reflects the additional pleasure gained from knowing that the best alternative was selected.

Many studies have focused on regret theory because of its descriptive power. Özerol et al. [29] investigated the findings of regret theory and explored the parallel between regret theory and the PROMETHEE II method. Zhang et al. [58] proposed a decision analysis method based on regret theory to solve stochastic MCDM problems in which both the criteria values and probabilities of the states take the form of interval numbers. Zhang et al. [59] considered the regret aversion characteristic of decision-makers’ psychological behavior and proposed a method for dealing with stochastic MCDM problems. However, few attempts have been made to apply regret theory to solving grey stochastic MCDM problems.

As a result, this paper proposes a stochastic MCDM method using regret theory, in which criteria values take the form of EGNs. For simplicity, the classical TOPSIS method is combined with the aforementioned grey stochastic MCDM method. Other methods may be substituted and studied in the future. In the proposed method, the utility value function and regret value function of EGNs are constructed, and then the perceived utility value is obtained by calculating the sum of the utility value and regret value for the given criteria. By using the function of interval probability, the overall perceived utility of alternatives is calculated in the form of interval grey numbers. Based on the TOPSIS method, all alternatives are ranked according to their relative closeness. The contributions and benefits of the proposed method include (1) applying EGNs to evaluate the incomplete information, (2) modeling the decision problem while considering the decision-makers’ regret and rejoice values, and (3) combining regret theory with the TOPSIS method to solve grey stochastic MCDM problems.

This paper is organized as follows: Sect. 2 reviews key concepts, including interval probability, grey numbers and EGNs, grey random variables, and regret theory. Section 3 focuses on the grey stochastic MCDM approach based on regret theory and TOPSIS, with criteria values taking the form of EGNs. Section 4 provides two specific examples and compares our experimental results with results from the existing methods. Finally, Sect. 5 discusses conclusions.

2 Preliminaries

This section reviews the concepts of interval probability, grey numbers and EGNs, grey random variables, and regret theory.

2.1 Interval probability

Consider a variable \(x\) taking its values in a finite set \(X = \{ x_{1} ,x_{2} , \ldots ,x_{n} \}\) and a set of intervals \([L_{i} ,U_{i} ] \, (i = 1,2, \ldots ,n)\) satisfying \(0 \le L_{i} \le U_{i} \le 1 \, (i = 1,2, \ldots ,n)\). We can interpret these intervals as interval probability.

Definition 1 [12]:

A set of intervals \([L_{i} ,U_{i} ] \, (i = 1,2, \ldots ,n)\) with \(0 \le L_{i} \le U_{i} \le 1 \, (i = 1,2, \ldots ,n)\) that describe the probability of a fundamental event is called a \(n\)-dimensional probability interval \((n - PRI)\). Let \(L = (L_{1} ,L_{2} , \ldots ,L_{n} )\) and \(U = (U_{1} ,U_{2} , \ldots ,U_{n} )\), then \(n - PRI\) can be denoted as \(n - PRI(L,U)\).

Definition 2 [12]:

Let \(p_{1} ,p_{2} , \ldots ,p_{n}\) be a group of positive real numbers. If \(L_{i} \le p_{i} \le U_{i} \, (i = 1,2, \ldots ,n)\) and \(\sum\nolimits_{i = 1}^{n} {p_{i} = 1}\), then \(n - PRI(L,U)\) is reasonable. Otherwise, \(n - PRI(L,U)\) is unreasonable.

Lemma 1 [54]:

A \(n - PRI(L,U)\) is reasonable, iff \(\sum\nolimits_{i = 1}^{n} {L_{i} \le 1 \le } \sum\nolimits_{i = 1}^{n} {U_{i} }\).

Definition 3 [54]:

If a \(n - PRI(L,U)\) is reasonable, the probability intervals \([L_{i} ,U_{i} ] { }(i = 1,2, \ldots ,n)\) can be transformed into more precise probability intervals \([\overline{L}_{i} ,\overline{U}_{i} ] { }(i = 1,2, \ldots ,n)\). The formulae are given as follows:

$$\overline{L}_{i} = \hbox{max} \left( {L_{i} ,1 - \sum\limits_{j = 1,j \ne i}^{n} {U_{j} } } \right),\;\,\overline{U}_{i} = \hbox{min} \left( {U_{i} ,1 - \sum\limits_{j = 1,j \ne i}^{n} {L_{j} } } \right).$$

2.2 Grey numbers and extended grey numbers

A grey number is an uncertain number, which is usually represented as closed intervals and used to describe insufficient and incomplete information.

Definition 4 [25]:

A grey number is a number with clear upper and lower boundaries but has an unknown position within these boundaries, the mathematical expression of which is

$${\tilde{ \otimes }} = [a,b] .$$
(1)

Here, \({\tilde{ \otimes }}\) is a grey number, and \(a\) and \(b\) are the upper and lower limits of the given information, \(a,b \in R\). A grey number represents the range of the possible variance of the underlying number, which is the same as an interval value with the same upper and lower limit. Thus, it can also be called an interval grey number.

Definition 5 [19]:

Let \({\tilde{ \otimes }}_{1} \in [a,b]\) and \({\tilde{ \otimes }}_{2} \in [c,d]\) be two interval grey numbers, and let \(l({\tilde{ \otimes }}_{1} ) = b - a\) and \(l({\tilde{ \otimes }}_{2} ) = d - c\) be the lengths of the interval grey numbers, respectively. In this case, the expected probability degree of \({\tilde{ \otimes }}_{1}\) against \({\tilde{ \otimes }}_{2}\) is defined as follows:

$$P({\tilde{ \otimes }}_{1} \ge {\tilde{ \otimes }}_{2} ) = \hbox{max} \left\{ {1 - \hbox{max} \left( {\frac{d - a}{{l({\tilde{ \otimes }}_{1} ) + l({\tilde{ \otimes }}_{2} )}},0} \right),0} \right\}.$$
(2)

Definition 6 [19]:

The relationship between \({\tilde{ \otimes }}_{1}\) and \({\tilde{ \otimes }}_{2}\) can be determined as follows:

(1) If \(P({\tilde{ \otimes }}_{1} \ge {\tilde{ \otimes }}_{2} ) < 0.5\), then \({\tilde{ \otimes }}_{1}\) is less than \({\tilde{ \otimes }}_{2}\), which can be denoted as \({\tilde{ \otimes }}_{1} < {\tilde{ \otimes }}_{2}\).

(2) If \(P({\tilde{ \otimes }}_{1} \ge {\tilde{ \otimes }}_{2} ) = 0.5\), then \({\tilde{ \otimes }}_{1}\) is equal to \({\tilde{ \otimes }}_{2}\), which can be denoted as \({\tilde{ \otimes }}_{1} = {\tilde{ \otimes }}_{2}\).

(3) If \(P({\tilde{ \otimes }}_{1} \ge {\tilde{ \otimes }}_{2} ) > 0.5\), then \({\tilde{ \otimes }}_{1}\) is larger than \({\tilde{ \otimes }}_{2}\), which can be denoted as \({\tilde{ \otimes }}_{1} > {\tilde{ \otimes }}_{2}\).

Definition 7 [21]:

Let \({\tilde{ \otimes }}_{1} \in [a,b]\) and \({\tilde{ \otimes }}_{2} \in [c,d]\) be two interval grey numbers. The Euclidean distance between \({\tilde{ \otimes }}_{1}\) and \({\tilde{ \otimes }}_{2}\) can be calculated as follows:

$$d\left( {{\tilde{ \otimes }}_{1} ,{\tilde{ \otimes }}_{2} } \right) = \sqrt {\frac{1}{2}\left[ {\left( {a - c} \right)^{2} + \left( {b - d} \right)^{2} } \right]} .$$
(3)

Grey numbers are usually represented as closed intervals, but continuous intervals cannot fully reveal the uncertainty of the given information. In light of this, Yang [56] provided a new definition of grey numbers that considers both continuous and discrete grey numbers.

Definition 8 [56]:

Let \(\otimes\) be a union set of closed or open intervals. An EGN can be represented as follows:

$$\otimes = \bigcup\limits_{i = 1}^{n} {[a_{i} ,b_{i} ]} .$$
(4)

Here, \(i = 1,2, \ldots ,n\), \(n\) is an integer, and \(0 < n < \infty\), \(a_{i} ,b_{i} \in R\) and \(b_{i - 1} < a_{i} \le b_{i} < a_{i + 1}\). The set of all EGNs that meet Eq. (4) is represented as \(R( \otimes )\).

Theorem 1:

Let \(\otimes\) be an EGN. The following properties are true:

(1) \(\otimes\) is a continuous EGN \(\otimes = [a_{1} ,b_{n} ]\) iff \(a_{i} \le b_{i - 1}\) (\(\forall i > 1\)) or \(n = 1\);

(2) \(\otimes\) is a discrete EGN \(\otimes = \{ a_{1} ,a_{2} , \cdots ,a_{n} \}\) iff \(a_{i} = b_{i}\);

(3) \(\otimes\) is a mix EGN iff only part of its intervals shrink to crisp numbers and the others remain as intervals.

Definition 9 [56]:

Let \(\otimes_{1} = \bigcup\nolimits_{i = 1}^{n} {[a_{i} ,b_{i} ]}\) and \(\otimes_{2} = \bigcup\nolimits_{j = 1}^{m} {[c_{j} ,d_{j} ]}\) be two EGNs, \(a_{i} \le b_{i} (i = 1,2, \ldots ,n)\), \(c_{j} \le d_{j} (j = 1,2, \ldots ,m)\), \(\lambda \in R,\lambda \ge 0\), then:

(1) \(\otimes_{1} + \otimes_{2} = \bigcup\nolimits_{i = 1}^{n} {\bigcup\nolimits_{j = 1}^{m} {[a_{i} + c_{j} ,b_{i} + d_{j} ]} } ;\)

(2) \(- \otimes_{1} = \bigcup\nolimits_{i = 1}^{n} {[ - b_{i} , - a_{i} ]} ;\)

(3) \(\otimes_{1} - \otimes_{2} = \bigcup\nolimits_{i = 1}^{n} {\bigcup\nolimits_{j = 1}^{m} {[a_{i} - d_{j} ,b_{i} - c_{j} ]} } ;\)

(4) \(\otimes_{1} \times \otimes_{2} = \bigcup\nolimits_{i = 1}^{n} {\bigcup\nolimits_{j = 1}^{m} {[\hbox{min} \{ a_{i} c_{j} ,a_{i} d_{j} ,b_{i} c_{j} ,b_{i} d_{j} \} ,\hbox{max} \{ a_{i} c_{j} ,a_{i} d_{j} ,b_{i} c_{j} ,b_{i} d_{j} \} ]} } ;\)

(5) \(\frac{{ \otimes_{1} }}{{ \otimes_{2} }} = \bigcup\nolimits_{i = 1}^{n} {\bigcup\nolimits_{j = 1}^{m} {[\hbox{min} \{ \frac{{a_{i} }}{{c_{j} }},\frac{{a_{i} }}{{d_{j} }},\frac{{b_{i} }}{{c_{j} }},\frac{{b_{i} }}{{d_{j} }}\} ,\hbox{max} \{ \frac{{a_{i} }}{{c_{j} }},\frac{{a_{i} }}{{d_{j} }},\frac{{b_{i} }}{{c_{j} }},\frac{{b_{i} }}{{d_{j} }}\} ]} } (c_{j} \ne 0,d_{j} \ne 0(j = 1,2, \ldots ,m));\)

(6) \(\lambda \otimes_{1} = \bigcup\nolimits_{i = 1}^{n} {[\lambda a_{i} ,\lambda b_{i} ]} ;\)

(7) \(\otimes_{1}^{\lambda } = \bigcup\nolimits_{i = 1}^{n} {[min(a_{i}^{\lambda } ,b_{i}^{\lambda } ),max(a_{i}^{\lambda } ,b_{i}^{\lambda } )]} .\)

The expected probability degree of EGNs is provided, inspired by the expected probability degree of the grey numbers.

Definition 10:

Let \(\otimes_{1} = \bigcup\nolimits_{i = 1}^{n} {[a_{i} ,b_{i} ]}\) and \(\otimes_{2} = \bigcup\nolimits_{j = 1}^{m} {[c_{j} ,d_{j} ]}\) be two EGNs, \(\otimes_{1} , \otimes_{2} \in R( \otimes )\). The expected probability degree of \(\otimes_{1}\) against \(\otimes_{2}\) is defined as follows:

$$P( \otimes_{1} \ge \otimes_{2} ) = \frac{1}{n \times m}\sum\limits_{i = 1}^{n} {\sum\limits_{j = 1}^{m} {P\left( {[a_{i} ,b_{i} ] \ge [c_{j} ,d_{j} ]} \right)} } .$$
(5)

Definition 11:

Let \(\otimes_{1} = \bigcup\nolimits_{i = 1}^{n} {[a_{i} ,b_{i} ]}\) and \(\otimes_{2} = \bigcup\nolimits_{j = 1}^{m} {[c_{j} ,d_{j} ]}\) be any two EGNs, \(\otimes_{1} , \otimes_{2} \in R( \otimes )\). The order relationship between these EGNs is defined as follows:

(1) If \(P( \otimes_{1} \ge \otimes_{2} ) < 0.5\), then \(\otimes_{1}\) is less than \(\otimes_{2}\), which can be denoted as \(\otimes_{1} < \otimes_{2}\).

(2) If \(P( \otimes_{1} \ge \otimes_{2} ) = 0.5\), then \(\otimes_{1}\) is equal to \(\otimes_{2}\), which can be denoted as \(\otimes_{1} = \otimes_{2}\).

(3) If \(P( \otimes_{1} \ge \otimes_{2} ) > 0.5\), then \(\otimes_{1}\) is larger than \(\otimes_{2}\), which can be denoted as \(\otimes_{1} > \otimes_{2}\).

2.3 Grey random variables

Discrete grey random variables, called grey random variables in this paper, are a group of random variables made up of a finite set of EGNs. A grey random variable is denoted as \(\xi ( \otimes )\), and the probability distribution of \(\xi ( \otimes )\) is shown in Table 1.

Table 1 The probability distribution of \(\xi ( \otimes )\)

In Table 1, \(\otimes_{i}\) is the ith value that would be taken by \(\xi ( \otimes )\), while \(n\) is the number of values that a grey random variable can have. \(p_{i}\) is the probability with respect to \(\otimes_{i}\), and \(\sum\nolimits_{i = 1}^{n} {p_{i} } = 1\). The probability distribution function can be denoted as \(f(\xi ( \otimes ) = \otimes_{i} ) = p_{i}\).

2.4 Regret theory

Loomes and Sugden [26] and Bell [2] introduced regret theory separately. They defined regret as a reflection of the difference in a decision-maker’s position resulting from choosing one of the unselected alternatives instead of the selected alternative. They defined rejoice as a reflection of the additional pleasure gained from knowing that the best alternative was selected.

In regret theory, these authors developed a modified utility function to measure the expected value of satisfaction derived from choosing alternative \(A\) and rejecting alternative \(B\). Let \(x\) and \(y\) be the consequence of choosing alternative \(A\) and \(B\), respectively, then the modified utility function of achieving \(x\) can be denoted as

$$u(x,y) = v(x) + R\left( {v(x) - v(y)} \right) .$$
(6)

Here, \(v(x)\) represents the utility value that the decision-maker would derive from consequence \(A\) if he experiences it without having to choose. \(R\left( {v(x) - v(y)} \right)\) indicates the regret-rejoice value, and the difference \(v(x) - v(y)\) is the loss/gain value of having chosen \(A\) rather than the forgone choice \(B\). The regret-rejoice function \(R( \cdot )\) is monotonically increasing and decreasingly concave, with \(R^{\prime}( \cdot ) > 0\), \(R^{\prime\prime}( \cdot ) < 0\) and \(R(0) = 0\).

The regret theory was initially derived for pair-wise choices, and since then it has been extended to general choice sets [31]. Let \(x_{1} ,x_{2} , \ldots ,x_{m}\) be the consequence of choosing alternatives \(A_{1} ,A_{2,} \ldots ,A_{m}\), respectively. \(A_{i}\) is the ith alternative, and \(x_{i}\) is the consequence of \(A_{i}\). The modified utility function of achieving \(x_{i}\) can be denoted as

$$u_{i} = v(x_{i} ) + R\left( {v(x_{i} ) - v(x^{*} )} \right) .$$
(7)

Here, \(x^{*} = \hbox{max} \{ x_{i} \left| {i = 1,2, \ldots ,m} \right.\}\). \(R\left( {v(x_{i} ) - v(x^{*} )} \right)\) indicates the regret value, which is always non-positive. Rational decision-makers choose the optimal alternative by maximizing their expected modified utility of all possible alternatives.

3 Grey stochastic MCDM approach based on regret theory and TOPSIS

This section introduces the grey stochastic MCDM approach based on regret theory and TOPSIS.

3.1 Description of the decision-making problem

Consider the following grey stochastic MCDM problem: assume that \(A = \{ a_{1} ,a_{2} , \ldots ,a_{m} \}\) is a discrete alternative set of \(m\) possible alternatives, and that \(C = \{ c_{1} ,c_{2} , \ldots ,c_{n} \}\) is a set of \(n\) criteria. The weighted vector of the criteria is \(W = (w_{1} ,w_{2} , \ldots ,w_{n} )\), where \(w_{j} \in [ 0 , 1 ] { }\) and \(\sum\nolimits_{j = 1}^{n} {w_{j} } = 1\).

Due to the uncertainty of the decision-making environment, the alternatives would have some possible status. Let \(\varTheta_{j} = (\theta_{1} ,\theta_{2} , \ldots ,\theta_{{l_{j} }} )\) be the possible status belonging to the criterion \(c_{j}\), and let \(p_{j}^{t} = [p_{j}^{Lt} ,p_{j}^{Ut} ]\) be the interval probability of the status \(\theta_{t} \, (1 \le t \le l_{j} )\) that belongs to the criterion \(c_{j}\), where \(0 \le p_{j}^{Lt} \le p_{j}^{Ut} \le 1\), \(\sum\nolimits_{t = 1}^{{l_{j} }} {p_{j}^{Lt} } \le 1\) and \(\sum\nolimits_{t = 1}^{{l_{j} }} {p_{j}^{Ut} } \ge 1\). Suppose that the characteristics of the alternative \(a_{i}\) with respect to \(c_{j}\) are represented by the EGN \(\otimes u_{ij}\), the t-th status of which is denoted as \(\otimes u_{ij}^{t} = \bigcup\nolimits_{k = 1}^{l} {[a_{ijk}^{t} ,b_{ijk}^{t} ]}\), and \(a_{ij1}^{t} \le b_{ij1}^{t} < a_{ij2}^{t} \le b_{ij2}^{t} < \cdots < a_{ijl}^{t} \le b_{ijl}^{t}\). The grey stochastic decision matrix can be represented as \(R^{t} = \left( { \otimes u_{ij}^{t} } \right)_{m \times n}\). We are then able to rank the alternatives according to the presented information.

3.2 Decision-making procedure

For the aforementioned grey stochastic MCDM problem, the solving procedure can be summarized as follows.

  1. Step 1:

    Normalize the decision matrix.

    In order to eliminate the influence of different dimensions of physical quantity, the decision matrix needs to be normalized so as to transform the various criteria values into comparable values. The criteria are normally classified into two types: criteria of the maximizing type and criteria of the minimizing type. If the criteria are of the maximizing type, the transforming formula is as follows:

    $$r_{ij}^{t} = \frac{{ \otimes u_{ij}^{t} }}{{b_{ijk}^{t(max)} }} = \bigcup\limits_{k = 1}^{l} {\left[ {\frac{{a_{ijk}^{t} }}{{b_{ijk}^{t(max)} }},\frac{{b_{ijk}^{t} }}{{b_{ijk}^{t(max)} }}} \right]} .$$
    (8)

    Here, \(b_{ijk}^{t(max)} = \mathop {\hbox{max} }\nolimits_{1 \le k \le l,1 \le i \le m} b_{ijk}^{t}\).

    If the criteria are of the minimizing type, the formula is as follows:

    $$r_{ij}^{t} = \frac{{a_{ijk}^{t(\hbox{min} )} }}{{ \otimes u_{ij}^{t} }} = \bigcup\limits_{k = 1}^{l} {\left[ {\frac{{a_{ijk}^{t(\hbox{min} )} }}{{b_{ijk}^{t} }},\frac{{a_{ijk}^{t(\hbox{min} )} }}{{a_{ijk}^{t} }}} \right]} .$$
    (9)

    Here, \(a_{ijk}^{t(\hbox{min} )} = \mathop {\hbox{min} }\nolimits_{1 \le k \le l,1 \le i \le m} a_{ijk}^{t}\).

    The normalized decision matrix is denoted by \(N^{t} = \left( {r_{ij}^{t} } \right)_{m \times n}\), where \(r_{ij}^{t} = \bigcup\nolimits_{k = 1}^{l} {[\underline{r}_{ijk}^{t} ,\overline{r}_{ijk}^{t} ]}\).

  2. Step 2:

    Identify the ideal point.

    Identify the ideal point \(I = (I_{1}^{ + } ,I_{2}^{ + } , \ldots ,I_{n}^{ + } )\). Let \(I_{j}^{ + } = (r_{j}^{1 + } ,r_{j}^{2 + } , \ldots ,r_{j}^{{l_{j} + }} )\) be the ideal point related to criterion \(c_{j}\), and \(1 \le t \le l_{j}\); then \(r_{j}^{t + } = [\underline{r}_{j}^{t + } ,\overline{r}_{j}^{t + } ]\) can be calculated according to the following formula:

    $$\underline{r}_{j}^{t + } = \overline{r}_{j}^{t + } = \hbox{max} \{ \overline{r}_{ijk}^{t} \left| {1 \le i \le m,1 \le j \le n,1 \le t \le l_{j} } \right.,1 \le k \le l\} .$$
    (10)
  3. Step 3

    Calculate the utility values and regret values concerning the criteria.

    1. (a)

      The utility function should be built before calculating the criteria’s utility values. Due to the decision-maker’s risk aversion, the utility function \(v(x)\) is monotonically increasing concave, with \(v^{\prime}(x) > 0\) and \(v^{\prime\prime}(x) < 0\). Here, the power function is used as the criteria value of the utility function.

      $$v(x) = x^{\alpha } ,$$
      (11)

      where \(\alpha\) is the coefficient of risk aversion and \(0 < \alpha < 1\). If parameter \(\alpha\) is smaller, the decision-maker’s risk aversion is greater.

      For the criteria value \(r_{ij}^{t} = \bigcup\nolimits_{k = 1}^{l} {[\underline{r}_{ijk}^{t} ,\overline{r}_{ijk}^{t} ]}\), the actual criteria value \(x\) is in the range of \(\bigcup\nolimits_{k = 1}^{l} {[\underline{r}_{ijk}^{t} ,\overline{r}_{ijk}^{t} ]}\) and subject to various kinds of distribution [58]. Let \(f_{ij}^{t} (x)\) be the probability density function, then the utility value \(v_{ij}^{t}\) can be calculated with the following formula:

      $$v_{ij}^{t} = \frac{1}{l}\sum\limits_{k = 1}^{l} {\int_{{\underline{r}_{ijk}^{t} }}^{{\overline{r}_{ijk}^{t} }} {v(x)f_{ijk}^{t} (x)dx} } , \, 1 \le i \le m,1 \le j \le n,1 \le t \le l_{j} ,1 \le k \le l.$$
      (12)

      Here, two kinds of distribution are taken into consideration.

    1. (1)

      Uniform distribution

      Uniform distribution is one of the most common distributions. For a grey random variable \(x\) following uniform distribution, the probability density function is

      $$f_{ijk}^{t} (x) = \left\{ \begin{array}{ll} \frac{1}{{\overline{r}_{ijk}^{t} - \underline{r}_{ijk}^{t} }},&\quad \underline{r}_{ijk}^{t} \le x \le \overline{r}_{ijk}^{t} \hfill \\ 0,&\quad {\text{otherwise}} \end{array} \right.\quad1 \le i \le m,1 \le j \le n,1 \le t \le l_{j} ,1 \le k \le l.$$
      (13)
    2. (2)

      Normal distribution

      Normal distribution is extremely important in statistics and is often used in the natural and social sciences for real-valued random variables with unknown distributions [5]. For a grey random variable \(x\) following normal distribution, the probability density function is:

      $$f_{ijk}^{t} (x) = \left\{ \begin{array}{ll} \frac{1}{{\sqrt {2\pi } \sigma_{ijk}^{t} }}exp\left[ {{\raise0.7ex\hbox{${ - (x - \mu_{ijk}^{t} )^{2} }$} \!\mathord{\left/ {\vphantom {{ - (x - \mu_{ijk}^{t} )^{2} } {2(\sigma_{ijk}^{t} )^{2} }}}\right.\kern-0pt} \!\lower0.7ex\hbox{${2(\sigma_{ijk}^{t} )^{2} }$}}} \right], &\quad \underline{r}_{ijk}^{t} \le x \le \overline{r}_{ijk}^{t} \\ 0,&\quad{\text{otherwise}} \\ \end{array} \right. \quad 1 \le i \le m,1 \le j \le n,1 \le t \le l_{j} ,1 \le k \le l.$$
      (14)

      Here, the mean value is \(\mu_{ijk}^{t} = {{( {\underline{r}_{ijk}^{t} + \overline{r}_{ijk}^{t} } )} / 2}\), and the standard deviation is \(\sigma_{ijk}^{t} = {{( {\overline{r}_{ijk}^{t} - \underline{r}_{ijk}^{t} } )} / 6}.\)

    3. (b)

      The regret-rejoice function \(R(\Delta v)\) is constructed to determine the regret values concerning the criteria; it is monotonically increasing and decreasingly concave with \(R^{\prime}( \cdot ) > 0\), \(R^{\prime\prime}( \cdot ) < 0\) and \(R(0) = 0\). \(R(\Delta v)\) can be represented as follows:

      $$R(\Delta v) = 1 - \exp ( - \delta \Delta v).$$
      (15)

      Here, \(\delta\) is the regret aversion coefficient, and \(\delta > 0\). If the value of \(\delta\) is larger, then the decision-maker tends toward regret aversion. \(\Delta v\) represents the difference between the utility values of two alternatives. When \(R(\Delta v) > 0\), \(R(\Delta v)\) represents the rejoice value; conversely, \(R(\Delta v)\) represents the regret value when \(R(\Delta v) < 0\).

      Compared with the ideal point, the regret value of alternative \(a_{i}\) with respect to \(c_{j}\) under the t-th status can be calculated as follows:

      $$R_{ij}^{t} = 1 - \exp \left[ { - \delta \left( {v_{ij}^{t} - v_{j}^{t + } } \right)} \right], \, 1 \le i \le m,1 \le j \le n,1 \le t \le l_{j} .$$
      (16)

      Here, \(v_{j}^{t + } = \int_{{\underline{r}_{j}^{t + } }}^{{\overline{r}_{j}^{t + } }} {v(x)f_{j}^{t + } (x)dx} = \left( {\overline{r}_{j}^{t + } } \right)^{\alpha }\) and \(v_{ij}^{t} \le v_{j}^{t + }\); thus \(R_{ij}^{t} \le 0\), and \(R_{ij}^{t}\) represent the regret values.

  4. Step 4:

    Calculate the overall perceived utility values of alternatives.

    The perceived utility values of alternatives can be calculated on the basis of Step 3. These represent the sum of the utility value and regret value of the alternatives. Let \(u_{ij}^{t}\) be the perceived utility value of alternative \(a_{i}\) with respect to \(c_{j}\) under the t-th status, then

    $$u_{ij}^{t} = v_{ij}^{t} + R_{ij}^{t} , { }1 \le i \le m,1 \le j \le n,1 \le t \le l_{j} .$$
    (17)

    The overall perceived utility value of alternative \(a_{i}\) with respect to \(c_{j}\) can be calculated as follows:

    $$u_{ij} = \sum\limits_{t = 1}^{{l_{j} }} {u_{ij}^{t} \overline{p}_{j}^{t} } .$$
    (18)

    Here, \(\bar{p}_{j}^{t}\) is the more precise interval probability, which can be calculated by

    $$\bar{p}_{j}^{Lt} = \hbox{max} \left( {p_{j}^{Lt} ,1 - \sum\nolimits_{{t^{\prime} = 1,t^{\prime} \ne t}}^{{l_{j} }} {p_{j}^{Ut} } } \right),$$
    (19)
    $$\bar{p}_{j}^{Ut} = \hbox{min} \left( {p_{j}^{Ut} ,1 - \sum\limits_{{t^{\prime} = 1,t^{\prime} \ne t}}^{{l_{j} }} {p_{j}^{Lt} } } \right).$$
    (20)

    The overall perceived utility matrix is denoted by \(U = (u_{ij} )_{m \times n}\), where \(u_{ij} = [\underline{u}_{ij} ,\overline{u}_{ij} ]\).

  5. Step 5:

    Prioritize the alternatives according to the overall perceived utility intervals.

    Identify the positive ideal solution \(U^{ + }\) and the negative ideal solution \(U^{ - }\) of the overall perceived utility intervals as follows:

    $$U^{ + } = (u_{1}^{ + } ,u_{2}^{ + } , \ldots ,u_{m}^{ + } ),\; U^{ - } = (u_{1}^{ - } ,u_{2}^{ - } , \ldots ,u_{m}^{ - } ).$$

    The ideal solutions are chosen according to the following formula:

    $$u_{j}^{ + } = \mathop {\hbox{max} }\limits_{i = 1}^{m} u_{ij} ,{ 1} \le i \le m,1 \le j \le n,$$
    (21)
    $$u_{j}^{ - } = \mathop {\hbox{min} }\limits_{i = 1}^{m} u_{ij} , \, 1 \le i \le m,1 \le j \le n.$$
    (22)

    Using Eq. (3), calculate the distances between each overall perceived utility interval and both the positive ideal solution \(d_{i}^{ + }\) and the negative ideal solution \(d_{i}^{ - }\).

    $$d_{i}^{ + } = \sum\limits_{j = 1}^{m} {w_{j} d(u_{ij} ,u_{j}^{ + } )} ,$$
    (23)
    $$d_{i}^{ - } = \sum\limits_{j = 1}^{m} {w_{j} d(u_{ij} ,u_{j}^{ - } )} .$$
    (24)

    Then, estimate the relative closeness with the following formula:

    $$C_{i} = \frac{{d_{i}^{ - } }}{{d_{i}^{ - } + d_{i}^{ + } }} .$$
    (25)

    The bigger \(C_{i}\) is, the better \(a_{i}\) will be.

4 Illustrative examples

This section uses two examples to perform a detailed and precise demonstration of the proposed approach. The first example shows the use of the proposed approach with interval grey numbers, which can be considered a special case of EGNs. The second example demonstrates the application of the proposed approach with general EGNs. In addition to validating the effectiveness of the proposed method, we also perform comparison analyses.

4.1 Illustration of the proposed approach

To demonstrate the use of the proposed approach with interval grey numbers, consider the problem of evaluating companies for investment decisions [20].

Example 1

An investment bank is planning to invest in three companies denoted as \(A = \{ a_{1} ,a_{2} ,a_{3} \}\). Three criteria are taken into account: \(c_{1}\), annual product income; \(c_{2}\), social benefit; and \(c_{3}\), environmental pollution degree. The criteria’s weights vector is denoted as \(W = (0.1,0.2,0.7)\). All three companies have three possible statuses: good, \(\theta_{1}\); fair, \(\theta_{2}\); and poor, \(\theta_{3}\). The probability of each status is expressed as an interval probability, the values of which are \(P_{1} = [0.3,0.5]\), \(P_{2} = [0.4,0.9]\), and \(P_{3} = [0.1,0.5]\). The criteria value of each alternative takes the form of interval grey numbers, and the grey random variable is subject to normal distribution. Table 2 lists the associated assessment values. It is possible to choose the best alternative according to the information provided.

Table 2 Assessment values of alternatives under each criterion

The following procedure yields the most desirable alternative.

  1. Step 1:

    Normalize the decision matrix.

    In Example 1, \(c_{1}\) and \(c_{2}\) are of the maximizing type, while \(c_{3}\) is of the minimizing type. Applying Eqs. (8) and (9) yields the normalized decision matrix, as shown in Table 3.

    Table 3 The normalized decision matrix
  2. Step 2:

    Identify the ideal point.

    Applying Eq. (10) to the data shown in Table 3 yields the absolute ideal point \(I = (I_{1}^{ + } ,I_{2}^{ + } , \ldots ,I_{n}^{ + } )\). Here, \(I_{j}^{ + } = (r_{j}^{1 + } ,r_{j}^{2 + } , \ldots ,r_{j}^{{l_{j} + }} )\), and \(r_{j}^{t + } = [\underline{r}_{j}^{t + } ,\overline{r}_{j}^{t + } ] = [1,1] \, (1 \le j \le n,1 \le t \le l_{j} )\).

  3. Step 3:

    Calculate the utility values and regret values concerning the criteria.

    Calculating the utility values and regret values concerning the criteria requires consideration of two parameters, namely the risk aversion coefficient and the regret aversion coefficient. According to Tversky and Kahneman [40], the coefficient of risk aversion is \(\alpha = 0.88\). Assume the regret aversion coefficient is \(\delta = 0.3\); other values of \(\delta\) will be discussed later. The utility values and regret values concerning the criteria can be determined using Eqs. (11), (12), (14), and (16). For example, \(v_{11}^{1} = 0.860\), \(R_{11}^{1} = - 0.043\), \(v_{21}^{1} = 0.982\), and \(R_{21}^{1} = - 0.005\). All criteria values are compared with the absolute ideal point, so the regret values concerning the criteria are non-positive.

  4. Step 4:

    Calculate the overall perceived utility values of alternatives.

    By applying Eq. (17), the utility values and regret values calculated above can be added to get the perceived utility values for the alternatives, which are shown in Table 4.

    Table 4 The perceived utility values matrix

    More precise interval probabilities can be found using Eqs. (19) and (20), with the results \(\overline{p}_{1} = [0.3,0.5]\), \(\overline{p}_{2} = [0.4,0.6]\), and \(\overline{p}_{3} = [0.1,0.3]\). Table 5 shows the overall perceived utility values calculated using Eq. (18).

    Table 5 The overall perceived utility values matrix
  5. Step 5:

    Prioritize the alternatives according to the overall perceived utility intervals.

    First, identify the positive ideal solution \(U^{ + }\) and the negative ideal solution \(U^{ - }\) of the overall perceived utility intervals:

    $$u_{1}^{ + } = [0.751,1.312],\;u_{2}^{ + } = [0.719,1.267],\; u_{3}^{ + } = [0.577,0.989],$$
    $$u_{1}^{ - } = [0.659,1.154],\;u_{2}^{ - } = [0.584,1.024],\; u_{3}^{ - } = [0.336,0.647].$$

    Next, calculate the distances between each overall perceived utility interval and both the positive ideal solution \(d_{i}^{ + }\) and the negative ideal solution \(d_{i}^{ - }\) using Eqs. (23) and (24):

    $$d_{1}^{ + } = 0.218,\;d_{2}^{ + } = 0.113,\;d_{3}^{ + } = 0.052,$$
    $$d_{1}^{ - } = 0.042,\;d_{2}^{ - } = 0.148,\;d_{3}^{ - } = 0.207.$$

    Finally, apply Eq. (25) to obtain the relative closeness:

    $$C_{1} = 0.160,\;C_{2} = 0.567,\;C_{3} = 0.799.$$

    \(C_{3} > C_{2} > C_{1}\) thus determining that the best alternative is \(a_{3}\).

    The value of coefficient \(\delta\) reflects the degree of the decision-makers’ regret aversion. Table 6 shows the relative closeness and ranking results using different \(\delta\) values.

    Table 6 The relative closeness and ranking results using different \(\delta\) values

    Table 6 shows that the ranking results remain the same when different \(\delta\) values are used. However, in some cases the ranking results may change as the \(\delta\) value changes.

Example 2

Assume the assessment values of the alternatives under each criterion in Example 1 take the form of general EGNs, and the grey random variable is subject to uniform distribution. The associated assessment values are shown in Table 7.

Table 7 Assessment values of alternatives under each criterion

The following procedure will yield the most desirable alternative.

  1. Step 1:

    Normalize the decision matrix.

    In Example 2, \(c_{1}\) and \(c_{2}\) are of the maximizing type, while \(c_{3}\) is of the minimizing type. Applying Eqs. (8) and (9) yields the normalized decision matrix, as shown in Table 8.

    Table 8 The normalized decision matrix
  2. Step 2:

    Identify the ideal point.

    The absolute ideal point \(I = (I_{1}^{ + } ,I_{2}^{ + } , \ldots ,I_{n}^{ + } )\) can be obtained from the data shown in Table 8 by applying Eq. (10). Here, \(I_{j}^{ + } = (r_{j}^{1 + } ,r_{j}^{2 + } , \ldots ,r_{j}^{{l_{j} + }} )\) and \(r_{j}^{t + } = [\underline{r}_{j}^{t + } ,\overline{r}_{j}^{t + } ] = [1,1] \, (1 \le j \le n,1 \le t \le l_{j} )\).

  3. Step 3:

    Calculate the utility values and regret values concerning the criteria.

    The parameters used in this example are again \(\alpha = 0.88\) and \(\delta = 0.3\). The utility values and regret values concerning the criteria can be obtained using Eqs. (11), (12), (13), and (16). For example, \(v_{11}^{1} = 0.807\), \(R_{11}^{1} = - 0.060\), \(v_{21}^{1} = 0.752\), and \(R_{21}^{1} = - 0.078\). All the criteria values are compared with the absolute ideal point, so the regret values concerning the criteria are non-positive.

  4. Step 4:

    Calculate the overall perceived utility values of alternatives.

    The utility values and regret values can be calculated using Eq. (17), and then added to get the perceived utility values of alternatives, which are shown in Table 9.

    Table 9 The perceived utility values matrix

    Similar to Example 1, the more precise interval probabilities are \(\overline{p}_{1} = [0.3,0.5]\), \(\overline{p}_{2} = [0.4,0.6]\), and \(\overline{p}_{3} = [0.1,0.3]\). The overall perceived utility values are calculated using Eq. (18), as shown in Table 10.

    Table 10 The overall perceived utility values matrix
  5. Step 5

    Prioritize the alternatives according to the overall perceived utility intervals.

    First, identify the positive ideal solution \(U^{ + }\) and the negative ideal solution \(U^{ - }\) of the overall perceived utility intervals as follows:

    $$u_{1}^{ + } = [0.716,1.251],\;u_{2}^{ + } = [0.737,1.288],\; u_{3}^{ + } = [0.238,0.450],$$
    $$u_{1}^{ - } = [0.566,1.005],\;u_{2}^{ - } = [0.530,0.953],\; u_{3}^{ - } = [0.118,0.253].$$

    Next, calculate the distances between each overall perceived utility interval to both the positive ideal solution \(d_{i}^{ + }\) and the negative ideal solution \(d_{i}^{ - }\) using Eqs. (23) and (24):

    $$d_{1}^{ + } = 0.057,\;d_{2}^{ + } = 0.135,\;d_{3}^{ + } = 0.056,$$
    $$d_{1}^{ - } = 0.138,\;d_{2}^{ - } = 0.056,\;d_{3}^{ - } = 0.135.$$

    Finally, apply Eq. (25) to obtain the relative closeness:

    $$C^{\prime}_{1} = 0.708,\;C^{\prime}_{2} = 0.293,\;C^{\prime}_{3} = 0.707.$$

    \(C^{\prime}_{1} > C^{\prime}_{3} > C^{\prime}_{2}\), thus the best alternative is \(a^{\prime}_{1}\).

    Table 11 shows the relative closeness and ranking results using different \(\delta\) values.

    Table 11 Relative closeness and ranking results using different \(\delta\) values

    Table 11 shows that the ranking of \(a^{\prime}_{2}\) is certain, but whether \(a^{\prime}_{1}\) or \(a^{\prime}_{3}\) is the better alternative depends on the \(\delta\) value. When \(\delta = 0.1\), \(a^{\prime}_{3}\) is superior to \(a^{\prime}_{1}\), but the situation reverses when \(\delta \ge 0.2\). This shows that the decision-makers’ regret aversion exerts an impact on the ranking of the alternatives.

4.2 Comparison analysis and discussion

A comparison analysis is performed to verify the validity and feasibility of the proposed decision-making method. The first type of method uses the prospect theory and VIKOR, which is also based on bounded rationality. The second type of method uses the expected probability degree, which is based on complete rationality.

  1. 1.

    Comparison of the proposed approach with the method based on prospect theory and VIKOR

    Li and Zhao [20] proposed a VIKOR method based on prospect theory to achieve a comprehensive look at the alternatives. After normalization, the prospect value function of the interval grey numbers was defined, and the prospect value of each alternative was calculated using all other alternatives as a reference point. After this step the positive ideal solution and the negative ideal solution can be easily identified. Finally, the \(S_{i} ( \otimes )\), \(R_{i} ( \otimes )\), and \(Q_{i} ( \otimes )\) values can be obtained as shown in Table 12.

    Table 12 \(S_{i} ( \otimes )\), \(R_{i} ( \otimes )\), and \(Q_{i} ( \otimes )\) values

    Since \(S_{3} ( \otimes ) < S_{ 2} ( \otimes ) < S_{ 1} ( \otimes )\), \(R_{3} ( \otimes ) < R_{ 2} ( \otimes ) < R_{ 1} ( \otimes )\), and \(Q_{3} ( \otimes ) < Q_{ 2} ( \otimes ) < Q_{ 1} ( \otimes )\), the definitive ranking is \(a_{3} > a_{2} > a_{1}\), and the best alternative is \(a_{3}\).

    When we use the same example, Li and Zhou’s method [20] produces the same ranking of alternatives as our proposed method. These results verify the validity of the method proposed in this paper. Though there are some differences between the two methods, both the method based on prospect theory and the proposed method based on regret theory take the decision-makers’ psychological behavior into account. One main difference between the two methods is that the decision-makers utilizing prospect theory need to give one or more reference points, which are not necessary when utilizing regret theory. Also, five parameters used in the prospect theory calculation formula are difficult to determine. Compared with prospect theory, the regret theory calculation formula involves only two parameters, giving it superior ease of use. In addition, the EGNs used in this paper are more powerful in expressing evaluation information than interval grey numbers. Thus, to some extent, the method proposed in this paper is more practical than the method proposed by Li and Zhao [20].

  2. 2.

    Comparison of the proposed approach with the method based on expected probability degree

    Wang et al. [47] proposed a grey stochastic MCDM method based on expected probability degree. For comparison, the interval probability should be transformed to the known probability by substituting each interval value with the mean value of its upper and lower limits and normalizing the values. After transformation, the probability values of the statuses in Example 1 are \(p_{1} = 0.364\), \(p_{2} = 0.455\), and \(p_{3} = 0.181\).

    In this method, the expected probability degrees of the alternative under criteria were calculated to form the expected probability degree judgment matrix. With respect to each criterion, the comprehensive judgment matrix of expected probability degrees was easily established. Then, the comprehensive sorting value was determined with the calculation formula \((\omega_{i} = \frac{{\left( {\sum\nolimits_{l = 1}^{m} {\sum\nolimits_{k = 1}^{n} {w_{k} \times E\left( {P\left( {x_{ik} > x_{lk} } \right)} \right)} } } \right) + \frac{m}{2} - 1}}{{m \times \left( {m - 1} \right)}},\) where \(E\left( {P\left( {x_{ik} > x_{lk} } \right)} \right)\) is the expected probability degree of the grey stochastic variable \(x_{ik}\) against \(x_{lk}\)). The values are \(\omega_{1} = 0.291\), \(\omega_{2} = 0.360\), and \(\omega_{3} = 0.255\). Since \(\omega_{2} > \omega_{1} > \omega_{3}\), the ranking is \(a_{2} > a_{1} > a_{3}\), and the best alternative is \(a_{2}\).

    The ranking obtained by the method based on the expected probability degree is visibly different from that obtained by the method based on regret theory proposed in this paper. The difference lies in the position of \(a_{3}\). When the method based on the expected probability degree is used, \(a_{3}\) is inferior to \(a_{2}\) and \(a_{1}\), but with the use of the proposed method, the situation is reversed. Compared with Wang et al.’s method [47], which is based on complete rationality using the expected probability degree, the proposed method is based on bounded rationality using regret theory. The proposed method fully considers decision-makers’ bounded rationality for decision-making, making it superior to traditional decision-making methods.

Compared with the existing methods, the proposed grey stochastic MCDM method based on regret theory and TOPSIS has the following advantages.

  1. 1.

    The EGNs, combining both intervals and discrete sets of numbers, can express the evaluation information with more flexibility. They have a more powerful capability to express uncertainty due to missing information, and thus have a wider range of applications.

  2. 2.

    The method proposed in this paper is based on bounded rationality, which fully considers the decision-makers’ psychological behavior. Compared with traditional decision-making methods based on complete rationality, the proposed method can address decision-making problems more reasonably and effectively.

  3. 3.

    The proposed method is based on regret theory, which takes the decision-maker’s attitude of regret into account. Compared with prospect theory, regret theory does not use reference points, and the calculation formula includes fewer parameters. Therefore, the proposed method is more advantageous than prospect theory in practical application.

  4. 4.

    The proposed method based on regret theory employs a new form of criteria values, EGNs, thus expanding the application of regret theory.

5 Conclusions

Stochastic MCDM methods are widely used in real-life decision-making processes, and the decision-making method based on regret theory is more congruent with people’s actual decision-making behavior. In this paper, we proposed a grey stochastic MCDM approach based on regret theory using EGNs to express criteria values. Regret theory was first used to obtain the utility value and regret value concerning the criteria, and then the perceived utility values of the alternatives were calculated. Next, the TOPSIS method was used to prioritize the alternatives according to the overall perceived utility intervals. Finally, two examples and comparison analyses were presented to demonstrate the feasibility and usability of the proposed method.

The proposed method makes several important contributions. First, dealing with stochastic MCDM problems with EGNs expands the research scope of MCDM. The proposed method also incorporates the decision-makers’ regrets into the decision-making model, which makes the method more applicable to real-world choices. Additionally, the proposed method expands the application of regret theory and provides a new method through which to solve stochastic MCDM problems.

It is worth noting that there are a considerable number of MCDM problems with criteria values that are expressed as grey and show randomness. Grey stochastic MCDM problems have a wide range of applications, such as investment, new product development project selection, medical care evaluation, supplier selection, and project risk assessment. Therefore, the study of grey stochastic MCDM problems has important practical applications.

In the future, we expect to investigate the generalized distance measures and aggregation operators of EGNs. Furthermore, we may apply our distance measures and aggregation operators to other grey stochastic MCDM methods based on bounded rationality. Meanwhile, we will also consider expanding the scope of application of our method.