1 Introduction

Probability theory has been extensively applied for handling indeterminate phenomena. As important numerical characteristics, the expected value, variance, and covariance of random variables have been studied thoroughly involving their mathematical properties and applications. While dealing with practical decision problems, those indices are often used as important criteria. Additionally, in order to solve problems appropriately via probability theory, it is often required to estimate probability distributions by using statistical data in accordance with the law of large numbers.

However, in the real world, due to technological and economical difficulties or low frequency of events occurring, there exists a shortage of sufficient data for deriving precise probability distributions. In such cases, especially when there are no samples available to evaluate probability distributions, as an alternative approach, some field experts are asked to estimate the degrees of belief that events happen. It is widely known that due to subjectivity human beings may overweight unlikely events, which makes the estimated degrees of belief be very different from the real frequency. For the purpose of tackling this kind of problems, Liu (2007) initiated uncertainty theory. As an efficient tool for dealing with indeterminate phenomena, uncertainty theory has been studied by many researchers. It has been applied to many fields, such as uncertain programming (Ke et al. 2015; Zhong et al. 2017; Zhou et al. 2014), uncertain process (Yao and Li 2012; Yao and Zhou 2016, 2017), uncertain network (Zhang et al. 2013; Zhou et al. 2014a, b), uncertain logic (Li and Liu 2009; Zhang and Li 2014), uncertain finance (Chen et al. 2017; Ji and Zhou 2015b; Zhou et al. 2017), uncertain differential equation (Ji and Zhou 2015a; Su et al. 2016), uncertain agency problem (Wu et al. 2014; Yang et al. 2014), among others.

In theoretical research on uncertainty theory, much attention was given to the numerical characteristics of uncertain variables especially expected value and variance due to their useful practical interpretation. The notion of expected value was first defined by Liu (2007) as the mean value of all possible values of an uncertain variable. Concerning uncertain variables whose distributions are regular, Liu (2010) further proposed a convenient equivalent formula for expected value in terms of inverse distribution. As an extension, Liu and Ha (2010) suggested a formula to calculate the expected value of a strictly monotone function in regard to independent uncertain variables whose distributions are regular. Based upon the concept of expected value, Liu (2007) introduced the variance of an uncertain variable. Due to the subadditivity of an uncertain measure, a formula for calculating the variance was shown in Liu (2007) by virtue of its distribution. Yao (2015) then derived an equivalent formula for calculating the variance of an uncertain variable using its inverse distribution. In the same paper, Yao (2015) proved some inequalities for variances of uncertain variables useful in real applications. The expected value and variance of uncertain variables have been widely applied into practical problems. For example, Liu et al. (2014) proposed a new uncertain expected value operator approach for determining the importance of engineering characteristics and their rankings in quality function deployment. Zhou et al. (2015, 2016) extended the concept of minimum spanning tree to its uncertain version by using the expected value as one of judgement criteria. Qin (2015) presented the calculation formulae for variances of hybrid portfolio returns on the basis of uncertainty theory and then formulated corresponding mean-variance models to solve the hybrid portfolio selection problem.

In probability theory, covariance and correlation coefficient are very important measures to interpret the degree of association between two random variables, and have been useful especially in the field of regression analysis. In this paper, the concepts of covariance and correlation coefficient are initiated in the field of uncertainty theory involving their mathematical properties. The relationships between covariance and variance are also investigated.

The rest of this paper is organized as follows. Some fundamental concepts of uncertain variables are recalled in Sect. 2. The concepts and some calculation formulae of covariance and correlation coefficient for uncertain variables are presented, and some of their properties are put forward in Sect. 3. Subsequently, the relationships between variance and covariance of uncertain variables are discussed and described by means of some equalities and inequalities in Sect. 4. Some conclusions are drawn in Sect. 5.

2 Preliminaries

In this section, some basic definitions and theorems of uncertainty theory are briefly recalled, as they will be used throughout this paper.

Definition 1

(Liu 2007) Let \(\Gamma \) be a nonempty set, and a \(\sigma \)-algebra over \(\Gamma \). The set function is called an uncertain measure if it satisfies the following three axioms:

Axiom 1

(Normality Axiom) for the universal set \(\Gamma \);

Axiom 2

(Duality Axiom) for any event \(\Lambda \);

Axiom 3

(Subadditivity Axiom) For every countable sequence of events \(\Lambda _1,\) \(\Lambda _2,\ldots \), we have

(1)

Liu (2009) defined the product uncertain measure as follows:

Axiom 4

(Product Axiom) Let be uncertainty spaces for \(k = 1, 2,\ldots \). The product uncertain measure is an uncertain measure satisfying

(2)

where \(\Lambda _k\) are arbitrarily chosen events from \(\text{ L }_k\) for \(k = 1, 2,\ldots \), respectively.

Definition 2

(Liu 2007) An uncertain variable is a measurable function \(\mu \) from an uncertainty space to the set of real numbers, i.e., for any Borel set B of real numbers, the set

$$\begin{aligned} \{\mu \in B\}=\{\gamma \in \Gamma \ \big |\ \mu (\gamma )\in B\} \end{aligned}$$
(3)

is an event.

Definition 3

(Liu 2007) The uncertainty distribution \(\Psi \) of an uncertain variable \(\mu \) is defined by

(4)

for any real number x.

Example 1

An uncertain variable \(\mu \) is called linear, denoted as \(\mu \sim \mathcal{L}(a, b)\) with \(a<b\), if its uncertainty distribution is

$$\begin{aligned} \Psi (x)=\left\{ \begin{array}{ll} 0, &{}\quad \text{ if } x< a\\ (x-a)/(b-a), &{}\quad \text{ if } a\le x < b \\ 1, &{}\quad \text{ if } x\ge b. \end{array} \right. \end{aligned}$$
(5)

Example 2

An uncertain variable \(\mu \) is called zigzag, denoted as \(\mu \sim {\mathcal {Z}}(a,b,c)\) with \(a<b<c\), if its uncertainty distribution is

$$\begin{aligned} \Psi (x)=\left\{ \begin{array}{ll} 0, &{}\quad \text{ if } x< a\\ (x-a)/2(b-a), &{}\quad \text{ if } a \le x< b \\ (x+c-2b)/2(c-b), &{}\quad \text{ if } b \le x < c \\ 1, &{}\quad \text{ if } x\ge c. \end{array} \right. \end{aligned}$$
(6)

Example 3

An uncertain variable \(\mu \) is called normal, denoted as with \(\sigma >0\), if its uncertainty distribution is

$$\begin{aligned} \Psi (x)=\displaystyle \left( 1+\exp \displaystyle \left( \frac{\pi (e-x)}{\sqrt{3}\sigma }\displaystyle \right) \displaystyle \right) ^{-1},\quad x\in \mathfrak {R}. \end{aligned}$$
(7)

Definition 4

(Liu 2010) If an uncertainty distribution \(\Psi (x)\) is continuous and strictly increasing at \(0<\Psi (x)<1\), and \(\lim \limits _{{{x\rightarrow -\infty }}} \Psi (x)=0, \lim \limits _{{{x\rightarrow +\infty }}}\Psi (x)=1\), then \(\Psi (x)\) is called regular.

Definition 5

(Liu 2010) If \(\mu \) is an uncertain variable whose distribution \(\Psi (x)\) is regular, then \(\Psi ^{-1}(\theta )\) is called the inverse uncertainty distribution of \(\mu \).

From Definition 4, we know that \(\Psi ^{-1}(\theta )\) is well defined on (0, 1). If necessary, the domain can be extended by letting \(\Psi ^{-1}(0)=\lim _{\theta \downarrow 0}\Psi ^{-1}(\theta )\) and \(\Psi ^{-1}(1)=\lim _{\theta \uparrow 1}\Psi ^{-1}(\theta )\).

It is easy to see that the distributions of a linear uncertain variable \(\mu _1\sim \mathcal{L}(a, b)\) in Example 1, a zigzag uncertain variable \(\mu _2\sim {\mathcal {Z}}(a,b,c)\) in Example 2, and a normal uncertain variable in Example 3 are all regular, and their inverse uncertainty distributions are

$$\begin{aligned} \Psi ^{-1}_1(\theta )= & {} a + (b-a)\theta , \end{aligned}$$
(8)
$$\begin{aligned} \Psi ^{-1}_2(\theta )= & {} \left\{ \begin{array}{ll} a+2(b-a)\theta , &{}\quad \text{ if } \theta \le 0.5 \\ 2b-c+2(c-b)\theta , &{}\quad \text{ if } \theta > 0.5, \end{array} \right. \end{aligned}$$
(9)

and

$$\begin{aligned} \Psi ^{-1}_3(\theta )=e+\frac{\sqrt{3}\sigma }{\pi }\ln \frac{\theta }{1-\theta }, \end{aligned}$$
(10)

respectively.

Definition 6

(Liu 2009) The uncertain variables \(\mu _1\), \(\mu _2\), \(\ldots \), \(\mu _n\) are said to be independent if

(11)

for any Borel sets \(B_1,B_2,\ldots ,B_n\) of real numbers.

Theorem 1

(Liu 2009) The uncertain variables \(\mu _1\), \(\mu _2\), \(\ldots \), \(\mu _n\) are independent if and only if

(12)

for any Borel sets \(B_1,B_2,\ldots ,B_n\) of real numbers.

Theorem 2

(Liu 2010) Let \(\mu _1\), \(\mu _2\),\(\ldots \), \(\mu _n\) be independent uncertain variables that have regular distributions \(\Psi _1\), \(\Psi _2\),\(\ldots \), \(\Psi _n\), respectively. If the function \(f(y_1,y_2,\ldots ,y_n)\) is strictly increasing in \(y_1, y_2,\ldots , y_m\) and strictly decreasing in \(y_{m+1}, y_{m+2},\ldots , y_n\), then the uncertain variable \(\mu =f(\mu _1,\mu _2,\ldots ,\mu _n)\) has an inverse distribution

$$\begin{aligned} \Upsilon ^{-1}(\theta )=f\left( \Psi ^{-1}_1(\theta ),\ldots , \Psi ^{-1}_m(\theta ), \Psi ^{-1}_{m+1}(1-\theta ),\ldots , \Psi ^{-1}_{n}(1-\theta )\right) . \end{aligned}$$
(13)

Theorem 3

(Liu 2010) If \(\mu \) is an uncertain variable that has regular distribution \(\Psi \), then

$$\begin{aligned} E[\mu ]=\displaystyle \int _{0}^{1}\Psi ^{-1}(\theta )\mathrm{d}\theta . \end{aligned}$$
(14)

According to Eqs. (8)–(10) and (14), the expected value of uncertain variables \(\mu _1\sim \mathcal{L}(a, b)\), \(\mu _2\sim {\mathcal {Z}}(a,b,c)\), and are

$$\begin{aligned} E[\mu _1]= & {} \displaystyle \int _{0}^{1}\left( a+(b-a)\theta \right) \mathrm{d}\theta =\displaystyle \frac{a+b}{2}, \end{aligned}$$
(15)
$$\begin{aligned} E[\mu _2]= & {} \displaystyle \int _{0}^{0.5}\left( a+2(b-a)\theta \right) \mathrm{d}\theta \nonumber \\&+\displaystyle \int _{0.5}^{1}\left( 2b-c+2(c-b)\theta \right) \mathrm{d}\theta =\displaystyle \frac{a+2b+c}{4}, \end{aligned}$$
(16)

and

$$\begin{aligned} E[\mu _3]=\displaystyle \int _{0}^{1}\left( e+\frac{\sqrt{3}\sigma }{\pi }\ln \frac{\theta }{1-\theta }\right) \mathrm{d}\theta =e, \end{aligned}$$
(17)

respectively.

Theorem 4

(Liu 2010) If \(\mu \) and \(\nu \) are two independent uncertain variables that have finite expected values, then we have

$$\begin{aligned} E[a\mu +b\nu ]=aE[\mu ]+bE[\nu ] \end{aligned}$$
(18)

for any real numbers a and b.

Definition 7

(Liu 2007) If \(\mu \) is an uncertain variable that has a finite expected value \(E[\mu ]\), then the variance of \(\mu \) is defined by

$$\begin{aligned} V[\mu ]=E[\left( \mu -E[\mu ]\right) ^{2}]. \end{aligned}$$
(19)

Theorem 5

(Yao 2015) If \(\mu \) is an uncertain variable that has a regular distribution \(\Psi \) and a finite expected value \(E[\mu ]\), then the variance of \(\mu \) is

$$\begin{aligned} V[\mu ]=\displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )-E[\mu ]\right) ^{2}\mathrm{d}\theta . \end{aligned}$$
(20)

Theorem 6

(Liu 2007) If \(\mu \) is an uncertain variable that has a finite expected value, then we have

$$\begin{aligned} V[a\mu +b]=a^2V[\mu ] \end{aligned}$$
(21)

for any real numbers a and b.

3 Covariance and correlation coefficient

In this section, we first give a definition for the covariance of two uncertain variables. A method is also suggested for determining the covariance of two uncertain variables with regular distributions via their inverse distributions. Afterwards, the definition of the correlation coefficient of two uncertain variables is presented. The properties of covariance and correlation coefficient are further studied and the meaning behind the mathematical formulation is then revealed through some examples. For simplicity, an uncertain variable with a regular distribution is called a regular uncertain variable in the rest of this paper.

3.1 Definition and calculation formulae of covariance

Definition 8

Let \(\mu \) and \(\nu \) be two uncertain variables. The covariance of \(\mu \) and \(\nu \) is defined by

$$\begin{aligned} Cov[\mu ,\nu ] =E[\left( \mu -E[\mu ]\right) \left( \nu -E[\nu ]\right) ], \end{aligned}$$
(22)

where \(E[\mu ]\) and \(E[\nu ]\) are the expected values of \(\mu \) and \(\nu \), respectively.

Remark 1

It is known that for two random variables \(\mu \) and \(\nu \), we have

$$\begin{aligned} Cov[\mu ,\nu ]= & {} E\left[ \left( \mu -E[\mu ]\right) \left( \nu -E[\nu ]\right) \right] \\= & {} E\left[ \mu \nu -E[\nu ]\mu -E[\mu ]\nu +E[\mu ]E[\nu ]\right] \\= & {} E[\mu \nu ]-E[\mu ]E[\nu ]. \end{aligned}$$

Since \(E[\mu \nu ]=E[\mu ]E[\nu ]\) if the two random variables are independent, as a consequence, \(Cov[\mu ,\nu ]=0\) holds for any two independent random variables.

On the other hand, if the two variables \(\mu \) and \(\nu \) are uncertain variables, from Definition 8, we have

$$\begin{aligned} Cov[\mu ,\nu ]= & {} E\left[ \left( \mu -E[\mu ]\right) \left( \nu -E[\nu ]\right) \right] \\= & {} E\left[ \mu \nu -E[\nu ]\mu -E[\mu ]\nu +E[\mu ]E[\nu ]\right] \\= & {} E\big [\mu \nu -E[\nu ]\mu -E[\mu ]\nu \big ]+E[\mu ]E[\nu ]. \end{aligned}$$

Even though \(\mu \) and \(\nu \) are independent, the uncertain variables \(\mu \nu \), \(E[\nu ]\mu \), and \(E[\mu ]\nu \) are not independent in general. Since the linearity of expected value of uncertain variables is based on independence (see Theorem 4), it cannot be deduced that \(E\big [\mu \nu - E[\nu ]\mu - E[\mu ]\nu \big ]=E[\mu \nu ]-2E[\mu ]E[\nu ]\). Furthermore, the equation \(E[\mu \nu ]=E[\mu ]E[\nu ]\) does not hold for two independent uncertain variables \(\mu \) and \(\nu \). Therefore, the conclusion \(Cov[\mu ,\nu ]=0\) does not follow if two uncertain variables \(\mu \) and \(\nu \) are independent. The main reason is the difference between the concepts of independence for the two types of variables. Due to this, the covariance of uncertain variables has a completely different interpretation compared with the covariance of random variables, and this will be explained thoroughly later in the paper.

As mentioned above, the uncertain measure is a subadditive measure. Just like the variance, it is not easy to express the covariance of uncertain variables defined in Eq. (22) simply by distributions. From Definitions 7 and 8, it is clear that variance can be considered as a special type of covariance. In view of formula (20) for the variance of uncertain variable (see Theorem 5), we provide the following stipulation for the calculation of covariance via inverse distributions.

Stipulation 1

Let \(\mu \) and \(\nu \) be two regular uncertain variables with distributions \(\Psi \) and \(\Upsilon \) and finite expected values \(E[\mu ]\) and \(E[\nu ]\), respectively. Then the covariance of \(\mu \) and \(\nu \) is

$$\begin{aligned} Cov[\mu ,\nu ]=\displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )-E[\mu ])(\Upsilon ^{-1}(\theta )-E[\nu ]\right) \mathrm{d}\theta . \end{aligned}$$
(23)

Theorem 7

Let \(\mu \) and \(\nu \) be two regular uncertain variables with distributions \(\Psi \) and \(\Upsilon \) and finite expected values \(E[\mu ]\) and \(E[\nu ]\), respectively. Then

$$\begin{aligned} Cov[\mu ,\nu ]=\displaystyle \int _{0}^{1}\Psi ^{-1}(\theta )\Upsilon ^{-1}(\theta )\mathrm{d}\theta -E[\mu ]E[\nu ]. \end{aligned}$$
(24)

Proof

From Stipulation 1, it is easy to obtain the covariance of \(\mu \) and \(\nu \) as

$$\begin{aligned} Cov[\mu ,\nu ]= & {} \displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )\Upsilon ^{-1}(\theta )-E[\nu ]\Psi ^{-1}(\theta )-E[\mu ]\Upsilon ^{-1}(\theta )+E[\mu ]E[\nu ] \right) \mathrm{d}\theta \\= & {} \displaystyle \int _{0}^{1}\Psi ^{-1}(\theta )\Upsilon ^{-1}(\theta )\mathrm{d}\theta -E[\nu ]\displaystyle \int _{0}^{1}\Psi ^{-1}(\theta )\mathrm{d}\theta \\&-\,E[\mu ]\displaystyle \int _{0}^{1}\Upsilon ^{-1}(\theta )\mathrm{d}\theta +E[\mu ]E[\nu ]. \end{aligned}$$

Then according to (14) (see Theorem 3), we get

$$\begin{aligned} Cov[\mu ,\nu ]=\displaystyle \int _{0}^{1}\Psi ^{-1}(\theta )\Upsilon ^{-1}(\theta )\mathrm{d}\theta -E[\mu ]E[\nu ]. \end{aligned}$$

\(\square \)

Theorem 8

Let \(\mu \) be a regular uncertain variable with a finite expected value. Then

$$\begin{aligned} V[\mu ]=Cov[\mu ,\mu ]. \end{aligned}$$
(25)

Proof

For convenience, denote the distribution of \(\mu \) by \(\Psi \). From Stipulation 1, we obtain

$$\begin{aligned} Cov[\mu ,\mu ]= & {} \displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )-E[\mu ]\right) \left( \Psi ^{-1}(\theta )-E[\mu ]\right) \mathrm{d}\theta \\= & {} \displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )-E[\mu ]\right) ^2\mathrm{d}\theta . \end{aligned}$$

Then from (20) (see Theorem 5), it follows that \(Cov[\mu ,\mu ]=V[\mu ]\). \(\square \)

Example 4

Consider the covariance of two linear uncertain variables \(\mu \sim \mathcal{L}(a, b)\) and \(\nu \sim \mathcal{L}(c, d)\). Since the expected values of \(\mu \) and \(\nu \) are \(E[\mu ]=(a+b)/2\) and \(E[\nu ]=(c+d)/2\), and the inverse distributions of \(\mu \) and \(\nu \) are

$$\begin{aligned} \Psi ^{-1}(\theta )=a+(b-a)\theta \end{aligned}$$

and

$$\begin{aligned} \Upsilon ^{-1}(\theta )=c+(d-c)\theta , \end{aligned}$$

respectively, it follows from Stipulation 1 that

$$\begin{aligned} Cov[\mu ,\nu ]= & {} \displaystyle \int _0^1\left( a+(b-a)\theta -\displaystyle \frac{a+b}{2}\right) \left( c+(d-c)\theta -\displaystyle \frac{c+d}{2}\right) \mathrm{d}\theta \\ = & {} \displaystyle \frac{(b-a)(d-c)}{12}. \end{aligned}$$

Example 5

Consider the covariance of two zigzag uncertain variables \(\mu \sim {\mathcal {Z}}(a_1,b_1,c_1)\) and \(\nu \sim {\mathcal {Z}}(a_2,b_2,c_2)\). Since the expected values of \(\mu \) and \(\nu \) are \(E[\mu ]=(a_1+2b_1+c_1)/4\) and \(E[\nu ]=(a_2+2b_2+c_2)/4\), and the inverse distributions of \(\mu \) and \(\nu \) are

$$\begin{aligned} \Psi ^{-1}(\theta )=\left\{ \begin{array}{ll} a_1+2(b_1-a_1)\theta , &{}\quad \text{ if } \theta \le 0.5 \\ 2b_1-c_1+2(c_1-b_1)\theta , &{}\quad \text{ if } \theta > 0.5 \end{array} \right. \end{aligned}$$

and

$$\begin{aligned} \Upsilon ^{-1}(\theta )=\left\{ \begin{array}{ll} a_2+2(b_2-a_2)\theta , &{}\quad \text{ if } \theta \le 0.5 \\ 2b_2-c_2+2(c_2-b_2)\theta , &{}\quad \text{ if } \theta > 0.5, \end{array} \right. \end{aligned}$$

respectively, from Stipulation 1, we obtain

$$\begin{aligned}&Cov[\mu ,\nu ]\\&\quad =\displaystyle \int _0^{0.5}\left( a_1+2(b_1-a_1) \theta -\frac{a_1+2b_1+c_1}{4}\right) \\&\qquad \times \left( a_2+2(b_2-a_2)\theta -\frac{a_2+2b_2+c_2}{4}\right) \mathrm{d}\theta \\&\quad \quad +\displaystyle \int _{0.5}^1\left( 2b_1-c_1+2(c_1-b_1) \theta -\frac{a_1+2b_1+c_1}{4}\right) \\&\qquad \times \left( 2b_2-c_2+2(c_2-b_2)\theta -\displaystyle \frac{a_2+2b_2+c_2}{4}\right) \mathrm{d}\theta \\&\quad =\displaystyle \frac{1}{48}\left[ (b_2\!-\!a_2) \left( 5(b_1\!-\!a_1)\!+\!3(c_1\!-\!b_1)\right) +(c_2\!-\!b_2) \left( 3(b_1\!-\!a_1)+5(c_1-b_1)\right) \right] . \end{aligned}$$

Example 6

Consider the covariance of two normal uncertain variables and . Since the expected values of \(\mu \) and \(\nu \) are \(E[\mu ]=e_1\) and \(E[\nu ]=e_2\), and the inverse distributions of \(\mu \) and \(\nu \) are

$$\begin{aligned} \Psi ^{-1}(\theta )=e_1+\frac{\sqrt{3}\sigma _1}{\pi }\ln \frac{\theta }{1-\theta } \end{aligned}$$

and

$$\begin{aligned} \Upsilon ^{-1}(\theta )=e_2+\frac{\sqrt{3}\sigma _2}{\pi }\ln \frac{\theta }{1-\theta }, \end{aligned}$$

respectively, following from Stipulation 1, we get

$$\begin{aligned} Cov[\mu ,\nu ]= & {} \displaystyle \int _0^1\left( e_1+\frac{\sqrt{3}\sigma _2}{\pi }\ln \frac{\theta }{1-\theta }-e_1\right) \left( e_2+\frac{\sqrt{3}\sigma _2}{\pi }\ln \frac{\theta }{1-\theta }-e_2\right) \mathrm{d}\theta \\= & {} \displaystyle \frac{3\sigma _1\sigma _2}{\pi ^2}\int _0^1\left( \ln \frac{\theta }{1-\theta }\right) ^2\mathrm{d}\theta \\= & {} \sigma _1\sigma _2. \end{aligned}$$

Example 7

Consider the covariance of two uncertain variables \(\mu \sim \mathcal{L}(a, b)\) and . Using the results of Examples 4 and 6, and Stipulation 1, we get

$$\begin{aligned} Cov[\mu ,\nu ]= & {} \displaystyle \int _0^1\left( a+(b-a)\theta -\displaystyle \frac{a+b}{2}\right) \left( e+\frac{\sqrt{3}\sigma }{\pi }\ln \frac{\theta }{1-\theta }-e\right) \mathrm{d}\theta \\= & {} \displaystyle \frac{\sqrt{3}(b-a)\sigma }{\pi }\displaystyle \int _0^1\left( \theta -\displaystyle \frac{1}{2}\right) \ln \displaystyle \frac{\theta }{1-\theta }\mathrm{d}\theta \\= & {} \displaystyle \frac{\sqrt{3}(b-a)\sigma }{2\pi }. \end{aligned}$$

Theorem 9

Assume that \(\mu _1, \mu _2,\ldots , \mu _n\) and \(\nu _1, \nu _2,\ldots , \nu _n\) are independent regular uncertain variables with distributions \(\Psi _1, \Psi _2,\ldots , \Psi _n\), and \(\Upsilon _1, \Upsilon _2,\ldots , \Upsilon _n\), respectively. If \(f(y_1, y_2,\ldots , y_n)\) is strictly increasing in \(y_1, y_2,\ldots , y_m\) and strictly decreasing in \(y_{m+1}, y_{m+2},\ldots , y_n\), and if \(g(y_1, y_2,\ldots , y_n)\) is strictly increasing in \(y_1, y_2,\ldots , y_k\) and strictly decreasing in \(y_{k+1}, y_{k+2},\ldots , y_n\), then the uncertain variables \(\mu =f(\mu _1, \mu _2,\ldots ,\mu _n)\) and \(\nu =g(\nu _1, \nu _2,\ldots , \nu _n)\) have a covariance

$$\begin{aligned} Cov[\mu ,\nu ]=\displaystyle \int _0^1\left( f\left( \Psi _1^{-1}(\theta ),\ldots , \Psi _m^{-1}(\theta ), \Psi _{m+1}^{-1}(1-\theta ),\ldots , \Psi _n^{-1}(1-\theta )\right) -\tau _1\right) \nonumber \\ \times \left( g\left( \Upsilon _1^{-1}(\theta ),\ldots , \Upsilon _k^{-1}(\theta ), \Upsilon _{k+1}^{-1}(1-\theta ),\ldots , \Upsilon _n^{-1}(1-\theta )\right) -\tau _2\right) \mathrm{d}\theta ,\nonumber \\ \end{aligned}$$
(26)

where \(\tau _1\) and \(\tau _2\) are the expected values of \(f(\mu _1, \mu _2,\ldots , \mu _n)\) and \(g(\nu _1, \nu _2,\ldots , \nu _n)\), respectively, with

$$\begin{aligned} \tau _1=\displaystyle \int _0^1f\left( \Psi _1^{-1}(\theta ),\ldots , \Psi _m^{-1}(\theta ), \Psi _{m+1}^{-1}(1-\theta ),\ldots , \Psi _n^{-1}(1-\theta )\right) \mathrm{d}\theta \end{aligned}$$

and

$$\begin{aligned} \tau _2=\displaystyle \int _0^1g\left( \Upsilon _1^{-1}(\theta ),\ldots , \Upsilon _k^{-1}(\theta ), \Upsilon _{k+1}^{-1}(1-\theta ),\ldots , \Upsilon _n^{-1}(1-\theta )\right) \mathrm{d}\theta . \end{aligned}$$

Proof

By the operational law (see Theorem 2), the inverse distributions of \(\mu \) and \(\nu \) are

$$\begin{aligned} \Psi ^{-1}(\theta )=f\left( \Psi _1^{-1}(\theta ),\ldots , \Psi _m^{-1}(\theta ), \Psi _{m+1}^{-1}(1-\theta ),\ldots , \Psi _n^{-1}(1-\theta )\right) , \end{aligned}$$

and

$$\begin{aligned} \Upsilon ^{-1}(\theta )=g\left( \Upsilon _1^{-1}(\theta ),\ldots , \Upsilon _k^{-1}(\theta ), \Upsilon _{k+1}^{-1}(1-\theta ),\ldots , \Upsilon _n^{-1}(1-\theta )\right) , \end{aligned}$$

respectively. Then according to Stipulation 1 and formula (14) (see Theorem 3), the theorem is easily proved. \(\square \)

3.2 Correlation coefficient

The normalized version of the covariance of uncertain variables, called the correlation coefficient, is a dimensionless quantity, defined by dividing the covariance by the product of the square roots of the variances of \(\mu \) and \(\nu \).

Definition 9

Let \(\mu \) and \(\nu \) be two regular uncertain variables with finite expected values and non-zero variances. The correlation coefficient of \(\mu \) and \(\nu \) is defined by

$$\begin{aligned} Corr[\mu ,\nu ]=\displaystyle \frac{Cov[\mu ,\nu ]}{\sqrt{V[\mu ]}\sqrt{V[\nu ]}}. \end{aligned}$$
(27)

Theorem 10

Let \(\mu \) and \(\nu \) be two regular uncertain variables with finite expected values. Then

$$\begin{aligned} |Corr[\mu ,\nu ]|\le 1. \end{aligned}$$
(28)

Proof

First denote the distributions of \(\mu \) and \(\nu \) as \(\Psi \) and \(\Upsilon \), respectively. Then by Stipulation 1 and Definition 9, we only need to prove the inequality

$$\begin{aligned} \left| \displaystyle \int _{0}^{1}\frac{(\Psi ^{-1}(\theta )-E[\mu ])(\Upsilon ^{-1}(\theta )-E[\nu ])}{\sqrt{V[\mu ]}\sqrt{V[\nu ]}}\mathrm{d}\theta \right| \le 1. \end{aligned}$$
(29)

It is known that \(|\int _{a}^{b}f(x)\mathrm{d}x|\le \int _{a}^{b}|f(x)|\mathrm{d}x\) and \(|ab|\le (a^2+b^2)/2\), so we have

$$\begin{aligned}&\left| \displaystyle \int _{0}^{1}\frac{(\Psi ^{-1}(\theta )-E[\mu ])(\Upsilon ^{-1}(\theta )-E[\nu ])}{\sqrt{V[\mu ]}\sqrt{V[\nu ]}}\mathrm{d}\theta \right| \nonumber \\&\quad \le \displaystyle \int _{0}^{1}\left| \frac{(\Psi ^{-1}(\theta )-E[\mu ])(\Upsilon ^{-1}(\theta )-E[\nu ])}{\sqrt{V[\mu ]}\sqrt{V[\nu ]}}\right| \mathrm{d}\theta \end{aligned}$$
(30)

and

$$\begin{aligned}&\left| \displaystyle \frac{\left( \Psi ^{-1}(\theta )-E[\mu ]\right) \left( \Upsilon ^{-1}(\theta )-E[\nu ]\right) }{\sqrt{V[\mu ]}\sqrt{V[\nu ]}}\right| \nonumber \\&\quad \le \displaystyle \frac{1}{2}\displaystyle \frac{\left( \Psi ^{-1}(\theta )-E[\mu ]\right) ^2}{V[\mu ]}+\displaystyle \frac{1}{2}\displaystyle \frac{\left( \Upsilon ^{-1}(\theta )-E[\nu ]\right) ^2}{V[\nu ]}. \end{aligned}$$
(31)

It follows from Inequalities (30) and (31) that

$$\begin{aligned}&\left| \displaystyle \int _{0}^{1}\frac{\left( \Psi ^{-1}(\theta )-E[\mu ]\right) \left( \Upsilon ^{-1}(\theta )-E[\nu ]\right) }{\sqrt{V[\mu ]}\sqrt{V[\nu ]}}\mathrm{d}\theta \right| \\&\quad \le \displaystyle \int _{0}^{1}\left| \frac{(\Psi ^{-1}(\theta )-E[\mu ])(\Upsilon ^{-1}(\theta )-E[\nu ])}{\sqrt{V[\mu ]}\sqrt{V[\nu ]}}\right| \mathrm{d}\theta \\&\quad \le \displaystyle \frac{1}{2}\displaystyle \int _{0}^{1}\frac{\left( \Psi ^{-1}(\theta )-E[\mu ]\right) ^2}{V[\mu ]}\mathrm{d}\theta + \displaystyle \frac{1}{2}\displaystyle \int _{0}^{1}\frac{\left( \Upsilon ^{-1}(\theta )-E[\nu ]\right) ^2}{V[\nu ]}\mathrm{d}\theta . \end{aligned}$$

Using formula (20) (see Theorem 5), it follows that

$$\begin{aligned} \displaystyle \frac{1}{2}\displaystyle \int _{0}^{1}\frac{\left( \Psi ^{-1}(\theta )-E[\mu ]\right) ^2}{V[\mu ]}\mathrm{d}\theta + \displaystyle \frac{1}{2}\displaystyle \int _{0}^{1}\frac{\left( \Upsilon ^{-1}(\theta )-E[\nu ]\right) ^2}{V[\nu ]}\mathrm{d}\theta =\displaystyle \frac{1}{2}+\displaystyle \frac{1}{2}=1. \end{aligned}$$

\(\square \)

Remark 2

Notice that the equality in Inequality (28) holds if and only if the equalities in Inequalities (30) and (31) hold concurrently, which means

$$\begin{aligned} \displaystyle \frac{\Psi ^{-1}(\theta )-E[\mu ]}{\sqrt{V[\mu ]}}=\displaystyle \frac{\Upsilon ^{-1}(\theta )-E[\nu ]}{\sqrt{V[\nu ]}}, \quad 0\le \theta \le 1. \end{aligned}$$

Example 8

Consider the correlation coefficient of two linear uncertain variables \(\mu \sim \mathcal{L}(a, b)\) and \(\nu \sim \mathcal{L}(c, d)\). According to the calculation formula for variance and the result of Example 4, we have

$$\begin{aligned} V[\mu ]=\displaystyle \frac{(b-a)^2}{12},\quad V[\nu ]=\displaystyle \frac{(d-c)^2}{12},\quad Cov[\mu ,\nu ]=\displaystyle \frac{(b-a)(d-c)}{12}. \end{aligned}$$

Then the correlation coefficient of \(\mu \) and \(\nu \) is

$$\begin{aligned} Corr[\mu ,\nu ] = \displaystyle \frac{\displaystyle \frac{(b-a)(d-c)}{12}}{\displaystyle \frac{b-a}{2\sqrt{3}}\times \displaystyle \frac{d-c}{2\sqrt{3}}}=1. \end{aligned}$$

In summary, the correlation coefficient of any two linear uncertain variables is equal to 1. Figure 1 shows the distributions of four linear uncertain variables \(\mathcal{L}(0,3),\mathcal{L}(1,3),\mathcal{L}(3,4)\) and \(\mathcal{L}(2,6)\). It is clear that the correlation coefficients of any two of them are all equal to 1.

Fig. 1
figure 1

The distributions of four linear uncertain variables

Example 9

Consider the correlation coefficient of two zigzag uncertain variables \(\mu \sim {\mathcal {Z}}(a_1, b_1\), \(c_1)\) and \(\nu \sim {\mathcal {Z}}(a_2, b_2, c_2)\). According to the calculation formula for variance and the result of Example 5, we have

$$\begin{aligned} V[\mu ]= & {} \displaystyle \frac{1}{48}\left[ 5(b_1-a_1)^2+6(b_1-a_1)(c_1-b_1)+5(c_1-b_1)^2\right] ,\\ V[\nu ]= & {} \displaystyle \frac{1}{48}\left[ 5(b_2-a_2)^2+6(b_2-a_2)(c_2-b_2)+5(c_2-b_2)^2\right] , \end{aligned}$$

and

$$\begin{aligned} Cov[\mu ,\nu ]= & {} \displaystyle \frac{1}{48}[(b_2-a_2) (5(b_1-a_1)\\&+\,3(c_1-b_1))+(c_2-b_2)(3(b_1-a_1)+5(c_1-b_1))]. \end{aligned}$$

Then the correlation coefficient of \(\mu \) and \(\nu \) is

$$\begin{aligned}&Corr[\mu ,\nu ]\\&\quad =\displaystyle \frac{(b_2-a_2)[5(b_1-a_1)+3(c_1-b_1)]+(c_2-b_2)[3(b_1-a_1)+5(c_1-b_1)]}{\sqrt{\prod \nolimits _{i=1}^{2}[5(b_i-a_i)^2+6(b_i-a_i)(c_i-b_i)+5(c_i-b_i)^2]}}. \end{aligned}$$

Denoting \(\displaystyle \frac{c_1-b_1}{b_1-a_1}=m\), \(\displaystyle \frac{c_2-b_2}{b_2-a_2}=n\), and \(\displaystyle \frac{m}{n}=k\), we get

$$\begin{aligned} Corr[\mu ,\nu ]= & {} \displaystyle \frac{5+3m+3n+5mn}{\sqrt{5+6m+5m^2}\sqrt{5+6n+5n^2}}\nonumber \\= & {} \displaystyle \frac{5kn^2+3(k+1)n+5}{\sqrt{5k^2n^2+6kn+5}\sqrt{5n^2+6n+5}}. \end{aligned}$$
(32)

From Eq. (32), we know that the value of \(Corr[\mu ,\nu ]\) changes with different values of k and n. As a further investigation, numerical experiments of calculating the correlation coefficient \(Corr[\mu , \nu ]\) with different combinations (nk) are performed and the results are shown in Table 1, in which \(Corr[\mu ,\nu ]=\widetilde{1}\) represents the correlation coefficient of \(\mu \) and \(\nu \) is approximately equal to the supremum 1 (the absolute difference is less than \(10^{-4}\)), and \(Corr[\mu ,\nu ]=\underline{ 0.6}\) indicates that the correlation coefficient of \(\mu \) and \(\nu \) is approximately equal to the infimum 0.6. Table 1 shows that the correlation coefficient of any two zigzag uncertain variables \(\mu \) and \(\nu \) takes the value in (0.6, 1]. Moreover, \(Corr[\mu ,\nu ]=1\) if \(k=1\), and \(Corr[\mu ,\nu ]<1\) if \(k\ne 1\).

Table 1 Correlation coefficients of zigzag uncertain variables

Figure 2 shows the distributions of four zigzag uncertain variables \({\mathcal {Z}}(0,2,6),{\mathcal {Z}}(1,2,4),{\mathcal {Z}}(1,3,7),\) and \({\mathcal {Z}}(3,4,6)\). It is easy to see that the correlation coefficients between each other are all equal to 1 since \(k=1\) holds. Figure 3 shows the distributions of four zigzag uncertain variables \({\mathcal {Z}}(0,2,5),{\mathcal {Z}}(1,2,5),{\mathcal {Z}}(1,3,4),\) and \({\mathcal {Z}}(2,4,6)\). It can also be seen that the correlation coefficients between each other are all less than 1.

Fig. 2
figure 2

The distributions of four zigzag uncertain variables with correlation coefficient 1

Fig. 3
figure 3

The distributions of four zigzag uncertain variables with correlation coefficient less than 1

Example 10

Consider the correlation coefficient of two normal uncertain variables and . According to the calculation formula for variance and the result in Example 6, we have

$$\begin{aligned} V[\mu ]=\sigma _1^2, \quad V[\nu ]=\sigma _2^2, \quad Cov[\mu ,\nu ]=\sigma _1\sigma _2. \end{aligned}$$

Correspondingly, the correlation coefficient of \(\mu \) and \(\nu \) is

$$\begin{aligned} Corr[\mu ,\nu ] =\displaystyle \frac{\sigma _1\sigma _2}{\sigma _1\sigma _2}=1. \end{aligned}$$

In other words, the correlation coefficient of any two normal uncertain variables is equal to 1. Figure 4 shows the distributions of four normal uncertain variables and . The correlation coefficients between any two of them are equal to 1.

Fig. 4
figure 4

The distributions of four normal uncertain variables

Example 11

Consider the correlation coefficient of two uncertain variables \(\mu \sim \mathcal{L}(a, b)\) and . According to the calculation formula for variance and the result of Example 7, we obtain

$$\begin{aligned} V[\mu ]=\displaystyle \frac{(b-a)^2}{12},\quad V[\nu ]=\sigma ,\quad Cov[\mu ,\nu ]=\displaystyle \frac{\sqrt{3}(b-a)\sigma }{2\pi }. \end{aligned}$$

Then the correlation coefficient of \(\mu \) and \(\nu \) is

$$\begin{aligned} Corr[\mu ,\nu ] =\displaystyle \frac{\displaystyle \frac{\sqrt{3}(b-a)\sigma }{2\pi }}{\displaystyle \frac{(b-a)\sigma }{2\sqrt{3}}}=\displaystyle \frac{3}{\pi }. \end{aligned}$$

In probability theory, the correlation coefficient is a measure to describe the degree of linear dependence. That the absolute value of correlation coefficient equals to 1 implies that the two random variables have a linear relationship. That is, if \(|Corr[\mu ,\nu ]|=1\) holds for two random variables \(\mu \) and \(\nu \), then we have \(\mu =a\nu +b\) for real numbers \(a\ne 0\) and b. However, the understanding of the correlation coefficient of two uncertain variables is different from its probability counterpart. From Examples 811, it can be seen that the correlation coefficient of two uncertain variables is equal to 1 if the two uncertain variables have the same type of distributions, for instance, two linear uncertain variables, two normal uncertain variables, or two zigzag uncertain variables that have a proportional relation as described in Example 9. As a reasonable deduction, the correlation coefficient of two uncertain variables can be used as an effective tool for measuring the degree relevance (similarity) between their distributions, which also explains why \(Cov[\mu ,\nu ]=0\) does not hold for two independent uncertain variables \(\mu \) and \(\nu \) as mentioned in Remark 1.

3.3 Properties of covariance and correlation coefficient

In the following, it is proved that the covariance and correlation coefficient of uncertain variables have some important properties including symmetry, linearity, and distributivity.

Theorem 11

Let \(\mu \) and \(\nu \) be two regular uncertain variables with finite expected values. Then

$$\begin{aligned} Cov[\mu ,\nu ]=Cov[\nu ,\mu ], \end{aligned}$$
(33)

and

$$\begin{aligned} Corr[\mu ,\nu ]=Corr[\nu ,\mu ]. \end{aligned}$$
(34)

Proof

The proof is elementary,and will be omitted. \(\square \)

Theorem 12

Let \(\mu \) and \(\nu \) be two regular uncertain variables with finite expected values. Then

$$\begin{aligned} Cov[a\mu ,b\nu ]=abCov[\mu ,\nu ] \end{aligned}$$
(35)

and

$$\begin{aligned} Corr[a\mu ,b\nu ]=Corr[\mu ,\nu ] \end{aligned}$$
(36)

for any real numbers a and b with \(ab>0\).

Proof

Let us denote the distributions of \(\mu \) and \(\nu \) by \(\Psi \) and \(\Upsilon \), respectively. If \(a>0\) and \(b>0\), on the basis of formula (26) for the covariance of strictly monotone functions (see Theorem 9) and the linearity of expected value (see Theorem 4), it follows that

$$\begin{aligned} Cov[a\mu ,b\nu ]= & {} \displaystyle \int _{0}^{1}(a\Psi ^{-1}(\theta )-E[a\mu ])(b\Upsilon ^{-1}(\theta )-E[b\nu ])\mathrm{d}\theta \\= & {} \displaystyle \int _{0}^{1}(a\Psi ^{-1}(\theta )-aE[\mu ])(b\Upsilon ^{-1}(\theta )-bE[\nu ])\mathrm{d}\theta \\= & {} ab\displaystyle \int _{0}^{1}(\Psi ^{-1}(\theta )-E[\mu ])(\Upsilon ^{-1}(\theta )-E[\nu ])\mathrm{d}\theta \\= & {} abCov[\mu ,\nu ]. \end{aligned}$$

Similarly, if \(a<0\) and \(b<0\), we have

$$\begin{aligned} Cov[a\mu ,b\nu ]= & {} \displaystyle \int _{0}^{1}(a\Psi ^{-1}(1-\theta )-aE[\mu ])(b\Upsilon ^{-1}(1-\theta )-bE[\nu ])\mathrm{d}\theta \\= & {} ab\displaystyle \int _{0}^{1}(\Psi ^{-1}(\theta )-E[\mu ])(\Upsilon ^{-1}(\theta )-E[\nu ])\mathrm{d}\theta \\= & {} abCov[\mu ,\nu ]. \end{aligned}$$

In conclusion, (35) holds for any real numbers a and b with \(ab>0\). In addition, based on Definition 9 and Eq. (35), we have

$$\begin{aligned} Corr[a\mu ,b\nu ]=\displaystyle \frac{Cov[a\mu ,b\nu ]}{\sqrt{V[a\mu ]}\sqrt{V[b\nu ]}}=\displaystyle \frac{abCov[\mu ,\nu ]}{\sqrt{V[a\mu ]}\sqrt{V[b\nu ]}}. \end{aligned}$$

According to the linearity of variance (see Theorem 6), we get \(V[a\mu ]=a^2V[\mu ]\) and \(V[b\nu ]=b^2V[\nu ]\). Then we immediately obtain

$$\begin{aligned} Corr[a\mu ,b\nu ]=\displaystyle \frac{abCov[\mu ,\nu ]}{ab\sqrt{V[\mu ]}\sqrt{V[\nu ]}}=Corr[\mu ,\nu ]. \end{aligned}$$

\(\square \)

Remark 3

It should be noted that following from the original definition of covariance (see Definition 8) and the linearity of expected value (see Theorem 4), it can be quickly deduced that Eq. (35) holds for any real numbers a and b without the assumption \(ab>0\).

Theorem 13

Let \(\mu \), \(\nu \) and \(\delta \) be independent regular uncertain variables with finite expected values. Then

$$\begin{aligned} Cov[\mu +\nu ,\delta ]=Cov[\mu ,\delta ]+Cov[\nu ,\delta ]. \end{aligned}$$
(37)

Proof

Denote the distributions of \(\mu \), \(\nu \) and \(\delta \) by \(\Psi \), \(\Upsilon \) and \(\Phi \), respectively. From formula (26) (see Theorem 9), we get

$$\begin{aligned} Cov[\mu +\nu ,\delta ]=\displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )+\Upsilon ^{-1}(\theta )-E[\mu +\nu ]\right) \left( \Phi ^{-1}(\theta )-E[\delta ]\right) \mathrm{d}\theta . \end{aligned}$$

Next, based on the linearity of expected value (see Theorem 4), we obtain that

$$\begin{aligned} Cov[\mu +\nu ,\delta ]= & {} \displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )+\Upsilon ^{-1}(\theta )-E[\mu ]-E[\nu ]\right) \left( \Phi ^{-1}(\theta )-E[\delta ]\right) \mathrm{d}\theta \\= & {} \displaystyle \int _{0}^{1}\left( \Psi ^{-1}(\theta )-E[\mu ]\right) \left( \Phi ^{-1}(\theta )-E[\delta ]\right) \mathrm{d}\theta \\&+\,\displaystyle \int _{0}^{1}\left( \Upsilon ^{-1}(\theta )-E[\nu ]\right) \left( \Phi ^{-1}(\theta )-E[\delta ]\right) \mathrm{d}\theta . \end{aligned}$$

Finally, according to Stipulation 1,

$$\begin{aligned} Cov[\mu +\nu ,\delta ]=Cov[\mu ,\delta ]+Cov[\nu ,\delta ]. \end{aligned}$$

\(\square \)

Theorem 14

Let \(\mu _1,\mu _2,\ldots , \mu _n\) and \(\nu \) be independent regular uncertain variables with finite expected values. Then

$$\begin{aligned} Cov\left[ \sum \limits _{i=1}^{n}\mu _i,\nu \right] =\sum \limits _{i=1}^{n}Cov[\mu _i,\nu ]. \end{aligned}$$
(38)

Proof

The proof follows from Theorem 13, by induction. \(\square \)

Theorem 15

Let \(\mu _1,\mu _2,\ldots , \mu _n\) and \(\nu _1,\nu _2,\ldots , \nu _m\) be independent regular uncertain variables with finite expected values. Then

$$\begin{aligned} Cov\left[ \sum \limits _{i=1}^{n}\mu _i,\sum \limits _{j=1}^{m}\nu _j\right] =\sum \limits _{i=1}^{n}\sum \limits _{j=1}^{m}Cov[\mu _i,\nu _j]. \end{aligned}$$
(39)

Proof

It follows immediately from Theorem 14. \(\square \)

4 Relation between variance and covariance

From Theorem 8, we know that variance can be considered as a special type of covariance, that is, \(V[\mu ]=Cov[\mu ,\mu ]\). In this section, we further discuss and analyze the relation between the variance and covariance of uncertain variables including some equalities and inequalities.

Theorem 16

Let \(\mu _1,\ldots ,\mu _n\) be independent regular uncertain variables with finite expected values. Then

$$\begin{aligned} V[\mu _{1}+\cdots +\mu _{n}]=\sum \limits _{i=1}^{n}V[\mu _{i}]+2\sum \limits _{i=1}^{n-1}\sum \limits _{j=i+1}^{n}Cov[\mu _{i},\mu _{j}]. \end{aligned}$$
(40)

Proof

Let us denote the distributions of \(\mu _1,\ldots ,\mu _n\) by \(\Psi _1,\Psi _2,\ldots ,\Psi _n\), respectively. It follows from the operational law (see Theorem 2) and formula (20) (see Theorem 5) that

$$\begin{aligned} V[\mu _{1}+\cdots +\mu _{n}]=\displaystyle \int _0^1\left( \Psi _{1}^{-1}(\theta )+\cdots +\Psi _{n}^{-1}(\theta )-E[\mu _{1}+\cdots +\mu _{n}]\right) ^{2}\mathrm{d}\theta . \end{aligned}$$

Then based on the linearity of the expected value (see Theorem 4), we obtain

$$\begin{aligned}&V[\mu _{1}+\cdots +\mu _{n}]\\&\quad =\displaystyle \int _0^1\left( \Psi _{1}^{-1}(\theta )+\cdots +\Psi _{n}^{-1}(\theta )-\left( E[\mu _{1}]+\cdots +E[\mu _{n}]\right) \right) ^{2}\mathrm{d}\theta \\&\quad =\displaystyle \int _0^1\left( (\Psi _{1}^{-1}(\theta )-E[\mu _{1}])+\cdots +(\Psi _{n}^{-1}(\theta )-E[\mu _{n}])\right) ^{2}\mathrm{d}\theta \\&\quad =\displaystyle \int _0^1\left( \sum \limits _{i=1}^{n} \left( \Psi _{i}^{-1}(\theta )-E[\mu _{i}]\right) ^{2}\right. \\&\quad \quad \left. +\, 2\displaystyle \sum \limits _{i=1}^{n-1}\sum \limits _{j=i+1}^{n}\left( \Psi _{i}^{-1}(\theta )-E[\mu _{i}]\right) \left( \Psi _{j}^{-1}(\theta )-E[\mu _{j}]\right) \right) \mathrm{d}\theta \\&\quad =\displaystyle \int _0^1\sum \limits _{i=1}^{n}\left( \Psi _{i}^{-1}(\theta )-E[\mu _{i}]\right) ^{2}\mathrm{d}\theta \\&\quad \quad +\, 2\sum \limits _{i=1}^{n-1}\sum \limits _{j=i+1}^{n}\displaystyle \int _0^1\left( \Psi _{i}^{-1}(\theta )-E[\mu _{i}]\right) \left( \Psi _{j}^{-1}(\theta )-E[\mu _{j}]\right) \mathrm{d}\theta . \end{aligned}$$

Finally, according to formula (20) and Stipulation 1, we get

$$\begin{aligned} V[\mu _{1}+\cdots +\mu _{n}]=\sum \limits _{i=1}^{n}V[\mu _{i}]+2\sum \limits _{i=1}^{n-1}\sum \limits _{j=i+1}^{n}Cov[\mu _{i},\mu _{j}]. \end{aligned}$$

\(\square \)

Example 12

Let \(\mu \) and \(\nu \) be two independent regular uncertain variables with finite expected values. Then according to Theorem 16, it follows that

$$\begin{aligned} V[\mu +\nu ]=V[\mu ]+V[\nu ]+2Cov[\mu ,\nu ]. \end{aligned}$$

Theorem 17

Let \(\mu \) and \(\nu \) be two regular uncertain variables with finite expected values. Then

$$\begin{aligned} |Cov[\mu ,\nu ]|\le \sqrt{V[\mu ]}\times \sqrt{V[\nu ]}. \end{aligned}$$
(41)

Proof

It follows immediately from Theorem 10. \(\square \)

Example 13

Consider two uncertain variables \(\mu \sim \mathcal{L}(a,b)\) and . On the basis of the calculation formula of variance and the result of Example 7, we have

$$\begin{aligned} V[\mu ]=\displaystyle \frac{(b-a)^2}{12},\quad V[\nu ]=\sigma ^2,\quad Cov[\mu ,\nu ]=(b-a)\displaystyle \frac{\sqrt{3}\sigma }{2\pi }. \end{aligned}$$

Then

$$\begin{aligned} |Cov[\mu ,\nu ]|=\displaystyle \frac{\sqrt{3}(b-a)\sigma }{2\pi }<\displaystyle \frac{\sqrt{3}(b-a)\sigma }{6}=\sqrt{V[\mu ]}\times \sqrt{V[\nu ]}. \end{aligned}$$

Furthermore, if \(\mu \) and \(\nu \) are two normal uncertain variables and , by Example 6, it follows that

$$\begin{aligned} |Cov[\mu ,\nu ]|=\sigma _1\sigma _2=\sqrt{V[\mu ]}\times \sqrt{V[\nu ]}. \end{aligned}$$

A similar conclusion can be derived for two linear uncertain variables by Example 4.

Remark 4

It is obvious that the equality in (41) holds if and only if \(Corr[\mu ,\nu ]=1\). Examples 811 have provided some special cases through which the underlying meaning of the covariance and correlation coefficient can be better understood. An interesting problem would be to derive sufficient and necessary conditions for \(Corr[\mu ,\nu ]=1\); this needs to be studied further.

5 Conclusions

Numerical characteristics, like expected value and variance, contain important information about uncertain variables, which can be used in decision-making processes under uncertain environments. On account of the concepts of expected value \(E[\mu ]\) and variance \(V[\mu ]\) of an uncertain variable \(\mu \) introduced by Liu (2010), we defined in this paper the covariance of two uncertain variables \(\mu \) and \(\nu \) as \(Cov[\mu ,\nu ] =E[\left( \mu -E[\mu ]\right) \left( \nu -E[\nu ]\right) ]\), which is similar with the covariance in probability theory. However, since the uncertain measure is subadditive, the covariance of two uncertain variables cannot be calculated directly by using their distributions. For the sake of tackling this problem, we proposed a formula for computing the covariance inspired by the formula for variance given in Liu (2010). Subsequently, based upon this formula, we derived the covariance by means of inverse distributions.

As another important concept, the correlation coefficient of two uncertain variables was also introduced in this paper as the normalized version of the covariance. Although the forms of covariance and correlation coefficient of uncertain variables are similar with those in probability theory, their practical meanings are different. Through the calculation results of the correlation coefficient in some specific examples (see Examples 811), we can conclude that the correlation coefficient of two uncertain variables indicates the degree of relevance between their distributions. In other words, the larger the correlation coefficient \(Corr[\mu ,\nu ]\) is, the higher the degree of similarity between the distributions of \(\mu \) and \(\nu \) is. One consequence is that the covariance of two independent uncertain variables is not equal to zero. Such results are different from those for random variables, and the essential reason is the difference between an uncertain measure and a probability measure. As a future theoretical study of the covariance and correlation coefficient, necessary and sufficient conditions for \(Corr[\mu ,\nu ]=1\) should be investigated.

Moreover, the results and conclusions proposed in this paper should make an important contribution to practical applications of the covariance. For instance, in order to evaluate the value at risk (VaR) of portfolio investment with uncertain returns and control investment risk, the covariance as well as variance should be analyzed based upon the calculation formulae for the covariance of uncertain variables suggested in this paper. In addition, the analysis of covariance can also be used to examine the result of uncertain regression. More applications related to covariance and correlation coefficient will be carried out in future work.