Keywords

1 Introduction

Fuzzy random variables originated as a branch of fuzzy mathematics in [1,2,3]. They are widely used in financial mathematics, forecasting, decision theory, and others. In particular, the mathematical model of a random experiment with fuzzy outcomes is interpreted as a fuzzy random variable. The current state of the theory of fuzzy random variables is reflected in [4,5,6,7] and others.

In this paper, a new definition of the quasi-scalar product between fuzzy random variables is introduced, and its relationship with the covariance of fuzzy random variables proposed in [8] is revealed. Extreme properties of expectations and fuzzy expectations of fuzzy random variables are discussed.

The main results of this work are devoted to the law of large numbers (LLN) for fuzzy random variables and linear regression of fuzzy random variables. The difference between our result and those known from the LLN consists in using a special metric associated with the quasi-scalar product introduced by the author. The specificity of our result on linear regression is to derive a formula for optimal linear regression coefficients similar to the Fourier coefficients for an orthonormal system in Hilbert space. This is provided by introduction the definition quasi-scalar product.

In addition, it is shown that under certain conditions, the optimal solution has the maximum cosine of the angle with the predicted fuzzy random variable in the class of linear estimates.

Below, by the fuzzy set A given on the universal space U, we mean the set of ordered pairs \((u, \mu _A(u))\), where the membership function \(\mu _A: U\rightarrow [0,1]\), determines the degree to which \(\forall u\in U\) belongs to the set A.

We rely on the following definition of a fuzzy number (cf. [9] chap. 2–4). A fuzzy number is called a fuzzy set whose universal set is the set of real numbers R, and which additionally satisfies the following conditions:

  1. 1)

    the support (supp) of a fuzzy number is a closed and bounded (compact) set of real numbers;

  2. 2)

    the membership function of a fuzzy number is convex;

  3. 3)

    the membership function of a fuzzy number is normal, i.e. the supremum of the membership function is equal to one;

  4. 4)

    the membership function of a fuzzy number is semi-continuous from above.

We will use the interval representation of fuzzy numbers. Namely, we will assign a set of \(\alpha \)-intervals to each fuzzy number.

The set of fuzzy numbers satisfying the conditions 1)–4) is denoted by J.

As known, the set \( \alpha \)-level of a fuzzy number \(\tilde{z}\in J\) with the membership function \(\mu _{\tilde{z}}(x)\) is defined by the relation

$$\begin{aligned} Z_{\alpha } = \{x| \mu _{\tilde{z}}(x)\ge \alpha \}\,\,\,(\alpha \in (0, 1]), \,\,\,Z_0 = supp(\tilde{z}). \end{aligned}$$

According to the above assumptions 1)–4) all \(\alpha \)-levels of a fuzzy number are closed and bounded intervals at the real axis. Let’s denote the left (lower) border of the interval \(z^{-}(\alpha )\), and the right (upper) - \(z^{+}(\alpha )\), i.e. \(Z_{\alpha } = [z^{-}(\alpha ), z^{+}(\alpha )]\). Sometimes \(z^{-}(\alpha )\) and \(z^{+}(\alpha )\) they are called the left and right indices of a fuzzy number, respectively.

On a set of fuzzy numbers, you can enter the definition of distances between them in different ways. The interval approach sometimes uses the Hausdorff distance, which for fuzzy numbers \( \tilde{z}\), \(\tilde{u}\in J\) with \(\alpha \)-level sets \(Z_{\alpha } = [z^{-}(\alpha ), z^{+}(\alpha )]\) and \(U_{\alpha } = [u^{-}(\alpha ), u^{+}(\alpha )]\) in accordance with [5] is defined by the formula \(\rho _H(\tilde{z}, \tilde{u}) = \sup \limits _{0<\alpha \le 1}d_H(Z_{\alpha }, U_{\alpha })\), where

$$\begin{aligned}d_H(Z_{\alpha }, U_{\alpha }) = \max \left[ \sup \limits _{z\in Z_{\alpha }}\inf \limits _{u\in u_{\alpha }} |z-u|, \sup \limits _{u\in u_{\alpha }}\inf \limits _{z\in z_{\alpha }} |z-u|\right] \end{aligned}$$

- Hausdorff metric. Some other distances are also considered (see, for example., [7, 8, 10]).

Denote by \(J_0\) - the set of fuzzy numbers in the interval representation of which the left and right indices are quadratically summable.

We use the distance \(\rho (\tilde{z}, \tilde{u})\) between the fuzzy numbers \(\tilde{z}\) and \(\tilde{u}\) from \(J_0\) with \(\alpha \)-level sets \(Z_{\alpha } = [z^{-}(\alpha ), z^{+}(\alpha )]\) and \(U_{\alpha } = [u^{-}(\alpha ), u^{+}(\alpha )]\), which is defined by the formula

$$\begin{aligned} \rho (\tilde{z}, \tilde{u}) = \left( \int _0^1(z^{-}(\alpha ) - u^{-}(\alpha ))^2+(z^{+}(\alpha ) - u^{+}(\alpha ))^2 d\alpha \right) ^{\frac{1}{2}}. \end{aligned}$$
(1)

Here and below, integration is understood by Lebesgue.

This kind of distance was previously used, for example, in [11].

Let the fuzzy number \( \tilde{z}\) correspond to \(\alpha \) - levels \(Z_{\alpha } = [z^{-}(\alpha ), Z^{+} (\alpha )]\), with \(\alpha \in (0, 1]\). Suppose, as is customary in interval analysis,

$$\begin{aligned} mid Z_{\alpha } = \frac{1}{2} (z^{+}(\alpha ) + z^{-}(\alpha )),\,\,\, rad Z_{\alpha } = \frac{1}{2} (z^{+}(\alpha ) - z^{-}(\alpha )). \end{aligned}$$

For fuzzy numbers \(\tilde{z}\) and \(\tilde{u}\) from \(J_0\), with sets of \(\alpha \) - levels \([z^{-}(\alpha ), z^{+}(\alpha )]\) and \([u^{-}(\alpha ), u^{+}(\alpha )]\), we define quasi-scalar product

$$\begin{aligned} \left\langle \tilde{z}, \tilde{u} \right\rangle = \int \limits _0^1 (mid Z_{\alpha }, mid U_{\alpha } + rad U_{\alpha } \,rad Z_{\alpha })d\alpha \end{aligned}$$
$$\begin{aligned} =0.5\int \limits _0^1 (z^{+}({\alpha })u^{+}({\alpha }) + z^{-}({\alpha })u^{-}({\alpha }))d\alpha . \end{aligned}$$
(2)

The quasinorm \(\tilde{z}\) is \(\left\langle \tilde{z}, \tilde{z} \right\rangle ^{1/2}\).

According to (1), (2) the distance between the fuzzy numbers \( \tilde{z}\) and \(\tilde{u}\) from \(J_0\) with sets of \(\alpha \) - levels \([z^{-}(\alpha ), z^{+}(\alpha )]\) and \([u^{-}(\alpha ), u^{+}(\alpha )]\) matches the quasinorm of a fuzzy number whose left index is \(z^{-}(\alpha ) - u^{-}(\alpha )\), and the right index \(z^{+}(\alpha ) - u^{+}(\alpha )\).

Under the sum of the fuzzy numbers \(\tilde{z}\) and \(\tilde{u}\) we will understand a fuzzy number with \(\alpha \) - levels \([z^{-}(\alpha ) + u^{-}(\alpha ), z^{+}(\alpha ) + u^{+}(\alpha )]\). The product of a fuzzy number \( \tilde{z}\) by a positive number c is a fuzzy number with \(\alpha \) - levels \([cz^{-}(\alpha ), cz^{+} (\alpha )]\). In the case of \(c<0\) - a fuzzy number with \(\alpha \) - levels \([cz^{+}(\alpha ), cz^{-} (\alpha )]\).

The following properties of the introduced quasi-scalar product are valid:

  1. 1)

    \(\left\langle \tilde{z}, \tilde{u} \right\rangle = \left\langle \tilde{u}, \tilde{z} \right\rangle \)  \((\forall \tilde{z}, \tilde{u}\in J_0)\);

  2. 2)

    \(\left\langle c_1\tilde{z}, c_2\tilde{u} \right\rangle = c_1c_2\left\langle \tilde{z}, \tilde{u} \right\rangle \)   \((\forall \tilde{z}, \tilde{u}\in J_0)\), provided that the product of the numbers \(c_1c_2>0\);

  3. 3)

    \(\left\langle \tilde{z}_1+\tilde{z}_2, \tilde{u} \right\rangle = \left\langle \tilde{z}_1, \tilde{u} \right\rangle + \left\langle \tilde{z}_2, \tilde{u} \right\rangle \)  \((\forall \tilde{z}_1, \tilde{z}_2, \tilde{u}\in J_0)\);

  4. 4)

    \(\left\langle \tilde{z}, \tilde{z} \right\rangle \ge 0\)   \((\forall \tilde{z}\in J_0)\) , and the condition    \(\left\langle \tilde{z}, \tilde{z} \right\rangle = 0\) is equivalent to vanishing the left and right indexes \(\tilde{z}\);

  5. 5)

    Cauchy-Bunyakovsky Inequality \(|\left\langle \tilde{z}, \tilde{u} \right\rangle |\le \left\langle \tilde{z}, \tilde{z} \right\rangle ^{1/2}\left\langle \tilde{u}, \tilde{u} \right\rangle ^{1/2}\)   \((\forall \tilde{z}, \tilde{u}\in J_0)\).

The quasi-scalar product of the form \(\left\langle \tilde{z}, \tilde{u} \right\rangle _1 = \int \limits _0^1 (mid\,\, Z_{\alpha } \, mid\,\, U_{\alpha }) d\alpha \) is considered in [6]. It is easy to see that in this case, turning the \(\left\langle \tilde{z}, \tilde{z} \right\rangle ^{1/2}_1\) to zero does not guarantee that the left and right indexes of \(\tilde{z}\) are equal to zero.

Other definitions of the scalar product of fuzzy numbers are also found in the literature (see, for example, [10]).

Note the following relationship between the quasi-scalar product (2) and the distance (1).

Set the fuzzy number \(\tilde{z}\) with indexes \(z^{-}(\alpha )\) and \(z^{+}(\alpha )\) match the vector function \(\bar{z}(\alpha ) = (z^{-}( \alpha ), z^{+}(\alpha ))^T\). Scalar product of \(\left\langle \bar{z}, \bar{u}\right\rangle \) vector functions \(\bar{z}\) and \(\bar{u}\) define by equality (2). Then \( \rho (\tilde{z}, \tilde{u}) = ||\bar{z} - \bar{u}||\).

We introduce the concept of the cosine of the angle between fuzzy numbers. For fuzzy numbers \(\tilde{z}\), \(\tilde{u}\in J_0\), we put

$$\begin{aligned} cos(\tilde{z}, \tilde{u}) = \frac{\left\langle \tilde{z}, \tilde{u}\right\rangle }{\left\langle \tilde{z}, \tilde{z}\right\rangle ^{1/2}\left\langle \tilde{u}, \tilde{u}\right\rangle ^{1/2}}. \end{aligned}$$

Note the properties of the cosine.

  1. 1.

    \(|cos(\tilde{z}, \tilde{u})|\le 1\,\,\,(for all \tilde{z}, \tilde{u}\in J_0)\).

    This follows from the Cauchy-Bunyakovsky inequality.

  2. 2.

    \(cos(\tilde{z}, \tilde{u}) = 0\), if and only if \(\tilde{z}\) and \(\tilde{u}\) quasi-orthogonal.

  3. 3.

    \(cos(\tilde{z}, \tilde{u}) = 1\), if and only if \(\tilde{z}\) and \(\tilde{u}\) are collinear, i.e. there is a number \(\lambda >0\) such that \(\tilde{z} = \lambda \tilde{u}\).

Indeed, \(cos(\tilde{z}, \tilde{u})\) matches \(cos(\bar{z}, \bar{u})\), where \(\bar{z}, \bar{u}\) are vector functions corresponding to \(\tilde{z}\) and \(\tilde{u}\), respectively. Then the condition \(cos(\tilde{z}, \tilde{u}) = cos(\bar{z}, \bar{u}) = 1\) means that \(\bar{z} = \lambda \bar{u}\), where \(\lambda >0\). This is equivalent to \(\tilde{z} = \lambda \tilde{u}\), i.e. the fuzzy numbers \(\tilde{z}\) and \(\tilde{u}\) collinear.

2 Fuzzy Random Variables and Their Averages

Let \((\varOmega , \varSigma , P)\) be a probability space, where \(\varOmega \) is a set of elementary events, \(\varSigma \) is a \(\sigma \)-algebra consisting of subsets of the set \(\varOmega \), and P is a probability measure.

A measurable map \(\tilde{X}:\varOmega \rightarrow J_0\) is called a fuzzy random variable if, for any \(\omega \in \varOmega \), the set \(\tilde{X}(\omega )\) is a fuzzy number from \(J_0\).

Consider the intervals of \(\alpha \) - levels of a fuzzy random variable \(\tilde{X}\) for a fixed \(\omega \). Namely, \(X_{\alpha }(\omega )=\{t\in R: \mu _{\tilde{X}(\omega )}\ge \alpha \}\), where \(\mu _{\tilde{X}(\omega )} \) - membership function of a fuzzy number \(\tilde{X}(\omega )\) , and \(\alpha \in (0,1]\). The interval \(X_{\alpha }(\omega )\) represent as \(X_{\alpha }(\omega ) = [X^{-}(\omega , \alpha ), X^{+}(\omega , \alpha )]\), where the boundaries are \(X^{-}(\omega , \alpha )\) and \(X^{+}(\omega , \alpha )\) - random variables. They are called, respectively, the left and right index of the fuzzy random variable \( \tilde{X}\).

Below, we will consider the class \( \mathfrak {X}\) of fuzzy random variables \(\tilde{X}\), for which indexes \(X^{-}(\omega , \alpha )\) and \(X^{+}(\omega , \alpha )\) are functions that are quadratically summable by \(\varOmega \times [0,1]\).

Put

$$\begin{aligned} x^{-}(\alpha ) = \int \limits _{\varOmega }X^{-}(\omega , \alpha )dP,\,\,\,\,x^{+}(\alpha ) = \int \limits _{\varOmega }X^{+}(\omega , \alpha )dP. \end{aligned}$$
(3)

A fuzzy number \(\tilde{x}\) with indexes defined by formula (3) is called the fuzzy expectation of a fuzzy random variable \(\tilde{X}\).

Let \(X_{\alpha }(\omega ) = [X^{-}(\omega , \alpha ), X^{+}(\omega , \alpha )] \) - interval \(\alpha \) - level of the fuzzy random variable \(\tilde{X}\). Put \(mid\,\, X_{\alpha }(\omega ) = \frac{1}{2}(X^{+}(\omega , \alpha ) + X^{-}(\omega , \alpha ))\) and \(rad\, X_{\alpha }(\omega ) = \frac{1}{2}(X^{+}(\omega , \alpha ) - X^{-}(\omega , \alpha ))\).

Expectation \(E(\tilde{X})\) a fuzzy random variable \(\tilde{X}\) is a number defined by the expression

$$\begin{aligned} E(\tilde{X}) = \int \limits _{0}^{1}\int \limits _{\varOmega }mid\,\,X_{\alpha }(\omega )dPd\alpha = 0.5\int \limits _{0}^{1}\int \limits _{\varOmega }(X^{-}(\omega , \alpha ) + X^{+}(\omega , \alpha ))dPd\alpha . \end{aligned}$$
(4)

Note the equality

$$\begin{aligned}E(\tilde{X}) = \int \limits _{0}^{1}mid\,\, X_{\alpha }d\alpha = 0.5\int \limits _{0}^{1}(X^{-}(\alpha ) + X^{+}(\alpha ))d\alpha , \end{aligned}$$

where \(X^{-}(\alpha )\) and \(X^{+}(\alpha )\) are determined by formulas (3).

We define a quasi-scalar product for fuzzy random variables \( \tilde{X}\) and \(\tilde{Y}\) with \(\alpha \) - level sets \(X_{\alpha }(\omega ) = [X^{-}(\omega , \alpha ), X^{+}(\omega , \alpha )]\) and \(Y_{\alpha }(\omega ) = [Y^{-}(\omega , \alpha ), Y^{+}(\omega , \alpha )]\) formula

$$\begin{aligned} \left\langle \tilde{X}, \tilde{Y}\right\rangle = \int \limits _{0}^{1}\int \limits _{\varOmega } (mid\,\,X_{\alpha }(\omega )\,mid\,\,Y_{\alpha }(\omega ) + \,rad\,\,X_{\alpha }(\omega )\,rad\,\,Y_{\alpha }(\omega ))dPd\alpha \end{aligned}$$
$$\begin{aligned} = 0.5\int \limits _{0}^{1}\int \limits _{\varOmega }(X^{+}(\omega , \alpha )Y^{+}(\omega , \alpha ) + X^{-}(\omega , \alpha )Y^{-}(\omega , \alpha ))dPd\alpha . \end{aligned}$$
(5)

In this case, the quasinorm of the fuzzy random variable \( \tilde{X}\) will be denoted \(||\tilde{X}|| = \left\langle \tilde{X}, \tilde{X} \right\rangle ^{1/2}\).

Note that the same properties 1)–5) hold for the quasi-scalar product (5) as for the quasi-scalar product of fuzzy numbers.

Some other definitions of the scalar product of fuzzy random variables may be found in [6, 10], and others.

Fuzzy random variables \(\tilde{X}\) and \(\tilde{Y}\) with \(\alpha \)-level intervals \([X(\omega , \alpha )^{- }, X(\omega , \alpha )^{+}]\) and \([Y(\omega , \alpha )^{- }, Y(\omega , \alpha )^{+}]\) are called independent if the random variables \(X(\omega , \alpha )^{-}\) and \(Y(\omega , \alpha )^{ -}\), as well as \(X(\omega , \alpha )^{+}\) and \(Y(\omega , \alpha )^{+}\) are pairwise independent for all \(\alpha \in (0, 1]\).

It is easy to check.

Statement 1. for independent fuzzy random variables \(\tilde{X}\) and \(\tilde{Y}\) their quasiscalar product \(\left\langle \tilde{X}, \tilde{Y}\right\rangle = \left\langle \tilde{x}, \tilde{y}\right\rangle \) , where \( \tilde{x}, \tilde{y}\) -fuzzy expectations \(\tilde{x}\) and \(\tilde{y}\) , respectively.

Define the distance between fuzzy random variables \(\tilde{X}\) and \(\tilde{Y}\) of class \(\mathfrak {X}\) expression

$$\begin{aligned} d(\tilde{X}, \tilde{Y}) =(\int \limits _{0}^{1}\int \limits _{\varOmega }(\left[ X^{-}(\omega , \alpha ) - Y^{-}(\omega , \alpha )\right] ^2 \end{aligned}$$
$$\begin{aligned} + \left[ X^{+}(\omega , \alpha ) - Y^{+}(\omega , \alpha )\right] ^2)dPd\alpha )^{1/2}. \end{aligned}$$
(6)

Definition (6) corresponds to the definition of the distance between fuzzy numbers (1). Other definitions of the distance between fuzzy-random variables are used, for example, in the works [3, 7, 10].

It turns out that the expectation \(E(\tilde{X})\) and the fuzzy expectation \(\tilde{x}\) of a fuzzy random variable \(\tilde{X}\) have certain extreme properties with respect to distances (1) and (6), respectively.

Denote by \( \hat{y}\) a singleton corresponding to the number \(y\in R\), i.e. a fuzzy number characterized by the membership function \( \mu _{\hat{y}}(x)\) equal to 1 for \(x=y\) and zero in other cases. By definition, all left and right indexes of \( \hat{y}\) are equal to y.

The following statements take place.

Statement 2. For a given fuzzy random variable \(\tilde{X}\) with indexes \(X^{-}(\omega , \alpha )\), \(X^{+}(\omega , \alpha )\) its expectation is \(E(\tilde{X})\) is the only solution to the extreme problem

$$\begin{aligned} d(\tilde{X}, \hat{y}) \rightarrow \min \,\,\,\,(\forall y\in R), \end{aligned}$$

where the distance is \(d(\tilde{X}, \hat{y})\) is defined by the formula (6).

Statement 3. Expectation \(E(\tilde{X})\) is the only solution to the following extreme problem

$$\begin{aligned} \rho (\tilde{x}, \hat{y})\rightarrow \min \,\,\,(\forall y\in R), \end{aligned}$$

where is the distance \(\rho \) defined by the formula (2).

These statements are verified by applying an extreme sign for scalar differentiable functions \(f(y)=d^2 (\tilde{X}, y)\) and \(g = \rho ^2(\tilde{x}, y)\), respectively, taking into account the expectation definition \(E(\tilde{X})\) and fuzzy expectation \(\tilde{x}\) of a fuzzy random variable \(\tilde{X}\).

The following theorem is true.

Theorem 1

The fuzzy expectation \( \tilde{x}\) of a fuzzy random variable \(\tilde{X}\) is the solution to the following extreme problem

$$\begin{aligned} d(\tilde{X}, \tilde{y})\rightarrow \min \,\,\,(\forall \tilde{y}\in J_0). \end{aligned}$$

The proof is just to check equality

$$\begin{aligned} d^2(\tilde{X}, \tilde{y}) = d^2(\tilde{X}, \tilde{x}) + \rho ^2(\tilde{x}, \tilde{y}) \,\,\,(\forall \tilde{y}\in J_0). \end{aligned}$$

We emphasize that the average fuzzy random variables in various aspects are widely discussed in the literature. However, their extreme properties were not previously observed.

3 The Law of Large Numbers

According to [8] we define the covariance between fuzzy random variables \(\tilde{X}\) and \(\tilde{Y}\) with intervals of \(\alpha \)-level \([X^{-}(\omega , \alpha ) X^{+}(\omega , \alpha )]\) and \([Y^{-}(\omega , \alpha ), Y^{+}(\omega , \alpha )]\) by formula

$$\begin{aligned} cov[\tilde{X}, \tilde{Y}] = 0.5 \int \limits _{0}^1\int \limits _{\varOmega } ((X^{-}(\omega , \alpha )-x^{-}(\alpha )) (Y^{-}(\omega , \alpha )-y^{-}(\alpha )) \end{aligned}$$
$$\begin{aligned} +(X^{+}(\omega , \alpha )-x^{+}(\alpha )) (Y^{+}(\omega , \alpha )-y^{+}(\alpha )))dPd\alpha .\end{aligned}$$
(7)

where \(x^{-}(\alpha )\) and \(x^{+}(\alpha )\) defined by formulas (3) and similarly \(y^{-}(\alpha )\) and \(y^{+}(\alpha )\).

This definition is convenient for us because it is closely related to the quasi-scalar product (5) and distance (6) that we have introduced. Various definitions of covariance of fuzzy random variables are found in the literature. In particular, in [6], the covariance \(cov[\tilde{X}, \tilde{Y}]\) fuzzy random variables \(\tilde{X}\), \(\tilde{Y}\) is similar to (7) expression

$$\begin{aligned} cov[\tilde{X}, \tilde{Y}] = 0.25\int \limits _0^1\int \limits _{\varOmega }(X^{-}(\omega , \alpha ) + X^{+}(\omega , \alpha ) - x^{-}(\alpha ) \end{aligned}$$
$$\begin{aligned} -x^{+}(\alpha ))(Y^{-}(\omega , \alpha ) + Y^{+}(\omega , \alpha ) - y^{-}(\alpha ) - y^{+}(\alpha ))dPd\alpha . \end{aligned}$$

However, it is easy to see that this formula actually includes covariances of random variables \(mid\,\,X_{\alpha }(\omega )\) and \(mid\,\, Y_{\alpha }(\omega )\), but not for \(rad\,\, X_{\alpha }(\omega )\) and \(rad \,\, Y_{\alpha }(\omega )\).

In this sense, the formula (7) used below more adequately reflects the structure of fuzzy random variables.

The definition of covariance (7) has a lot of properties, that are such a modification of the case of real random variables (see, [8]).

$$\begin{aligned} 1) \,cov[\tilde{X}+\tilde{Z}, \tilde{Y}] = cov[\tilde{X}, \tilde{Y}]+ cov[\tilde{Z}, \tilde{Y}]; \end{aligned}$$
$$\begin{aligned} 2) \, cov[c_1\tilde{X}, c_2\tilde{Y}] =c_1c_2cov[\tilde{X}, \tilde{Y}], \end{aligned}$$

for any real \(c_1, c_2\in R\) such that \(c_1c_2>0\).

This definition of sum fuzzy random variables and product of fuzzy random variable with real number understands as the respective definition for fuzzy numbers above.

A specific property of the covariance of fuzzy random variables defined by formula (7) with the quasi-scalar product (5) introduced by us (and not noted in [8]) is the following

$$\begin{aligned} 3.\,cov[\tilde{X}, \tilde{Y}] = \left\langle \tilde{X}, \tilde{Y}\right\rangle - \left\langle \tilde{x}, \tilde{y}\right\rangle . \end{aligned}$$

This property (for other definitions of covariance) was considered, for example, in [6, 10].

As usual, fuzzy random variables \(\tilde{X}_1, \tilde{X}_2\) are called uncorrelated if \(cov[\tilde{X}_1, \tilde{X}_2] = 0\).

Remark 1

If the fuzzy random variables \(\tilde{X}_1, \tilde{X}_2\) are independent, they are uncorrelated.

This follows from property 3 of the covariance given statement 1.

Remark 2

If the fuzzy random variables \(\tilde{X}_1, \tilde{X}_2\) are uncorrelated, then \(\left\langle \tilde{X}_1, \tilde{X}_2\right\rangle = \left\langle \tilde{x}_1, \tilde{x}_2\right\rangle \), where \(\tilde{x}_1, \tilde{x}_2\) - fuzzy expectations \(\tilde{x}_1\) and \(\tilde{x}_2\). Conversely, if the previous equality is satisfied, the fuzzy random variables \(\tilde{X}_1, \tilde{X}_2\) are uncorrelated.

This follows from property 3 of the covariance.

We define the variance of the fuzzy random variable \(\tilde{X}\) the equation \(D(\tilde{X}) = cov[\tilde{X}, \tilde{X}]\) and note its properties (cf. [8]):

  1. 1.

    \( D(c\tilde{X}) = c^2D(\tilde{X})\) for any real number c.

  2. 2.

    \(D(\tilde{X}+\tilde{Y}) = D(\tilde{X}) + D (\tilde{Y}) + 2cov[\tilde{X}, \tilde{Y}]\) for \(\forall \tilde{X}, \tilde{Y}\in \mathfrak {X}\).

  3. 3.

    \(D (\tilde{z}) = 0\) for any fuzzy number \(\tilde{z}\in J_0\).

    Important for us is the following special property of the dispersion

  4. 4.
    $$\begin{aligned} D(\tilde{X}) = \frac{1}{2}d^2(\tilde{X}, \tilde{x}) \,\,\,(\forall \tilde{X}\in \mathfrak {X}), \end{aligned}$$

where \(\tilde{x}\) is the fuzzy expectation of a fuzzy random variable \(\tilde{X}\), and \(d^2(\tilde{X}, \tilde{x})\) is the distance defined by the formula (6).

It follows from the equality \(D(\tilde{X}) = cov[\tilde{X}, \tilde{X}]\) and the definitions (6), (7).

Consider for fuzzy random variables looks Chebyshev’s inequality (see, e.g., [12], Chap. 6, Sect. 32 to “normal” random variables).

Lemma 1

(Chebyshev’s Inequality). For a fuzzy random variable \(\tilde{X}\) with a fuzzy expectation \(\tilde{x}\) and a given \(\varepsilon >0\), the inequality occurs

$$\begin{aligned} P(d(\tilde{X}, \tilde{x})\ge \varepsilon )\le \frac{2}{\varepsilon ^2}D(\tilde{X}). \end{aligned}$$
(8)

Indeed, by the probability properties

$$\begin{aligned} P(d(\tilde{X}, \tilde{x})\ge \varepsilon ) = \int \limits _{d(\tilde{X}, \bar{x})\ge \varepsilon }dP. \end{aligned}$$

Since in the integration domain \(\frac{d^2(\tilde{X}, \bar{x})}{\varepsilon ^2}\ge 1\), then

$$\begin{aligned} \int \limits _{d(\tilde{X}, \bar{x})\ge \varepsilon }dP\le \frac{1}{\varepsilon ^2}\int \limits _{\varOmega }d^2(\tilde{X}, \bar{x})dP = \frac{1}{\varepsilon ^2}d^2(\tilde{X}, \bar{x}). \end{aligned}$$

Whence, taking into account the property 4 of the variance, follows (8).

Inequality (8) is similar to the corresponding inequality from [8], but it uses a different definition of distance.

Let’s look how the law of large numbers turns out in the case of fuzzy random variables. There are a significant number of publications on this subject (see, for example, [7, 13,14,15]). The main difference is in determining the distance between fuzzy numbers (respectively, between fuzzy random variables).

Theorem 2

(Law of large numbers). Let \(\tilde{X}_1, \tilde{X}_2,...,\tilde{X}_n\) be a collection of pairwise uncorrelated fuzzy random variables with fuzzy expectations \(\tilde{x}_i\). Let their variances be uniformly bounded, i.e. there is a constant \(c>0\) such that \(D(X_i)\le c\,\,(i=1,..., n)\). Then the relation is valid

$$\begin{aligned} P(d(\frac{1}{n}\sum \limits _{i=1}^n\tilde{X}_i, \frac{1}{n}\sum \limits _{i=1}^n\tilde{x}_i)\ge \varepsilon )\le \frac{2 c}{n\varepsilon ^2}. \end{aligned}$$
(9)

Indeed, putting the Chebyshev’s inequality \(\tilde{X} = \frac{1}{n}\sum \limits _{i=1}^{n}\tilde{X}_i\), get

$$\begin{aligned} P(d(\frac{1}{n}\sum \limits _{i=1}^n\tilde{X}_i, \frac{1}{n}\sum \limits _{i=1}^n\tilde{x}_i)\le \frac{2}{\varepsilon ^2}D(\frac{1}{n}\sum \limits _{i=1}^n\tilde{X}_i). \end{aligned}$$

Further, under the properties 1, 2 of the variance we have

$$\begin{aligned} D(\frac{1}{n}\sum \limits _{i=1}^n\tilde{X}_i) = \frac{1}{n^2}D(\sum \limits _{i=1}^n\tilde{X}_i) = \frac{1}{n^2}\sum \limits _{i=1}^nD(\tilde{X}_i)\le \frac{c}{n}. \end{aligned}$$

Hence the result.

Inequality (9) implies

Corollary 1

In the conditions of Theorem 2 the relation is valid

$$\begin{aligned} P(d(\frac{1}{n}\sum \limits _{i=1}^n\tilde{X}_i, \frac{1}{n}\sum \limits _{i=1}^n\tilde{x}_i)<\varepsilon )\ge 1 - \frac{2 c}{n\varepsilon ^2}. \end{aligned}$$
(10)

The law of large numbers means that the probability on the left in (10) tends to 1 for \(n\rightarrow \infty \).

Let’s consider an important special case of the law of large numbers. It is said that fuzzy random variables \(\tilde{X}\) and \(\tilde{Y}\) with intervals of \(\alpha \) - levels \([X^{-}(\omega , \alpha ), X^{+}(\omega , \alpha )]\) and \([Y^{-}(\omega , \alpha ), Y^{+}(\omega , \alpha )]\) are equally distributed if \(X^{-}(\omega , \alpha )\) and \(Y^{-}(\omega ,\alpha )\), and \(X^{+}(\omega , \alpha )\) and \(Y^{+}(\omega , \alpha )\), are equally distributed for all \(\alpha \in [0,1]\).

It is said that \(\tilde{X}_1, \tilde{X}_2,..., \tilde{X}_n\) is a fuzzy random sample if \(\tilde{X}_i\) are independent and equally distributed. Theorem 2 implies

Corollary 2

Let \(\tilde{X}_1, \tilde{X}_2,..., \tilde{X}_n\) be a fuzzy random sample and \(\tilde{x}\) be a fuzzy expectation for each of the fuzzy random variables \(\tilde{X}_i\). Then

$$\begin{aligned} P(d(\frac{1}{n}\sum \limits _{i=1}^n\tilde{X}_i, \tilde{x})<\varepsilon )\ge 1 - \frac{2 c}{n\varepsilon ^2}, \end{aligned}$$

where c is the variance of the fuzzy random variable \( \tilde{X}_i\).

Moreover, under conditions of Corollary 2 and properties 1)–4) of the variance the convergence on metric (6) of \(\frac{1}{n}\sum \limits _{i=1}^n\tilde{X}_i\) to \(\tilde{x}\) is valid, when \(n\rightarrow \infty \).

4 Linear Regression

Let’s consider the optimal linear approximation of a (predicted) fuzzy random variable \( \tilde{Y}\) using a system of (predictive) fuzzy 1random variables \(\tilde{X}_1, \tilde{X}_2,..., \tilde{X}_n\). In a number of works [8, 11, 15,16,17] and other tasks of this kind were considered. In this case, the specifics of the problem are determined by the choice of the distance to be minimized. We investigate the question of approximating a fuzzy random variable \(\tilde{Y}\) - the linear combinations \(\sum \limits _{i=1}^n\beta _i\tilde{X}_i\)   with real coefficients \(\beta _i\)   \((i=1,...,n)\) by the criterion of minimizing the distance (6).

Consider first the extreme challenge with nonnegative coefficients \(\beta _i\ge 0\)   \((i=1,...,n)\)

$$\begin{aligned} d(\tilde{Y}, \sum \limits _{i=1}^n\beta _i\tilde{X}_i)\rightarrow \min \,\,\,\,(\forall \beta _i\ge 0). \end{aligned}$$
(11)

Takes place

Lemma 2

Let the fuzzy random variables \( \tilde{X}_i\) be quasi-orthogonal for \(i\ne j\), and their quasinorms \(\varkappa _j:=\left\langle \tilde{X}_j, \tilde{X}_j\right\rangle ^{1/2} \ne 0\)   \((j=1,..., n)\). Let the condition \(b_i = \left\langle \tilde{Y}, \tilde{X}_i\right\rangle \ge 0\)   \((i= 1,..., n)\). then problem (11) has a non-negative solution, and the only one. It has the form \( \beta _i^{*} = \frac{b_i}{\varkappa ^2_i}\),    \((i=1,..., n)\).

Indeed, due to the assumption that the coefficients \(\beta _i\) are non-negative, the left index \(\sum \limits _{i=1}^n\beta _i\tilde{X}_i\) is equal to \(\sum \limits _{i=1}^n\beta _iX_i^{-}(\omega , \alpha )\), and the right one is \( \sum \limits _{i=1}^n\beta _iX_i^{+}(\omega , \alpha )\). We will omit the arguments \(\omega , \alpha \) in the proof below. Put

$$\begin{aligned} F(\beta _1,..., \beta _n) = d^2(\tilde{Y}, \sum \limits _{i=1}^n\beta _i\tilde{X}_i) \end{aligned}$$
$$\begin{aligned} = \int \limits _0^1\int \limits _{\varOmega }((Y^{+} - \sum \limits _{i=1}^n\beta _iX_i^{+})^2 + (Y^{-} - \sum \limits _{i=1}^n\beta _iX_i^{-})^2)dPd\alpha . \end{aligned}$$
(12)

This is a quadratic form in \( \beta _1,..., \beta _n\).

Differentiate with respect to (12) for \(\beta _j\) and equate the derivative to zero

$$\begin{aligned} \frac{\partial F}{\partial \beta _j} = - 2\int \limits _0^1\int \limits _{\varOmega }((Y^{+} - \sum \limits _{i=1}^n\beta _iX_i^{+})X_j^{+} + (Y^{-} - \sum \limits _{i=1}^n\beta _iX_i^{-})X_j^{-})dPd\alpha = 0. \end{aligned}$$

Hence, for every \(j=1,2,..., n\) we have

$$\begin{aligned} \sum \limits _{i=1}^n\beta _i\int \limits _0^1\int \limits _{\varOmega }(X_i^{+}X_j^{+}+X_i^{-}X_j^{-})dPd\alpha = \int \limits _0^1\int \limits _{\varOmega }(Y^{+}X_j^{+}+Y^{-}X_j^{-})dPd\alpha , \end{aligned}$$

i.e.

$$\begin{aligned} \sum \limits _{i=1}^n\beta _i\left\langle \tilde{X}_i, \tilde{X}_j\right\rangle = \left\langle \tilde{Y}, \tilde{X}_j\right\rangle \,\,\,(j=1,...,n), \end{aligned}$$
(13)

We introduce the following notation. Vector B with coefficients \(b_i = \left\langle \tilde{Y}, \tilde{X}_i\right\rangle \), matrix A with coefficients \(A_{ij} = \left\langle \tilde{X}_i, \tilde{X}_j\right\rangle \), vector \( \beta \) with coefficients \(\beta _i\). In vector form, system (13) has the form \(A \beta = B. \) Matrix A due to the quasi-orthogonality of the system \(\{\tilde{X}_i\}\) has a diagonal form, with positive numbers on the main diagonal \(\varkappa _i^2\)   \((i=1,..., n)\). then the solution is \( \beta ^{*} = A^{-1}B\), i.e. \(\beta _i^{*} = \frac{b_i}{\varkappa ^2_i}\)   \((i=1...n)\).

The nonnegativity of the obtained coefficients \( \beta _i^{*}\) is provided by the condition \(b_i = \left\langle \tilde{Y}, \tilde{X}_i\right\rangle \ge 0\)   \((i= 1,..., n)\).

To verify that \(\beta ^{*} = A^{-1}B\) is the minimum point, consider the second derivative

$$\begin{aligned} \frac{\partial ^2\,F}{\partial \beta _j\partial \beta _s} = 2\int \limits _0^1\int \limits _{\varOmega }4(X_s^{+}X_j^{+}+X_s^{-}X_j^{-})dPd\alpha = \left\langle \tilde{X}_s, \tilde{X}_j\right\rangle \,\,\,\text {when}\,\, s\ne j. \end{aligned}$$
$$\begin{aligned} \frac{\partial ^2\,F}{\partial \beta _j^2} = 2\int \limits _0^1\int \limits _{\varOmega }((X_j^{+})^2+(X_j^{-})^2)dPd\alpha = 4\left\langle \tilde{X}_j, \tilde{X}_j\right\rangle \,\,\,\,\text {when}\,\, s = j. \end{aligned}$$

A sufficient sign of a minimum is the positive definiteness of the Hesse matrix \(\{\frac{\partial ^2 F}{\partial \beta _j\partial \beta _s}\}\) . And this is provided by the quasi-orthogonality of the system \(\{\tilde{X}_j\}\).

Remark 3

Condition \(\left\langle \tilde{Y}, \tilde{X}_i\right\rangle > 0\) means that there is an acute angle between the fuzzy random variables \(\tilde{Y}\) and \(\tilde{X}_i\). In other words, the fuzzy-random variables \( \tilde{Y}\) and \(\tilde{X}_i\) increase (in this sense) or decrease at the same time.

Remark 4

In the conditions of Lemma 2, we can reject the requirement of pairwise quasi-orthogonality of fuzzy random variables \(\tilde{X}_1,..., \tilde{X}_n\). It is sufficient to require positive invertibility of their Gram matrix A with coefficients \(a_{ij} = \left\langle \tilde{X}_i, \tilde{X}_j\right\rangle \). In the sense that the inverse matrix \(A^{-1}\) exists and converts vectors with non-negative coordinates back to vectors with non-negative coordinates.

Remark 5

If, under Lemma 2, we reject the requirement of pairwise quasi-orthogonality of fuzzy random variables \( \tilde{X}_1,..., \tilde{X}_n\), but additionally assume their pairwise uncorrelability, then it is sufficient to require positive invertibility of the Gram matrix from their fuzzy expectations \(\left\langle \tilde{x}_i, \tilde{x}_j\right\rangle \).

We emphasize that the coefficients \( \beta _i^{*}\) are analogous to the Fourier coefficients when decomposing in an orthogonal system in a Hilbert space (see, for example, [17], Chap. II, Sect. 11 for random variables). This is due to the relationship of the metric to be minimized in problem (11) with quasi-scalar product (5).

The proximity of fuzzy random variables (as well as any space with a scalar product) can be characterized by the cosine of the angle between them.

Define the cosine between the fuzzy random variables \(\tilde{Y}, \tilde{Z}\) by the equality

$$\begin{aligned} \cos (\tilde{Y}, \tilde{Z}) = \frac{\left\langle \tilde{Y}, \tilde{Z}\right\rangle }{||\tilde{Y}||||\tilde{Z}||}. \end{aligned}$$
(14)

According to the definition (14) and the properties of the cosine of the angle between the fuzzy numbers \(|cos (\tilde{Y}, \tilde{Z})|\le 1\). In this case, \(cos(\tilde{Y}, \tilde{Z}) = 0\), if and only, if \(\tilde{Y}\) and \(\tilde{Z}\) are quasi-orthogonal. And \(cos(\tilde{Y}, \tilde{Z}) = 1\), if and only, if \(\tilde{Z} = \lambda \tilde{Y}\) are collinear \((\lambda >0)\).

Denote, as in Lemma 2, \( \beta _i^{*} = \frac{1}{\varkappa ^2_i}\left\langle \tilde{Y}, \tilde{X}_i\right\rangle \) and consider

$$\begin{aligned} \tilde{Z}_n^{*} = \sum \limits _{i=1}^n\beta _i^{*}\tilde{X}_i \end{aligned}$$
(15)

- an optimal estimate of the predicted fuzzy random variable \( \tilde{Y}\) from Lemma 2.

Theorem 3

Let the conditions of Lemma 2. Then the optimal estimate (15) has the maximum cosine with the predicted fuzzy random variable \(\tilde{Y}\) in the class of linear estimates of the form \(Z_n = \sum \limits _{i=1}^n\beta _i\tilde{X}_i \,\,\,(\beta _i\ge 0)\).

Indeed, we will show that

$$\begin{aligned} |cos(\tilde{Y}, \tilde{Z}_n)|\le cos(\tilde{Y}, \tilde{Z}_n^ {*}). \end{aligned}$$

Due to the properties of the quasi-scalar product and the non-negativity of the coefficients \(\beta _i^{*} \ge 0\) we have

$$\begin{aligned} \left\langle \tilde{Y}, \tilde{Z}_n^{*}\right\rangle = \left\langle \tilde{Y}, \sum \limits _{i=1}^n\beta _i^{*}\tilde{X}_i\right\rangle = \sum \limits _{i=1}^n\beta _i^{*}\left\langle \tilde{Y}, \tilde{X}_i\right\rangle = \sum \limits _{i=1}^n(\beta _i^{*})^2\varkappa _i^2. \end{aligned}$$

In this case, due to the pairwise quasi-orthogonality of the system \(||Z_n^{*}||^2 = \sum \limits _{i=1}^n(\beta _i^{*})^2\varkappa _i^2\). Then

$$\begin{aligned} cos(\tilde{Y}, \tilde{Z}_n^ {*}) = \frac{\sum \limits _{i=1}^n\varkappa _i^2(\beta _i^{*})^2}{||\tilde{Y}||(\sum \limits _{i=1}^n\varkappa _i^2(\beta _i^{*})^2)^{1/2}} = \frac{1}{||\tilde{Y}||}(\sum \limits _{i=1}^n\varkappa _i^2(\beta _i^{*})^2)^{1/2}.\ \end{aligned}$$

Consider

$$\begin{aligned} \left\langle \tilde{Y}, \tilde{X}_i\right\rangle = \sum \limits _{i=1}^n\beta _i\left\langle \tilde{Y}, \tilde{X}_i\right\rangle = \sum \limits _{i=1}^n\beta _i\beta _i^{*}\varkappa _i^2\end{aligned}$$

and \(||\tilde{Z}_n||^2 = \sum \limits _{i=1}^n\beta _i^2\varkappa _i^2 \).

Then

$$\begin{aligned}cos(\tilde{Y}, \tilde{Z}_n) = \frac{\sum \limits _{i=1}^n\varkappa _i^2\beta _i\beta _i^{*}}{||\tilde{Y}||(\sum \limits _{i=1}^n\beta _i^2\varkappa _i^2)^{1/2}}. \end{aligned}$$

By the Cauchy-Schwarz inequality

$$\begin{aligned} |cos(\tilde{Y}, \tilde{Z}_n)|\le \frac{(\sum \limits _{i=1}^n\varkappa _i^2\beta _i^2)^{1/2}(\sum \limits _{i=1}^n\varkappa _i^2(\beta _i^{*})^2)^{1/2}}{||\tilde{Y}||(\sum \limits _{i=1}^n \varkappa _i^2\beta _i^2)^{1/2}} \end{aligned}$$
$$= \frac{1}{||\tilde{Y}||}(\sum \limits _{i=1}^n\varkappa _i^2(\beta _i^{*})^2)^{1/2} = cos(\tilde{Y}, \tilde{Z}_n^{*}),$$

which was required to be proved.

Let’s consider the optimal regression problem in a situation where all linear approximation coefficients are not assumed to be nonnegative and the condition \(\left\langle \tilde{Y}, \tilde{X}_i\right\rangle \ge 0\,\,\,(i=1,...,n)\) is met.

Note that the explicit form of the formula for the distance \(d(\tilde{Y}, \sum \limits _{i=1}^n\beta _i\tilde{X}_i)\) in the case of coefficients \(\beta _i\) of an arbitrary sign is inconvenient for research, since in this case the product of the interval \(\alpha \) - the level of a fuzzy number \(\tilde{z}\) by a clear number \(\beta \) is given by the cumbersome expression

$$ \beta [z^{-}, z^{+}] = [\min \{\beta z^{-}, \beta z^{+}\}, \max \{\beta z^{-}, \beta z^{+}\}]. $$

However, in the general situation, the following statement is true. Let’s say \(c_{*} = \max \limits _{j=1,..., n}\{\frac{||\tilde{Y}||}{||\tilde{X}_j||}\}\).

Theorem 4

Let the fuzzy-random variables \( \tilde{X}_i\) be quasi-orthogonal for \(i\ne j\), and all their quasinorms \(\varkappa _i\ne 0\,\, (i=1,..., n)\). then the problem is

$$\begin{aligned} d(\tilde{Y}, \sum \limits _{i=1}^n\beta _i\tilde{X}_i)\rightarrow \min \,\,\,(\beta _i\in [-c_{*}, \infty )) \end{aligned}$$
(16)

has a solution, and the only one. It has the form \( \beta _i^{*} = \frac{b_i}{\varkappa _i^2}\,\, (i=1,..., n)\) .

Note though problem (16) does not assume that the coefficients \(\beta _i\) are positive, the formula for the coefficients \(\beta _i\) has the same form as in Lemma 2. At the same time the coefficients \(b_i\) in the condition of Theorem 4 may have different signs.

In the proof of Theorem 4, the following special property of the distance (6) between fuzzy random variables will be used.

Lemma 3

For any fuzzy random variables \(\tilde{X}\), \(\tilde{Y}\) and \(\tilde{W}\) in \(\mathfrak {X}\) the next equality holds

$$ d(\tilde{X}+\tilde{W}, \tilde{Y}+\tilde{W}) = d(\tilde{X}, \tilde{Y}). $$

In fact, this is true because subject to the rules of interval addition on the left are

$$ (\tilde{X}+\tilde{W})^{-} = X^{-}+W^{-},\,\,(\tilde{Y}+\tilde{W})^{-} = Y^{-}+W^{-},\,\, $$

and similarly for the right indexes.

After substituting the corresponding expressions in (6), we obtain the required equality.

Proof

of Theorem 4. Let the condition \(\left\langle \tilde{Y}, \tilde{X}_i\right\rangle \ge 0\) not be satisfied for at least one j. Consider the fuzzy random variable \(\tilde{Z}=\tilde{Y}+c_{*}\sum \limits _{i=1}^n\tilde{X}_i\). According to the definition \(c_{*}>0\), \(\left\langle \tilde{Z}, \tilde{X}_j\right\rangle \ge 0\)    \((j = 1,..., n)\). Consider for \(\tilde{Z} \) task (11). Let \(\gamma _i\ge 0\) be the optimal coefficients of a linear combination \(\sum \limits _{i=1}^n\gamma _i\tilde{X}_i\) for \(\tilde{Z}\), obtained by solving problem (11). The vector \(\gamma \) with coordinates \(\gamma _i\) is defined by the formula \(\gamma = A^{-1}f\), for \(f_i = \left\langle \tilde{Z}, \tilde{X}_j\right\rangle \).

We show that the coefficients \( \gamma _i - c_{*}\) are optimal for linear approximation of a fuzzy random variable \(\tilde{Y}\) by the system \(\{\tilde{X}_i\}\).

Consider the distance \(d (\tilde{Y}, \sum \limits _{i=1}^n(\gamma _i-c_{*})X_i)\). By Lemma 3 and taking into account the definition of \( \tilde{Z}\), we have

$$\begin{aligned} d(\tilde{Y}, \sum \limits _{i=1}^n(\gamma _i-c_{*})X_i) = d(\tilde{Y} +\sum \limits _{i=1}^nc_{*}\tilde{X}_i, \sum \limits _{i=1}^n\gamma _i\tilde{X}_i) = d(\tilde{Z}, \sum \limits _{i=1}^n\gamma _i\tilde{X}_i). \end{aligned}$$
(17)

Since \(\{\gamma _i\}\) - solution of problem (11) for \(\tilde{Z}\), then in accordance with Lemma 2 for any set of numbers \(\xi _i\ge 0\)   \((i=1,...,n)\) can record

$$ d(\tilde{Z}, \sum \limits _{i=1}^n\gamma _i\tilde{X}_i)\le d(\tilde{Z}, \sum \limits _{i=1}^n\xi _i\tilde{X}_i) = d(\tilde{Y} +\sum \limits _{i=1}^nc\tilde{X}_i, \sum \limits _{i=1}^n(\xi _i-c_{*})X_i + \sum \limits _{i=1}^nc_{*}\tilde{X}_i) $$

Using Lemma 3 again, we get

$$ d (\tilde{Z}, \sum \limits _{i=1}^n\gamma _i\tilde{X}_i)\le d(\tilde{Y}, \sum \limits _{i=1}^n(\xi _i-c_{*})\tilde{X}_i). $$

Then (17) implies the inequality

$$ d(\tilde{Y}, \sum \limits _{i=1}^n(\gamma _i-c_{*})X_i) \le d(\tilde{Y}, \sum \limits _{i=1}^n(\xi _i-c_{*})X_i). $$

Since here \((\xi _i-c_{*})\) - arbitrary coefficients from a closed interval \([-c_{*}, \infty )\) , then \((\gamma _i-c_{*})\) - optimal coefficients.

Note that by definition \( \tilde{Z}\) and according to Lemma 2

$$ \gamma _j = \frac{1}{\varkappa ^2_j}\left\langle \tilde{Z}, \tilde{X}_j\right\rangle = \frac{1}{\varkappa ^2_j}\left\langle (\tilde{Y} +\sum \limits _{i=1}^nc_{*}\tilde{X}_i), \tilde{X}_j\right\rangle . $$

Then, taking into account quasiorthogonality system \(\{\tilde{X}_i\}\) will receive

$$ \gamma _j = \frac{1}{\varkappa ^2_j}\left( b_j + c_{*}\left\langle \tilde{X}_j, \tilde{X}_j \right\rangle \right) = \frac{b_j}{\varkappa ^2_j } + c_ {*}. $$

Hence, the optimal coefficients of \(\tilde{\beta }_j\) for \(\tilde{X}_j\) in the linear approximation \(\tilde{Y}\), having the form \((\gamma _j - c_{*})\), defined by the equality \(\frac{b_j}{\varkappa ^2_j} \)   \((j=1,..., n)\). this is what the statement implies.

Remark 6

Similarly to Remark 4, under the conditions of theorem 4, one can reject the quasi-orthogonality of the system \(\{\tilde{X}_i\}\) and instead assume positive invertibility of their Gram matrix. In addition, it is required that the sum of elements of all columns of the Gram matrix be positive.