Abstract
We propose a new kind of stochastic absolute value equations involving absolute values of variables. By utilizing an equivalence relation to stochastic bilinear program, we investigate the expected value formulation for the proposed stochastic absolute value equations. We also consider the expected residual minimization formulation for the proposed stochastic absolute value equations. Under mild assumptions, we give the existence conditions for the solution of the stochastic absolute value equations. The solution of the stochastic absolute value equations can be gotten by solving the discrete minimization problem. And we also propose a smoothing gradient method to solve the discrete minimization problem. Finally, the numerical results and some discussions are given.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Let \((\Omega , {{{{\mathcal {F}}}}}, \rho )\) be a probability space, where \(\Omega \subseteq R^n\) and \(\rho \) is a standard probability measure, we propose a new kind of stochastic absolute value equations, which is to find a vector \(x\in R^n\) such that
where \(A(\omega )\in R^{n\times n}\) and \(b(\omega )\in R^n\) for \(\omega \in \Omega \) are random quantities on a probability space \((\Omega , {{{{\mathcal {F}}}}}, \rho )\), |x| is the componentwise absolute value of vector \(x\in R^n\). We call (1.1) the stochastic absolute value equations (SAVE). When \(A(\omega )\) is a deterministic matrix and \(b(\omega )\) is a deterministic vector, then SAVE (1.1) reduces to the absolute value equation (AVE) which is equivalent to the general linear complementarity problem [1,2,3,4]. The AVE was widely used in solving linear programs, bimatrix games and fundamental problems of mathematical programming, one can see [2,3,4,5]. In the past few decades, the stochastic variational inequality problems [6, 7], the stochastic linear complementarity problems [8,9,10,11,12], the stochastic nonlinear complementarity problems [13, 14] and the stochastic tensor complementarity problems [15,16,17] were also widely studied in solving many optimization problems with uncertainty. However, no attention has been paid to SAVE (1.1) which contains the characteristics of AVE and stochastic optimization problems.
As the AVE is an NP hard problem [2], it is also a hard work to solve SAVE (1.1). Generally, for the stochastic optimization problems, there are two general approaches to get the solution of the problems [8,9,10]. The first approach applies the expected value (EV) method which formulates the problem as a deterministic problem by taking the expect of the stochastic quantity, and the second approach is the expected residual minimization (ERM) method, which is a natural extension of the least-squares method of minimizing the residual. In this paper, the equivalent relation between SAVE (1.1) and stochastic bilinear program is given. By using the EV formulation, we propose an expected value formulation for SAVE (1.1). We also study the ERM formulation for SAVE (1.1). We generate samples by the quasi-Monte Carlo methods and prove that every accumulation point of the discrete approximation problem is the solution of the expected residual minimization problem for SAVE (1.1).
The remainder of this paper is organized as follows. In Sect. 2, we show that SAVE (1.1) is equivalent to a stochastic bilinear program, which is a stochastic optimization problem with the formula as a stochastic generalized linear complementarity problem. Combined with an example, we give a discussion about the EV formulation. In Sect. 3, we first establish the boundedness of the solution set of the expected residual minimization problem, and then show that each accumulation point of the sequence generated by the ERM formulation is a solution of the expected residual minimization problem. In Sect. 4, we propose a smoothing gradient method for solving SAVE (1.1). Some numerical experiments are also given to verify the theoretical results of the ERM formulation. Finally, we complete our paper with some conclusions in Sect. 5.
2 Expected value formulation
We start by showing that SAVE (1.1) is equivalent to a stochastic bilinear program. By the equivalence of the stochastic bilinear program and the stochastic generalized linear complementarity problem, SAVE (1.1) can be reformulated as a stochastic generalized linear complementarity problem. Then the expected value formulation will be used to solve SAVE (1.1).
Theorem 2.1
SAVE (1.1) is equivalent to the stochastic bilinear program, i.e.,
Proof
By SAVE (1.1), from \(|x|\le A(\omega )x-b(\omega )\), we have
i.e., the above formulations are the equivalence of the constraints for the stochastic bilinear program. So we have
We complete the proof. \(\square \)
Theorem 2.2
SAVE (1.1) is equivalent to the stochastic generalized linear complementarity problem, i.e.,
Proof
By the equivalence of the stochastic bilinear program and the stochastic generalized linear complementarity problem, we get this theorem. \(\square \)
In the following of this paper, \(E[\cdot ]\) stands for the expectation of every elements of matrix and vector. \({\bar{A}}\) denotes the expectation of \(A(\omega )\) and \({\bar{b}}\) denotes the expectation of \(b(\omega )\), i.e.,
Then, we get the expected value formulation of the stochastic generalized linear complementarity problem as
In general, the solution set of (2.1) is not equivalent to the solution set of (2.2) for all \(\omega \in \Omega \). So, in this section, we consider a kind of discrete probability space, which has only finitely many elements, i.e., \(\Omega =\{\omega _{1},\omega _{2},\cdots ,\omega _{m}\}\). Now, (2.1) is equivalent to
where \(G(x)=({\bar{A}}+I)x-{\bar{b}}\), \(H(x)=({\bar{A}}-I)x-{\bar{b}}\), \(i=1,2,\cdots ,m\).
In the following of this section, we reformulate (2.3) as a nonlinear equations with nonnegative constraints, i.e., the expected value formulation of SAVE (1.1). (2.2) is a generalized linear complementarity problem, and it can be reformulated as a semismooth equations by Fischer-Burmeister (FB) function. FB function is an NCP function [1], which is defined as
where \(a,b\in R\). Then x is a solution of (2.3) if and only if
where
and
with
Now, we give a simple example to illustrate the transformation process.
Example 2.1
Consider SAVE (1.1), where
\(\Omega =\{\omega _1,\omega _2\}=\{0,2\},P_i={\mathcal {P}}(\omega _i\in \Omega )=\frac{1}{2}, i=1,2.\)
We know that \((1, 1, 1, 1)^T\) is the solution of this example. Now, we use the EV formulation to solve the above example. Firstly, we get
Then by (2.4), we know that Example 2.1 can be transformed into the following constrained equations
where \(y_i\ge 0,i=1,2,\cdots ,16\). The optimization solution of the above constrained equations is equivalence to the optimization solution of the following constrained optimization problem
We use fmincon function in Matlab Optimization Toolbox to solve the transformed constrained optimization problem. The numerical results are given in the following table, where \(x^0\) denotes the initial point, \(x^*\) denotes the optimum solution.
Remark
From the numerical results of the above example, we know that the SAVE (1.1) with finite discrete distribution can be solved by constrained optimization methods. But the EV transformation is a more complicated form with nonsmooth complementarity function and only solve SAVE (1.1) with finite discrete distribution. So, in the following section, we consider the expected residual minimization formulation, which can avoid transforming the SAVE into a complicated constrained optimization problem. And the expected residual minimization formulation can also be used to solve SAVE (1.1) with any distribution involving the finite discrete distribution.
3 Expected residual minimization formulation
To apply the expected residual minimization formulation to solve SAVE (1.1), we first formulate the problem as the following optimization problem
where \(F(x)=E[\Vert A(\omega )x-|x|-b(\omega )\Vert ^2]=\int _{\Omega }\Vert A(\omega )x-|x|-b(\omega )\Vert ^2\rho (\omega )\,{\mathrm d}\omega \). Discrete the involved problem by the quasi-Monte Carlo method, then the solution of the original problem can be approximated obtained by solving the discrete minimization problem.
To proceed, we give the following assumption.
Assumption 3.1
Let \(\rho :\Omega \rightarrow R_{+}\) be a continuous probability density function on probability space \((\Omega , {{{{\mathcal {F}}}}}, P)\). Suppose that
where \(A(\omega )\in R^{n\times n}\), \(b(\omega )\in R^n\), \(\omega \in \Omega \).
For \(\gamma >0\), we denote the level set of function F by \(\Xi (\gamma )\), i.e.,
Lemma 3.1
Suppose that there exists an \({\bar{\omega }}\in \Omega \), such that \(\rho ({\bar{\omega }})>0\) and \(A(\omega )\ne diag(sign(x))\). Then the level set \(\Xi (\gamma )\) is bounded.
Proof
By \(\rho \) is continuous, there exists \(\rho _0>0\) such that
where \(D({\bar{\omega }},\delta )\) is a closed sphere with center \({\bar{\omega }}\) and radius \(\delta \). We now consider a sequence \(\{x_k\}\in R^n\). By the continuity of \(\phi \), then there exists \(\omega _k\in B\), such that
where \(\phi (x_k,\omega )=A(\omega )x_k-|x_k|-b(\omega )\).
Denote \(\lambda =\int _{B}d\omega >0\). Then
Now, we only need to prove \(\Vert \phi (x_k,\omega _k)\Vert \rightarrow +\infty \) as \(\Vert x_k\Vert \rightarrow +\infty \). Suppose \(\Vert x_k\Vert \rightarrow +\infty \) holds, we know that \(x_k^i\rightarrow +\infty \) or \(x_k^i\rightarrow -\infty \) for some i. So, we get
for some i, i.e., we get \(\Vert \phi (x_k,\omega _k)\Vert \rightarrow +\infty \) holds for \(\Vert x_k\Vert \rightarrow +\infty \). Hence, the proof is completed. \(\square \)
In the following of this section, the quasi-Monte Carlo method for numerical integration is used as in [8, 18]. The transformation function \(\omega =u(\nu )\) is used to go from an integral on \(\Omega \) to the integral on the unit hypercube \([0,1]^m\). And the observations \(\{\nu _i\}, i=1,\cdots ,{\bar{N}}\) are generated in this unit hypercube.
Then, we get
where \({\bar{\rho }}(\nu )=\rho (u(\nu ))u'(\nu ),~{\bar{\Omega }}=[0,1]^m\).
For each k, we denote
where \(\phi (x,\nu _i)=A(\nu _i)x-|x|-b(\nu _i)\), \({\bar{\Omega }}_k:=\{\nu _i,i=1,\cdots ,\bar{N_k}\}\) is a set of observations generated by a quasi-Monte Carlo method such that \({\bar{\Omega }}_k\subseteq \Omega ,~\bar{N_k}\rightarrow \infty \) as \(k\rightarrow \infty \). In the remainder of this section, to simplify the natation, we suppose \(\Omega =[0,1]^m\) and let \(\omega \) denote \(\nu \).
Now, we consider
Obviously, (3.2) is the approximation problem to (3.1).
Lemma 3.2
For any fixed \(x\in R^n\), we get
Proof
From the definition of \(\phi \), we get
i.e.,
By Assumption 3.1, we know that \((\Vert A(\omega )\Vert +1)^2\rho (\omega )\) is a nonnegative continuous function and it is also bounded. Therefore, we get \(\Vert \phi (x,\cdot )\Vert ^2\rho (\cdot )\) is integrable and \(0\le F(x)<\infty \). By \(\Vert \phi (x,\cdot )\Vert ^2\rho (\cdot )\) is continuous, we have
for \(x\in R^n.\) This completes the proof. \(\square \)
Denote \(\vartheta \) as the optimal solution set of (3.1), and \(\vartheta _k\) as the optimal solution set of (3.2). Now, we give the following theorem to show the relation of the expected residual minimization problem (3.1) and the approximate expected residual minimization problem (3.2).
Theorem 3.1
If \(\rho ({\bar{\omega }})>0\) holds for \({\bar{\omega }}\in \Omega \), then \(\vartheta _k\) is nonempty and bounded when k is large enough. And every accumulation point of \(\{x_k\}\subseteq \vartheta _k\) is contained in the set \(\vartheta \).
Proof
We assume that \(x_k\rightarrow {\bar{x}}, k\rightarrow \infty \). Let \(F({\overline{x}})<\gamma \), by the continuity of F, we know that \(F(x_k)\le \gamma \) for all large k, i.e., \(x_k\in {\bar{D}}(\gamma )\) for all large k. Now, we show that
For all \(x,y\in R^n\), we get
Denote \(L(\omega )=\Vert A(\omega )\Vert +1\), we also get
By Lemma 3.1, \({\bar{D}}(\gamma )\) is closed and bounded, we have
where \(C_1=\max \{\Vert x\Vert |x\in {\bar{D}}(\gamma )\},\) \(x,y\in {\bar{D}}(\gamma )\). By
we obtain
where \(\digamma \) is a constant and for all large k satisfying
So, for \(k\rightarrow \infty \), we get
From the above results and Lemma 3.2, we obtain
when \(k\rightarrow \infty \). By \(x_k\in \vartheta _k\), we get
Therefore, we have
We complete the proof. \(\square \)
Now, we give two special kinds of SAVE (1.1), which can be solved without using discrete approximation.
Case I. Let \(\Omega =[{\tilde{\alpha }}_1,{\tilde{\beta }}_1]\times \cdots \times [{\tilde{\alpha }}_N,{\tilde{\beta }}_N]\) with \({\tilde{\alpha }}_j<{\tilde{\beta }}_j,~j=1,\cdots ,N\), and \({\tilde{\omega }}_j,j=1,\cdots ,N\) are independent. When \(\rho \) satisfies Assumption 3.1 and \(\rho _j\) denotes the density function for \({\tilde{\omega }}_j\), \(j=1,\cdots ,N\). We know that
where
Case II. Let \(A(\omega )\equiv A\) and \(b(\omega )\equiv {\tilde{b}}+T\omega \), where \(A\in R^{n\times n}\), \({\tilde{b}}\in R^n\) and \(T\in R^{n\times m}\) are given constants. For each i, the \(i\hbox {th}\) row of the matrix T has just one positive element \(t_i\), and the density function \(\rho \) is defined by
In this case, we get \(F(x)=\sum \limits _{i=1}^nF_i(x)\), where
4 A smoothing gradient method
In this section, we use the ERM formulation to transform SAVE (1.1) into an unconstrained optimization problem. For SAVE (1.1) contains nonsmooth term |x|, we consider smoothing method to solve it. Smoothing gradient method is an effective smoothing method to deal with this kind of problems [19,20,21], so we use the smoothing gradient method to solve SAVE (1.1).
Firstly, we generate samples \(\omega ^i, i=1,2,\cdots ,N\), i.e.,
and we choose the smoothing function of \(|x_i|\) as
where \(\mu \ge 0\). Denote \(\psi (x,\mu )=(\sqrt{x_1^2+\mu }, \sqrt{x_2^2+\mu }, \cdots , \sqrt{x_n^2+\mu })^{T}\). Then SAVE (1.1) can be transformed into the following unconstrained optimization problem
And the gradient of the objective function in (4.1) is
where \(J(A(\omega _i)x-\psi (x,\mu )-b(\omega _i))\) is the Jacobian of \((A(\omega _i)x-\psi (x,\mu )-b(\omega _i))\).
Next, we give the smoothing gradient method for SAVE (1.1).
It is easy to find that \({\widetilde{f}}(\cdot ,\mu )\ge 0\) for any constant \(\mu \ge 0\) and \(\forall x\in R^n\), \(\nabla _{x}{{\widetilde{f}}(\cdot ,\mu )}\) is uniformly continuous on the level set \(L(x_0,\mu )=\{x\in R^n|{\widetilde{f}}(x,\mu )\le {\widetilde{f}}(x_0,\mu )\}\). Next, we give the global convergence of the proposed smoothing gradient method. From [22], we get the following lemma.
Lemma 4.1
Given a constant \({\overline{\mu }}\ge 0\). Let \(\{x_k\}\) be the sequence generated by Algorithm 1, then
Theorem 4.1
Let \(\{\mu _k\}\) and \(\{x_k\}\) be the sequence generated by Algorithm 1. Then
Proof
Define \(K=\{k|\mu _{k+1}=\sigma \mu _k\}\). Suppose K is a finite set, then there exists an integer \({\widehat{k}}\) such that for all \(k>{\widehat{k}}\),
set \(\mu _k=\mu _{{\widehat{k}}}={\overline{\mu }}\), we get
From Lemma 4.1, we have
which contradicts with (4.2). So K is an infinite set, i.e., \(\underset{k\rightarrow \infty }{\lim }\mu _k=0\). Set \(K=\{k_0, k_1, \cdots \}\), for \(k_0<k_1<\cdots \), then
The proof is completed. \(\square \)
In the following of this section, we verify the effectiveness of Algorithm 1 via the following given examples. The parameters used in Algorithm 1 are chosen as \(\rho =0.5\), \(\sigma =0.5\), \(\delta =0.5\), \(\mu _0=0.01\), \({\bar{\gamma }}=0.5\). We terminate our algorithm if \(k\ge 10000\) or \(\Vert \nabla _{x}{\widetilde{f}}(x_{k},\mu _{k})\Vert <1.0e-5\) satisfied. We implement all numerical test in MATLAB R2019b and run the codes on a PC with 1.80GHz CPU.
Example 4.1
Consider SAVE (1.1), where
We generate samples \(\omega ^i\), \(i=1,2,\cdots ,N\), which obey the uniform distribution of [0, 1]. The numerical results of Example 4.1 are shown in Table 2 and Fig. 1, where N denotes the number of \(\omega ^i\), \(x^0\), \(x^*\) and \(f(x^*)\) denote the initial point, the optimum solution and the optimum value, respectively.
Example 4.2
Consider SAVE (1.1), where
We generate samples \(\omega ^i\), \(i=1,2,\cdots ,N\), which obey the uniform distribution of [0, 1]. The detailed numerical results are shown in Table 3 and Fig. 2, where N denotes the number of \(\omega ^i\), \(x^0\), \(x^*\) and \(f(x^*)\) denote the initial point, the optimum solution and the optimum value, respectively.
Example 4.3
Consider SAVE (1.1), where
and
We generate samples \(\omega ^i\), \(i=1,2,\cdots ,N\), which obey the uniform distribution of [0, 1]. The initial points are randomly generated. The detailed numerical results are shown in Table 4 and Fig. 3, where N denotes the number of \(\omega ^i\), \(x^*\) and \(f(x^*)\) denote the optimum solution and the optimum value respectively.
Example 4.4
Consider SAVE (1.1), where
and
We generate samples \(\omega ^i\), \(i=1,2,\cdots ,N\), which obey the uniform distribution of [0, 1]. The initial points are randomly generated. The detailed numerical results are shown in Tables 5, 6, Figs. 4 and 5, where n denotes the dimension, N denotes the number of \(\omega ^i\), \(x^*\) and \(f(x^*)\) denote the optimum solution and the optimum value respectively.
From the above numerical results, we can see that SAVE (1.1) can be solved by simple unconstrained optimization method. And the ERM formulation can avoid transforming SAVE (1.1) into a complicated constrained optimization problem.
5 Conclusions
In this paper, we propose a new kind of absolute value equation problem with random quantities, which is called stochastic absolute value equations. The properties of the proposed stochastic absolute value equations are studied. The expected value formulation and expected residual minimization formulation for solving the proposed stochastic absolute value equations are also given. Absolute value equations is widely used in studying engineering problems, economics and management problems. It is very meaningful to study this kind of stochastic absolute value equations, which is a more extensive problem than the absolute value equations.
References
Cottle, R., Pang, J., Stone, R.: The linear complementarity problem. Academic Press, San Diego (1992)
Mangasarian, O.: Absolute value programming. Comput. Optim. Appl. 36, 43–53 (2007)
Mangasarian, O., Mayer, R.: Absolute value equation. Linear Algebra Appl. 419, 359–367 (2006)
Mangasarian, O.: Absolute value equation solution via concave minimization. Optim. Lett. 1, 3–8 (2007)
Caccetta, L., Qu, B., Zhou, G.: A globally and quadratically convergent method for absolute value equations. Comput. Optim. Appl. 48, 45–58 (2011)
Gurkan, G., Ozge, A., Robison, S.: Sample-path solution of stochastic variational inequalities. Math. Program. 84, 313–333 (1994)
Ravat, U., Shanbhag, U.: On the existence of solutions to stochastic quasi-variational inequality and complementarity problems. Math. Program. 165, 291–330 (2017)
Chen, X., Fukushima, M.: Expected residual minimization method for stochastic linear complementarity problems. Math. Oper. Res. 30, 1022–1038 (2005)
Chen, X., Zhang, C., Fukushima, M.: Robust solution of monotone stochastic linear complementarity problem. Math. Program. 117, 51–80 (2009)
Fang, H., Chen, X., Fukushima, M.: Stochastic \(R_0\) matrix linear complementarity problems. SIAM J. Optim. 18, 482–506 (2007)
Zhang, C., Chen, X.: Smoothing projected gradient method and its application to stochastic linear complementarity problems. SIAM J. Optim. 20, 627–649 (2009)
Zhang, C.: Existence of optimal solution for general stochastic linear complementarity problems. Oper. Res. Lett. 39, 78–82 (2011)
Lin, G., Fukushima, M.: New reformulations for stochastic nonlinear complementarity problems. Optim. Methods Softw. 21, 551–564 (2006)
Ling, C., Qi, L., Zhou, G., Caccetta, L.: The \(SC^1\) property of an expected residual function arising from stochastic complementarity problems. Oper. Res. Lett. 36, 456–460 (2008)
Du, S., Che, M., Wei, Y.: Stochastic structured tensors to stochastic complementarity problems. Comput. Optim. Appl. 75, 649–668 (2020)
Ming, Z., Zhang, L., Qi, L.: Expected residual minimization method for monotone stochastic tensor complementarity problem. Comput. Optim. Appl. 77, 871–896 (2020)
Du, S., Cui, L., Chen, Y., Wei, Y.: Stochastic tensor complementarity problem with discrete distribution. J. Optim. Theory Appl. 192, 912–929 (2022)
Niederreiter, H.: Random number generation and quasi-monte carlo methods. J. Am. Stat. Assoc. 88, 147–153 (1992)
Burke, J., Lewis, A., Overton, M.: A robust gradient sampling algorithm for nonsmooth, nonconvex optimization. SIAM J. Optim. 15, 751–779 (2005)
Kiwiel, K.: Convergence of the gradient sampling algorithm for nonsmooth nonconvex optimization. SIAM J. Optim. 18, 379–388 (2007)
Chen, X.: Smoothing methods for nonsmooth, nonconvex minimization. Math. Program. 134, 71–99 (2012)
Ma, C.: Optimization method and its Matlab program design. Science Press, Beijing (2010)
Acknowledgements
This work was supported by National Natural Science Foundation of China (No.11671220, 11771244). The authors wish to give their sincere thanks to the editor and the anonymous referees for their valuable and helpful comments, which help improve the quality of this paper significantly.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Du, S., Sun, J., Niu, S. et al. Stochastic absolute value equations. J. Appl. Math. Comput. 69, 921–939 (2023). https://doi.org/10.1007/s12190-022-01777-0
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12190-022-01777-0
Keywords
- Stochastic absolute value equations
- Expected value formulation
- Expected residual minimization formulation