Abstract
In this paper, by combining the logarithmic-quadratic proximal (LQP) method and alternating direction method, we proposed an LQP alternating direction method for solving structured variational inequalities. The new iterate is generated by searching the optimal step size along a descent direction with a new step size αk. The choice of the descent direction and the step size selection strategies are important for the algorithm’s efficiency. The O(1/t) convergence rate of the proposed method is studied, and its efficiency is also verified by some numerical experiments.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
This paper considers the constrained convex programming problem with the following separate structure:
where \(\theta _{1}:\mathbb {R}_{+}^{n}\rightarrow \mathbb {R}\ \) and \(\theta _{2}:\mathbb {R}_{+}^{m}\rightarrow \mathbb {R}\ \) are closed proper convex functions not necessarily smooth, \(A\in \mathbb {R}^{l\times n}\) and \(B \in \mathbb {R}^{l\times m}\) are given matrices, and \(b\in \mathbb {R}^{l}\) is a given vector.
A very rich class of applications may be modeled as problem (1.1). Various methods have been developed, as we all know, a benchmark solver for problem (1.1) is the alternating direction method (ADM), originally proposed in Glowinski and Marrocco [25]. We refer to, e.g., [7, 17, 21, 24, 26,27,28, 30,31,32,33, 35, 38, 39] for some early references of ADM in PDE and optimization literatures. In particular, ADM has recently found impressive applications in many areas such as image processing and statistical learning (see, e.g., [20] and references therein). To alleviate the subproblems when a splitting method is applied to solve (1.1), one important effort is to regularize the objective functions appropriately and thus force the solutions to stay strictly within the interiors of \(\mathbb {R}_{+}^{n}\) and \(\mathbb {R}_{+}^{m}\). A very good choice for such regularization is the LQP regularization which was originally proposed in [1] and has been studied extensively in many articles such as [2,3,4,5, 19, 29, 37, 41].
Let ∂(.) denotes the sub-gradient operator of a convex function, and f(x) ∈ ∂𝜃1(x) and g(y) ∈ ∂𝜃2(y) are the sub-gradient of 𝜃1(x) and 𝜃2(y), respectively. By attaching a Lagrange multiplier vector \(\lambda \in \mathbb {R}^{l}\) to the linear constraint Ax + By = b, problem (1.1) can be written in terms of finding \(w \in {\mathcal W}=\mathbb {R}_{+}^{n} \times \mathbb {R}_{+}^{m} \times \mathbb {R}^{l}\) such that
where
and (.)⊤ denotes the transpose.
Problem (1.2)–(1.3) is referred to as structured variational inequalities (in short, SVI).
Yuan and Li [41] developed the following LQP-based decomposition method by applying the LQP terms to regularize the ADM subproblems: For a given \(w^{k}=(x^{k},y^{k},\lambda ^{k})\in \mathbb {R}^{n}_{++}\times \mathbb {R}^{m}_{++}\times \mathbb {R}^{l}\), and μ ∈ (0,1), the new iterative (xk+ 1, yk+ 1, λk+ 1) is obtained via solving the following system:
where \( X_{k}=\text {diag}({x^{k}_{1}},...,{x^{k}_{n}})\), x− 1 is an n-vector whose j th element is 1/xj, \( Y_{k}=\text {diag}({y^{k}_{1}},...,{y^{k}_{m}}),\) y− 1 is an m-vector whose j th element is 1/yj, and \(H \in \mathbb {R}^{l \times l}\), \(R_{1} \in \mathbb {R}^{n \times n}\), and \(S_{1} \in \mathbb {R}^{m \times m}\) are symmetric positive definite, where R1 = diag(r1,...,rn) and S1 = diag(s1,...,sm).
Later, Bnouhachem et al. [9, 10] and Li [36] proposed some LQP-ADMs and made the LQP-ADM more practical. Each iteration of the above methods contains a prediction and a correction; the predictor is obtained via solving (1.4)–(1.6), and the new iterate is obtained by a convex combination of the previous point and the one generated by a projection-type method along a descent direction for [9, 36], while the new iterate is computed directly by an explicit formula derived from the original LQP method for [10]. The main disadvantage of the methods in [8,9,10,11, 36, 37, 41] is that solving the (1.5) requires the solution of (1.4). Hence, the alternating direction methods are not eligible for parallel computing in the sense that the solutions of (1.4)–(1.5) cannot be obtained simultaneously. This characteristic excludes the possibility of applying some advanced computing technologies to solve (1.4)–(1.5). To overcome this difficulty, Bnouhachem and Hamdi [12] proposed a parallel descent LQP-ADM for solving SVI. The main advantage of this method is that the predictor is obtained via solving a system of non-linear equations in a parallel wise. Very recently, Bnouhachem and Rassias [18] proposed a new LQP alternating direction scheme for the separable constrained convex programming problem; the predictor \(\tilde {w}^{k}=(\tilde {x}^{k},\tilde {y}^{k},\tilde {\lambda }^{k})\) is obtained via solving LQP system approximately under significantly relaxed accuracy criterion and the new iterate wk+ 1(αk) = (xk+ 1, yk+ 1, λk+ 1) is given by
where
and
We notice that the convergence of this method was established under the assumption that the mappings f(x) and g(y) are Lipschitz continuous. However, in many applications, the mappings f(x) and g(y) may not be Lipschitz continuous. For this situation, it could be difficult to verify their Lipschitz continuity condition. It is interesting to design a new easily implementable method for the continuous mappings whose available information is only the mapping values.
In this paper, we propose a descent LQP alternating direction method for SVI. The global convergence of the proposed method can be guaranteed by the continuity rather than the Lipchitz continuity of these mappings. Since the self-adaptive adjustment rule is necessary in practice, we propose a self-adaptive method that adjusts the scalar parameter automatically. To illustrate the proposed method and demonstrate its efficiency, some applications and their numerical results are also provided. Our results can be viewed as significant extensions of the previously known results.
2 The proposed method
We recall some concepts and results which are needed in the sequel.
For any vector \(u\in \mathbb {R}^{n}, \|u\|^{2}=u^{\top }u, \|u\|_{\infty }=\max \limits \{|u_{1}|\ldots ,|u_{n}|\}\). Let \(D \in \mathbb {R}^{n \times n}\) be a symmetric positive definite matrix; the operators λl(D) and λm(D) denote the largest eigenvalue and the smallest eigenvalue of D, respectively. We denote the D-norm of u by \(\|u\|_{D}^{2}=u^{\top } Du\).
Some fundamental properties of the projection operator are listed without proof, (see, e.g., [6]).
Lemma 2.1
Let Ω be a nonempty closed convex subset of \(\mathbb {R}^{l}\) and denoted by PΩ[.] the projection on Ω with respect to the Euclidean norm, that is,
Then, we have the following inequalities:
We recall some basic definitions,which will be used in our later analysis.
Definition 2.1
The mapping \(T:\mathbb {R}^{n}\to \mathbb {R}^{n}\) is said to be
-
(a) monotone if
$$ (Tx-Ty)^{\top}(x-y)\ge 0,\qquad \forall x,y\in \mathbb{R}^{n}; $$ -
(b) L-Lipschitz continuous if there exists a constant L > 0 such that
$$ \|Tx-Ty\|\le L\|x-y\|,\qquad\forall x,y\in \mathbb{R}^{n}; $$
We make the following standard assumptions:
Assumption A
f is monotone and continuous on \(\mathbb {R}^{n}_{+}\) and g is monotone and continuous on \(\mathbb {R}^{m}_{+}\).
Assumption B
The solution set of SVI, denoted by \({\mathcal W}^{*}\), is nonempty.
Let βk > 0, r > 0,s > 0, \(H \in \mathbb {R}^{l \times l}\), \(R \in \mathbb {R}^{n \times n}\), and \(S \in \mathbb {R}^{m \times m}\) be positive definite diagonal matrices, where R = rIn×n and S = sIm×m. We propose the following inexact LQP alternating direction method for solving SVI:
Algorithm 2.1
-
Step 0. The initial step: Given ε > 0, μ ∈ (0,1),η ∈ (0,1),ρ > 0, and \(w^{0}=(x^{0}, y^{0}, \lambda ^{0}) \in \mathbb {R}^{n}_{++}\times \mathbb {R}^{m}_{++}\times \mathbb {R}^{l}\). \({\upbeta }_{0}=\min \limits \left \{\frac {(1-\eta )\lambda _{m}(R)}{3\lambda _{l}(H)\|A\|^{2}},\frac {(1-\eta )\lambda _{m}(S)}{3\lambda _{l}(H)\|B\|^{2}}\right \}.\) Set k = 0.
-
Step 1. Prediction step: Compute \(\tilde {w}^{k}=(\tilde {x}^{k},\tilde {y}^{k},\tilde {\lambda }^{k})\in \mathbb {R}^{n}_{++}\times \mathbb {R}^{m}_{++}\times \mathbb {R}^{l}\) by solving the following system:
$$ \begin{array}{@{}rcl@{}} &&{\upbeta}_{k}\left( f(x) - A^{\top} [\lambda^{k} - H (A x^{k} + B y^{k} - b) ]\right)\\ &+& R[(x-x^{k}) + \mu (x^{k} - {X_{k}^{2}} x^{-1})]=:{\xi^{k}_{x}}\approx0, \end{array} $$(2.3a)$$ \begin{array}{@{}rcl@{}} &&{\upbeta}_{k}\left( g(y) - B^{\top} [\lambda^{k} - H (A \tilde{x}^{k}+By^{k}- b)]\right)\\&+& S[(y-y^{k}) + \mu (y^{k} - {Y_{k}^{2}} y^{-1})]=:{\xi^{k}_{y}}\approx0, \end{array} $$(2.3b)$$ \tilde{\lambda}^{k} = \lambda^{k} - H (A\tilde{x}^{k} + B \tilde{y}^{k} - b), $$(2.3c)where βk is a proper parameter which satisfies
$$ \| {\xi^{k}_{x}}\|\leq\eta r\|x^{k}-\tilde{x}^{k}\|,\qquad \| {\xi^{k}_{y}}\|\leq\eta s\|y^{k}-\tilde{y}^{k}\| $$(2.4)and
$$ \begin{array}{@{}rcl@{}} \xi^{k}=\left( \begin{array}{c} {\xi^{k}_{x}}\\ {\xi^{k}_{y}} \\ 0 \end{array}\right)={\upbeta}_{k}\left( \begin{array}{c} f(\tilde{x}^{k})-f(x^{k})+\rho A^{\top}HA(x^{k}-\tilde{x}^{k})\\ g(\tilde{y}^{k})-g(y^{k})+\rho B^{\top}HB(y^{k}-\tilde{y}^{k}) \\ 0 \end{array}\right). \end{array} $$(2.5) -
Step 2. Convergence verification: If \( \max \limits \{\|x^{k}-\tilde {x}^{k}\|_{\infty },\|y^{k}-\tilde {y}^{k}\|_{\infty },\|\lambda ^{k}-\tilde {\lambda }^{k}\|_{\infty }\}<\epsilon ,\) then stop.
-
Step 3. Correction step: The new iterate wk+ 1(αk) = (xk+ 1, yk+ 1, λk+ 1) is given by
$$ w^{k+1}(\alpha_{k})= (1-\sigma) w^{k}+\sigma P_{\mathcal W}[w^{k}-\alpha_{k}d_{2}(w^{k},\tilde{w}^{k})], \quad \sigma\in(0,1) $$(2.6)where
$$ \alpha_{k}=\frac{\varphi(w^{k}, \tilde{w}^{k})}{\|d_{1}(w^{k}, \tilde{w}^{k})\|^{2}}, $$(2.7)$$ \begin{array}{c} d_{2}(w^{k},\tilde{w}^{k})= \end{array}\left( \begin{array}{c} {\upbeta}_{k}(f(\tilde{x}^{k}) - A^{\top} \tilde{\lambda}^{k})+{\upbeta}_{k} A^{\top}H(A(x^{k}-\tilde{x}^{k})+B(y^{k}-\tilde{y}^{k}))\\ {\upbeta}_{k}(g(\tilde{y}^{k}) - B^{\top} \tilde{\lambda}^{k})+{\upbeta}_{k} B^{\top}H(A(x^{k}-\tilde{x}^{k})+B(y^{k}-\tilde{y}^{k}))\\ {\upbeta}_{k}(A\tilde{x}^{k}+B\tilde{y}^{k}-b) \end{array}\right) $$(2.8)$$ \begin{array}{@{}rcl@{}} \varphi(w^{k}, \tilde{w}^{k})&=&(w^{k}-\tilde{w}^{k})^{\top}d_{1}(w^{k},\tilde{w}^{k})-\mu\| x^{k} -\tilde{x}^{k} \|_{R}^{2}-\mu\| y^{k} - \tilde{y}^{k} \|_{S}^{2} \\&&+{\upbeta}_{k}(\lambda^{k}-\tilde{\lambda}^{k})^{\top}(A(x^{k}-\tilde{x}^{k})+B(y^{k}-\tilde{y}^{k})), \end{array} $$(2.9)$$ \begin{array}{c} d_{1}(w^{k},\tilde{w}^{k}){\kern-.5pt}={\kern-4.5pt} \end{array}\left( \begin{array}{c} (1+\mu)R(x^{k}-\tilde{x}^{k})-{\upbeta}_{k}[f(x^{k})-f(\tilde{x}^{k})]+\rho{\upbeta}_{k} A^{\top}HA(x^{k}-\tilde{x}^{k}) \\ {\kern-4.5pt}(1+\mu)S(y^{k}-\tilde{y}^{k})-{\upbeta}_{k}[g(y^{k})-g(\tilde{y}^{k})] +{\upbeta}_{k} B^{\top}HA(x^{k}-\tilde{x}^{k})+\rho{\upbeta}_{k} B^{\top}HB(y^{k}-\tilde{y}^{k})\\ {\upbeta}_{k}H^{-1}(\lambda^{k}-\tilde{\lambda}^{k}) \end{array}{}\right) $$(2.10) -
Step 4. Adjusting:
Adaptive rule of choosing a suitable βk+ 1 as the start prediction step size for the next iteration
$$ {\upbeta}_{k+1} :=\left\{\begin{array}{ll} \min\{{\upbeta}_{0}, \tau*{\upbeta}_{k}\} & \text{ if }\max\{r_{1},r_{2}\} \le 0.5, \\\ {\upbeta}_{k} & \text{ otherwise}. \end{array}\right. $$(2.11)
where
and τ > 1. Set k := k + 1 and go to Step 1.
Remark 2.1
In general, the prediction step is implementable. Sometimes we can get the approximate solution of (2.3a)–(2.3c) directly via choosing a suitable βk > 0. Since R = rIn×n and S = sIm×m and if f is Lipschitz continuous on \(\mathbb {R}^{n}_{+}\) with Lipschitz constant kf, and g is Lipschitz continuous on \(\mathbb {R}^{m}_{+}\) with Lipschitz constant kg. Therefore, criterion (2.4) is satisfied when \({\upbeta }_{k}\leq \min \limits \{\frac {\eta r}{k_{f}+\rho \|A^{\top }HA\|},\frac {\eta s}{k_{g}+\rho \|B^{\top }HB\|}\}.\)
2.1 Relationship with some existing methods
Our method can be viewed as an extension and improvement for some well-known results, for example, the following:
-
It follows from (2.5) that (2.3a)–(2.3b) can be written as
$$ (R+\rho {\upbeta}_{k} A^{T}HA)(\tilde{x}^{k})^{2}-s^{k}\tilde{x}^{k}-\mu R {X_{k}^{2}}=0, $$(2.12)$$ (S+\rho {\upbeta}_{k} B^{T}HB)(\tilde{y}^{k})^{2}-p^{k}\tilde{y}^{k}-\mu S {Y_{k}^{2}}=0, $$(2.13)with
$$ s^{k}=(1-\mu)Rx^{k}-{\upbeta}_{k}(f(x^{k}) - A^{T} [\lambda^{k} - H ((1-\rho)A x^{k} + B {y}^{k} - b) ]), $$$$ p^{k}=(1-\mu)Sy^{k}-{\upbeta}_{k}(g(y^{k}) - B^{T} [\lambda^{k} - H (A \tilde{x}^{k} + (1-\rho)B y^{k} - b)]). $$The proposed method obtains the predictors \(\tilde {x}^{k}\) and \(\tilde {y}^{k}\) quite easily by an explicit formula. Since R = rIn×n and S = sIm×m, and if A = Il×n, B = −Il×m, and H = Il×l, the positive solution of (2.12)–(2.13) can be obtained explicitly by
$$ \tilde{x}^{k}_{i}=\frac{\left( \bar{s}^{k}_{i}+\sqrt{(\bar{s}^{k}_{i})^{2}+4\mu r(r+{\upbeta}_{k}\rho)({x^{k}_{i}})^{2}}\right)}{2(r+{\upbeta}_{k}\rho)}, $$(2.14)$$ \tilde{y}^{k}_{i}=\frac{\left( \bar{p}^{k}_{i}+\sqrt{(\bar{p}^{k}_{i})^{2}+4\mu s(s+{\upbeta}_{k}\rho)({y^{k}_{i}})^{2}}\right)}{2(s+{\upbeta}_{k}\rho)} $$(2.15)with
$$ \bar{s}^{k}=r(1-\mu)x^{k}-{\upbeta}_{k}\left( f(x^{k}) - [\lambda^{k} - ((1-\rho) x^{k} - {y}^{k} - b) ]\right), $$$$ \bar{p}^{k}=s(1-\mu)y^{k}-{\upbeta}_{k}\left( g(y^{k}) + [\lambda^{k} - (\tilde{x}^{k} - (1-\rho) y^{k} - b)]\right). $$Therefore, the proposed method is easily implementable. In contrast, the predictors \(\tilde {x}^{k}\) and \(\tilde {y}^{k}\) in [22, 34, 40, 42] could be computed by using the projection method but the applicability of the projection method is limited due to the fact that it is not easy to find the projection except in very special cases.
-
If βk = 1,∀k ≥ 0 and \(x^{k+1}=\tilde {x}^{k}, y^{k+1}=\tilde {y}^{k}\) and \(\lambda ^{k+1}=\tilde {\lambda }^{k}\) in (2.3a), (2.3b) and (2.3c), respectively, and \({\xi ^{k}_{x}}={\xi ^{k}_{y}}=0,\) we obtain the method proposed in [41].
-
The methods proposed in [8,9,10,11,12, 36] have suggested to solve problem (2.3a)–(2.3b) (with βk = 1, ∀k ≥ 0) exactly. It is more practical to find approximate solutions of problems (2.3a)–(2.3b) rather than the exact solutions due to the fact that in general this excludes some practical applications. Driven by the fact of eliminating this drawback, we solve problem (2.3a)–(2.3b) approximately. Since the self-adaptive adjustment rule is necessary in practice, we propose a self-adaptive method that adjusts the scalar parameter βk automatically.
-
The convergence of the method proposed in [18] was established under the assumption that the mappings f(x) and g(y) are Lipschitz continuous. However, in many applications, the mappings f(x) and g(y) may not be Lipschitz continuous. For this case, it could be difficult to verify their Lipschitz continuity condition. The convergence of the proposed method can be guaranteed under the assumption that the mappings f(x) and g(y) are continuous. Moreover, the parameter βk in the prediction procedure of the proposed method is selected self-adaptively.
-
Comparing the proposed method with the methods in [8,9,10,11,12,13,14,15,16, 18, 36], the new iterate is obtained by using a new direction with a new step size αk.
We need this lemma to analyze the convergence for the proposed method.
Lemma 2.2
[41] Let \(q(u)\in \mathbb {R}^{n}\) be a monotone mapping of u with respect to \(\mathbb {R}^{n}_{^{+}}\) and \(R\in \mathbb {R}^{n\times n}\) be a positive definite diagonal matrix. For a given uk > 0, if \(U_{k}:= \text {diag}({u^{k}_{1}},{u^{k}_{2}},\cdots , {u^{k}_{n}})\) and u− 1 be an n-vector whose j th element is 1/uj, then the equation
has a unique positive solution u. Moreover, for any v ≥ 0, we have
Now we are ready to present an inequality where a lower bound of \(\varphi (w^{k}, \tilde {w}^{k})\) is found for all \(w^{k}\in \mathbb {R}_{++}^{n} \times \mathbb {R}_{++}^{m}\times \mathbb {R}^{l}\). This inequality is also crucial for analyzing the contraction property and the convergence for the iterative sequence.
Theorem 2.1
For given \(w^{k}\in \mathbb {R}_{++}^{n} \times \mathbb {R}_{++}^{m}\times \mathbb {R}^{l},\) let \(\tilde {w}^{k}\) be generated by (2.3a)–(2.3c), then there exists two constants α1 > 0 and α2 > 0 such that
and
Proof
Since
It follows from the definition of \(\varphi (w^{k}, \tilde {w}^{k})\) that
By using Cauchy-Schwarz inequality in (2.4), we have
where
Then, we have
where α1 > 0 is a constant. The third inequality holds because βk ≤β0 for any k. The fourth inequality is obtained from the definition of β0.
Recalling the definition in (2.10), we rewrite \(d_{1}(w^{k}, \tilde {w}^{k})\) as
where
Note that, for any \(a, b \in \mathbb {R}^{n},\) we have
It follows that
where α2 > 0 is a constant. Therefore, it follows from (2.7) and (2.18) that
and this completes the proof. □
3 Basic results
In this section, to prove the global convergence for the proposed method, we first present some lemmas.
Lemma 3.1
For given \(w^{k}=(x^{k}, y^{k},\lambda ^{k})\in \mathbb {R}_{++}^{n} \times \mathbb {R}_{++}^{m}\times \mathbb {R}^{l},\) let \(\tilde {w}^{k}\) be generated by (2.3a)–(2.3c). Then for any \(w=(x,y, \lambda ) \in {\mathcal W}\), we have
Proof
Applying Lemma 2.2 to (2.3a) by setting uk = xk, \(u=\tilde {x}^{k}\), v = x in (2.17) and
we get
Recall
Adding (3.2) and (3.3), we obtain
Similarly, applying Lemma 2.2 to (2.3b); substituting uk = yk, \( u= \tilde {y}^{k}\), and v = y; and replacing R, n with S, m, respectively, in (2.17) and
we get
Recall
Adding (3.5) and (3.6), we have
It follows from (3.4), (3.7), (2.3c), and (2.5) that
and the assertion of this lemma is proved. □
Lemma 3.2
For given \(w^{k}=(x^{k}, y^{k},\lambda ^{k})\in \mathbb {R}_{++}^{n} \times \mathbb {R}_{++}^{m}\times \mathbb {R}^{l},\) let \(\tilde {w}^{k}\) be generated by (2.3a–2.3c). Then for any \(w^{*}=(x,y, \lambda ) \in {\mathcal W}^{*}\), we have
Proof
Recalling the definition in (2.9)
Using the monotonicity of f and g, we obtain
It follows from (3.11) that
Combining (3.10) and the above inequality, we can get the assertion of this lemma. □
The following theorem provides a unified framework for proving the convergence of the proposed method.
Theorem 3.1
Let \(w^{*}\in {\mathcal W}^{*}, w^{k+1}(\alpha _{k})\) be defined by (2.6) and
then
Proof
Since \(w^{*}\in {\mathcal W}^{*}\) and
it follows from (2.2) that
From (2.6), we get
Using the following identity
for \(a=w^{k}-{w^{k}_{p}}\), \(b={w^{k}_{p}}-w^{*},\) and (3.16), we obtain
Using the definition of Θ(αk) and (3.17), we get
Applying (3.1) (with \(w={w_{p}^{k}}\)), we obtain
Adding (3.9) and (3.19), we get
Applying (3.20) to the last term on the right side of (3.18), we obtain
and the theorem is proved. □
4 Convergence of the proposed method
In this section, we prove the global convergence of the proposed method. From the computational point of view, a relaxation factor γ ∈ (0,2) is preferable in the correction. The assertion (3.14) enables us to study the contraction property of the iterative sequence.
Theorem 4.1
Let \(w^{*}\in {\mathcal W}^{*}\) be a solution of SVI and let wk+ 1(γαk) be generated by (2.6). Then wk and \(\tilde {w}^{k}\) are bounded, and
where
Proof
It follows from (3.14), (2.18), and (2.19) that
Since γ ∈ (0,2), we have
and thus, {wk} is a bounded sequence.
It follows from (4.1) that
which means that
Since {wk} is a bounded sequence, we conclude that \(\{\tilde {w}^{k}\}\) is also bounded. □
With Lemma 3.1 and Theorem 4.1 at hand, we are able to prove the convergence of the proposed method. The following result can be proved by using the technique of Theorem 4.2 in [19].
Theorem 4.2
The sequence {wk} generated by the proposed method converges to some \(w^{\infty }\) which is a solution of SVI.
Proof
Since {wk} is bounded, it has at least one cluster point. Let \(w^{\infty }\) be a cluster point of {wk} and the subsequence \(\{ w^{k_{j}}\}\) converges to \(w^{\infty }\), since \({\mathcal W}\) is closed set, we have \(w^{\infty }\in {\mathcal W}\). By the construction of βk, we have that 0 < βk ≤β0,∀k. It follows from (4.2) that
Moreover, (4.3) and (3.1) imply that
and consequently
which means that \(w^{\infty }\) is a solution of SVI.
Now we prove that the sequence {wk} converges to \(w^{\infty }.\) Since
for any 𝜖 > 0, there exists an l > 0 such that
Therefore, for any k ≥ kl, it follows from (4.1) and (4.5) that
This implies that the sequence {wk} converges to \(w^{\infty }\) which is a solution of SVI. □
5 O(1/t) convergence rate
In this section, we show that the proposed method has the O(1/t) convergence rate. Recall that \({\mathcal W}^{*}\) can be characterized as (see (2.3.2) in pp. 159 of [23])
This implies that \(\hat {w}\) is an approximate solution of SVI with the accuracy 𝜖 > 0 if it satisfies
For the rest, our purpose is to show that after t iterations of the proposed method, we can find a \(\hat {w}\in {\mathcal W}\) such that (5.1) is satisfied with 𝜖 = O(1/t).
Our analysis needs a new sequence defined by
Based on (3.10) and (5.2), we easily have a relationship
Using (1.3), (2.8), and (5.2), we obtain
Lemma 5.1
Let \(\hat {w}^{k}\) be defined by (5.2) and \(w\in {\mathcal W}\), then, we have
Proof
It follows from (3.4) and (3.7) that
and
Then, by using the notation of \(\hat {w}^{k}\) in (5.2), (5.6), and (5.7) can be written as
and
In addition, it follows from (2.3c) and (5.2) that
Combining (5.8)–(5.10), recall the definition of \(d_{1}(w^{k},\tilde {w}^{k})\), we obtain the assertion (5.5). The proof is completed. □
Lemma 5.2
For given \(w^{k}\in \mathbb {R}_{++}^{n} \times \mathbb {R}_{++}^{m}\times \mathbb {R}^{l},\) let \({w_{p}^{k}}\) be defined by (3.15) and \(w\in {\mathcal W}\), then we have
Proof
Since \({w_{p}^{k}}\in {\mathcal W},\) substituting \(w={w_{p}^{k}}\) in (5.5) and using (5.3), we get
On the other hand, using (3.15) and (5.4), \({w_{p}^{k}}\) is the projection of \(w^{k}-\gamma \alpha _{k} {\upbeta }_{k}Q(\hat {w}^{k})\) on \(\mathcal W,\) it follows from (2.1) that
and consequently
Using the identity \(a^{\top }b =\frac {1}{2}\left (\|a\|^{2}-\|a-b\|^{2}+\|b\|^{2}\right )\) to the right hand side of the last inequality, we obtain
Adding (5.12) and (5.13), we get
and by using the monotonicity of Q, we obtain (5.11) and the proof is completed. □
Lemma 5.3
For given \(w^{k}\in \mathbb {R}_{++}^{n} \times \mathbb {R}_{++}^{m}\times \mathbb {R}^{l}\), let wk+ 1(γαk) be generated by (2.6) and \(w\in {\mathcal W}\), then we have
Proof
Using the following identity
we get
Substituting (5.16) into (5.15), we obtain
Substituting (5.17) into (5.11), we obtain (5.14), the required result. □
Now, we are ready to present the O(1/t) convergence rate of the proposed method.
Theorem 5.1
For any integer t > 0, we have a \(\hat {w}_{t}\in \mathcal W\) which satisfies
where
Proof
Summing the inequality (5.14) over k = 0,⋅⋅⋅,t, we obtain
Using the notations of Υt and \(\hat {w}_{t}\) in the above inequality, we derive
Indeed, \(\hat {w}_{t}\in \mathcal W\) because it is a convex combination of \(\hat {w}^{0},\hat {w}^{1},\cdot \cdot \cdot , \hat {w}^{t}.\)
The proof is completed. □
If \(\inf ^{\infty }_{k=0}{\upbeta }_{k}=\upbeta >0,\) then from (2.19) we have
Suppose that for any compact set \(\mathcal D\subset \mathcal W,\) let \(d =\sup \{\|w-w^{0}\||w\in \mathcal D\}.\) For any given 𝜖 > 0, after most
iterations, we have
That is, the O(1/t) convergence rate is established in an ergodic sense.
6 Preliminary computational results
In order to verify the theoretical assertions, we consider the following optimization problem with matrix variables:
where ∥⋅∥F is the matrix Fröbenius norm, that is,
It has been shown [9] that problem (6.1) can be converted to the following variational inequality: find \(u^{*}=(X^{*},Y^{*},Z^{*})\in \mathcal W \) \(= {S_{+}^{n} \times S_{+}^{n}\times \mathbb {R}^{n\times n}}\) such that
Problem (6.2) is a special case of (1.2)–(1.3) with matrix variables where A = In×n, B = −In×n, b = 0, f(X) = X − C, g(Y ) = Y − C, and \({\mathcal W} =S_{+}^{n} \times S_{+}^{n}\times \mathbb {R}^{n\times n}\).
For simplification, we take R = rIn×n, S = sIn×n, and H = In×n where r > 0 and s > 0 are scalars. In all tests, we take γ = 1.98, μ = 0.1,τ = 1.5,η = 0.95,ρ = 1.47,σ = 0.95, C = rand(n), and (X0, Y0, Z0) = (In×n, In×n,0n×n) as the initial point in the test. The iteration is stopped as soon as
All codes were written in Matlab; we compare the proposed method with those in [9, 12, 13, 16, 36]. The iteration numbers, denoted by k, and the computational time for problem (6.1) with different dimensions are given in tables 6.1–6.2.
Tables 1 and 2 report the comparison between the methods of [9, 12, 13, 16, 36] and the proposed method. The number of iteration has great advantage and a faster convergence speed.
7 Conclusions
In this paper, we proposed a new descent logarithmic-quadratic proximal alternating direction method for solving structured variational inequalities. The optimal step size along the descent direction is chosen to improve the efficiency of the new method. In addition, we proposed a self-adaptive method that adjusts the scalar parameter automatically. And the numerical efficiency of our algorithm is verified compared with several existing algorithms.
References
Auslender, A., Teboulle, M., Ben-Tiba, S.: A logarithmic-quadratic proximal method for variational inequalities. Comput. Optim. Appl. 12, 31–40 (1999)
Auslender, A., Teboulle, M.: Entropic proximal decomposition methods for convex programs and variational inequalities. Math. Program. 91, 33–47 (2001)
Auslender, A., Teboulle, M.: Interior gradient and epsilon-subgradient descent methods for constrained convex minimization. Math. Oper. Res. 29, 1–26 (2004)
Auslender, A., Teboulle, M.: Interior gradient and proximal methods for convex and conic optimization. SIAM J. Optim. 16, 697–725 (2006)
Auslender, A., Teboulle, M.: Lagrangian duality and related multiplier methods for variational inequality problems. SIAM J. Optim. 10, 1097–1115 (2000)
Bertsekas, D.P., Gafni, E.M.: Projection method for variational inequalities with applications to the traffic assignment problem. Math. Program. Study 17 (1982)
Bertsekas, D.P., Tsitsiklis, J.N.: Parallel and Distributed Computation: Numerical Methods. Athena Scientific, Belmont (1997)
Bnouhachem, A., Benazza, H., Khalfaoui, M.: An inexact alternating direction method for solving a class of structured variational inequalities. Appl. Math. Comput. 219, 7837–7846 (2013)
Bnouhachem, A.: On LQP alternating direction method for solving variational inequality problems with separable structure. J. Inequal. Appl. 2014(80), 1–15 (2014)
Bnouhachem, A., Xu, M.H.: An inexact LQP alternating direction method for solving a class of structured variational inequalities. Comput. Math. Appl. 67, 671–680 (2014)
Bnouhachem, A., Ansari, Q.H.: A descent LQP alternating direction method for solving variational inequality problems with separable structure. Appl. Math. Comput. 246, 519–532 (2014)
Bnouhachem, A., Hamdi, A.: Parallel LQP alternating direction method for solving variational inequality problems with separable structure. J. Inequal. Appl. 2014(392), 1–14 (2014)
Bnouhachem, A., Hamdi, A.: A hybrid LQP alternating direction method for solving variational inequality problems with separable structure. Appl. Math. Inf. Sci. 9(3), 1259–1264 (2015)
Bnouhachem, A., Al-Homidan, S., Ansari, Q.H.: New descent LQP alternating direction methods for solving a class of structured variational inequalities. Fixed Point Theory Appl. 2015(137), 1–11 (2015)
Bnouhachem, A., Latif, A., Ansari, Q.H.: On the O(1/t) convergence rate of the alternating direction method with LQP regularization for solving structured variational inequality problems. J. Inequal. Appl. 2016(297), 1–14 (2016)
Bnouhachem, A., Bensi, F., Hamdi, A.: On alternating direction method for solving variational inequality problems with separable structure. J. Nonlin. Sci. Apps. 10(1), 175–185 (2017)
Bnouhachem, A., Ansari, Q.H., Al-Homidan, S.: SQP Alternating direction for structured vriational inequality. J. Nonlinear Convex Anal. 19(3), 461–476 (2018)
Bnouhachem, A., Rassias, T.M.: A new descent alternating direction method with LQP regularization for the structured variational inequalities. Optim. Lett. 13 (1), 175–192 (2018)
Bnouhachem, A., Rassias, T.M.: On descent alternating direction method with LQP regularization for the structured variational inequalities, Optim. Lett., in press (2019)
Boyd, S., Parikh, N., Chu, E., Peleato, B., Eckstein, J.: Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends Mach. Learn. 3, 1–122 (2010)
Chan, T.F., Glowinski, R.: Finite Element Approximation and Iterative Solution of a Class of Mildly Non-Linear Elliptic Equations. Stanford University, Technical Report (1978)
Chen, Z., Wan, L., Yang, Y.: An inexact alternating direction method for structured variational inequalities. J. Optim. Theory Appl. 163(2), 439–459 (2014)
Facchinei, F., Pang, J.S.: Finite-Dimensional Variational Inequalities and Complementarity Problems, vol. I and II. Springer Series in Operations Research. Springer, New York (2003)
Fukushima, M.: Application of the alternating directions method of multipliers to separable convex programming problems. Comput. Optim. Appl. 1(1), 93–111 (1992)
Glowinski, R., Marrocco, A.: Sur l’approximation par éléments finis d’ordre un et la résolution par pénalisation-dualité d’une classe de problémes de Dirichlet non linéaires. Revue Fr. Autom. Inform. Rech. opér. Anal. Numé,r. 2, 41–76 (1975)
Glowinski, R.: Numerical Methods for Nonlinear Variational Problems. Springer, New York (1984)
Glowinski, R., Tallec, P.L.: Augmented Lagrangian and Operator-Splitting Methods in Nonlinear Mechanics. Society for Industrial and Applied Mathematics, Philadelphia (1989)
He, B.S., Liao, L.Z., Han, D.R., Yang, H.: A new inexact alternating directions method for monotone variational inequalities. Math. Program. 92, 103–118 (2002)
He, B.S., Liao, L.-Z., Yuan, X.M.: A LQP based interior prediction-correction method for nonlinear complementarity problems. J. Comput. Math. 24(1), 33–44 (2006)
He, B.S.: Parallel splitting augmented Lagrangian methods for monotone structured variational inequalities. Comput. Optim. Appl. 42, 195–212 (2009)
He, B.S., Tao, M., Yuan, X.M.: Alternating direction method with Gaussian back substitution for separable convex programming. SIAM J. Optim. 22, 313–340 (2012)
Hou, L.S.: On the O(1/t) convergence rate of the parallel descent-like method and parallel splitting augmented Lagrangian method for solving a class of variational inequalities. Appl. Math. Comput. 219, 5862–5869 (2013)
Jiang, Z.K., Bnouhachem, A.: A projection-based prediction-correction method for structured monotone variational inequalities. Appl. Math Comput. 202, 747–759 (2008)
Jiang, Z.K., Yuan, X.M.: New parallel descent-like method for sloving a class of variational inequalities. J. Optim. TheoryAppl. 145, 311–323 (2010)
Kontogiorgis, S., Meyer, R.R.: A variable-penalty alternating directions method for convex optimization. Math. Program. 83, 29–53 (1998)
Li, M.: A hybrid LQP-based method for structured variational inequalities. Int. J. Comput. Math. 89(10), 1412–1425 (2012)
Tao, M., Yuan, X.M.: On the O(1/t) convergence rate of alternating direction method with Logarithmic-quadratic proximal regularization. SIAM J. Optim. 22(4), 1431–1448 (2012)
Tseng, P.: Applications of a splitting algorithm to decomposition in convex programming and variational inequalities. SIAM J. Con. Optim. 29, 119–138 (1991)
Tseng, P.: Alternating projection-proximal methods for convex programming and variational inequalities. SIAM J. Optim. 7, 951–965 (1997)
Wang, K., Xu, L.L., Han, D.R.: A new parallel splitting descent method for structured variational inequalities. J. IndustrialMan. Optim. 10(2), 461–476 (2014)
Yuan, X.M., Li, M.: An LQP-based decomposition method for solving a class of variational inequalities. SIAM J. Optim. 21(4), 1309–1318 (2011)
Zhang, W., Han, D., Jiang, S.: A modified alternating projection based prediction-correction method for structured variational inequalities. Appl. Numer. Math. 83, 12–21 (2014)
Author information
Authors and Affiliations
Corresponding author
Additional information
This paper is dedicated to Mohamed Bnouhachem and Mohamed Khalfaoui
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Bnouhachem, A. A self-adaptive descent LQP alternating direction method for the structured variational inequalities. Numer Algor 86, 303–324 (2021). https://doi.org/10.1007/s11075-020-00890-0
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11075-020-00890-0
Keywords
- Variational inequalities
- Monotone operator
- Logarithmic-quadratic proximal method
- Projection method
- Alternating direction method