Abstract
In this paper, we study the completely positive (CP) tensor program, which is a linear optimization problem with the cone of CP tensors and some linear constraints. We reformulate it as a linear program over the cone of moments, then construct a hierarchy of semidefinite relaxations for solving it. We also discuss how to find a best CP approximation of a given tensor. Numerical experiments are presented to show the efficiency of the proposed methods.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
For given positive integers m and n, an m-way, n-dimension tensor \({\mathcal {A}}\) is a multidimensional array indexed by integer tuples \((i_1, \ldots , i_m)\) with \(1 \le i_1, \ldots , i_m \le n\), i.e.,
Let \({\mathbb {R}}\) be the set of real numbers and \(\mathrm {T}^m({\mathbb {R}}^{n})\) be the set of all such real-valued tensors. A tensor \({\mathcal {A}} \in \mathrm {T}^m({\mathbb {R}}^{n})\) is called symmetric if each entry \({\mathcal {A}}_{i_1 i_2 \ldots i_m}\) is invariant with respect to all permutations of \((i_1, \ldots , i_m)\). Let \(\mathrm {S}^m({\mathbb {R}}^{n})\) be the subspace of all symmetric tensors in \(\mathrm {T}^m({\mathbb {R}}^n)\). For a vector \(u \in {\mathbb {R}}^n\), denote \(u^{\otimes m}\) the m-way n-dimensional symmetric outer product tensor such that
A tensor in \(\mathrm {S}^m({\mathbb {R}}^{n})\) is called rank-1 if it has the form \(\lambda u^{\otimes m}\) where \(\lambda \in {\mathbb {R}}\) and \(u\in {\mathbb {R}}^n\). Clearly, if m is odd or \(\lambda > 0\), then the mth real root of \(\lambda \) always exists, so we can rewrite the tensor as
For the other case, the scalar \(\lambda \) cannot be absorbed into the vector v. Comon et al. [5] showed that every symmetric tensor \({\mathcal {A}} \in \mathrm {S}^m({\mathbb {R}}^{n})\) can be decomposed as
where \(\lambda _k\in {\mathbb {R}}\) and \(u^k \in {\mathbb {R}}^n\) for \(k = 1,\ldots , r\). Interested readers are referred to [4, 6, 14, 16,17,18, 25, 26] for numerical methods for computing real eigenvalues of symmetric tensors, tensor decompositions, as well as applications.
Let \({\mathbb {R}}^n_+\) be the nonnegative orthant of \({\mathbb {R}}^n\). A symmetric tensor \({\mathcal {A}}\in \mathrm {S}^m({\mathbb {R}}^{n})\) is completely positive (CP), if there exist vectors \(v^1,\ldots ,v^r\in {\mathbb {R}}^n_+\) such that
where r is called the length of the decomposition (1.2) (cf. [30, 31]). The smallest r in the above is called the cp-rank of \({\mathcal {A}}\). If \({\mathcal {A}}\) is CP, (1.2) is called a CP decomposition (or nonnegative decomposition) of \({\mathcal {A}}\). The CP tensor is a natural extension of the CP matrix. We refer to [1, 2, 21, 35, 36] for work on CP matrices and tensors.
For \({\mathcal {B}}\in \mathrm {S}^m({\mathbb {R}}^{n})\), we define
If \({\mathcal {B}}x^m \ge 0\) for all \(x \in {\mathbb {R}}^n_+\), we call \({\mathcal {B}}\) a copositive tensor (cf. [31]). The copositive tensor is an extension of the copositive matrix. Obviously, both symmetric nonnegative tensors and positive semidefinite tensors are copositive tensors.
For \({\mathcal {A}}, {\mathcal {B}}\in \mathrm {S}^m({\mathbb {R}}^{n})\), the inner product of \({\mathcal {A}}\) and \({\mathcal {B}}\) is defined as
and the norm of a tensor \({\mathcal {A}}\) is the square root of the sum of squares of its entries, i.e.,
For a cone \( C\subseteq \mathrm {S}^m({\mathbb {R}}^{n})\), the dual cone of C is defined as
Denote by \(CP_{m,n}\) and \(COP_{m,n}\) the sets of m-way, n-dimension CP tensors and copositive tensors, respectively. Both \(CP_{m,n}\) and \(COP_{m,n}\) are proper cones, and they are dual to each other [30].
CP tensor and nonnegative decomposition have wide applications in statistics, computer vision, exploratory multiway data analysis and blind source separation [3, 32, 33]. Generally, it is hard to check whether a tensor is CP or not. The problem is NP-hard even for the CP matrix case [8]. Kolda [18] assumed the length of the CP decomposition is known and formulated the CP tensor decomposition problem as a nonnegative constrained least squares problem, then solved it by a nonlinear optimization software. Qi et al. [30] showed that a strongly symmetric hierarchically dominated nonnegative tensor is CP and presented a hierarchical elimination algorithm for checking this. For a general symmetric tensor, we transformed the problem of checking whether it is CP to a truncated moment problem, then proposed a semidefinte algorithm for solving it [9].
Recently, Penã et. al [28] showed that under certain conditions a general polynomial optimization problem (POP), which are not necessarily quadratic, can be formulated as a linear optimization problem with the cone of CP tensors and some linear equality constraints. In [15], Kuang and Zuluaga gave the CP tensor relaxation for the POP and showed that it provides a better bound than the Lagrangian relaxation. They also proved that the CP tensor relaxation yields a better bound than the CP matrix relaxation for quadratic reformulations of some class of polynomial optimization problems. Since the checking of CP tensors is NP-hard, it is difficult to deal with the cone of CP tensors directly. A general way is to approximate it by some tractable cones, such as the cones of nonnegative tensors, the positive semidefinite tenors and the doubly nonnegative tensors [15]. However, this approach generally only provides a bound for the original problem and gives an approximate solution of the problem.
The aim of this paper is to develop a new method for the CP tensor program, which is stated as:
where \({\mathcal {A}}, {\mathcal {A}}_i\in \mathrm {S}^m({\mathbb {R}}^{n}), b_i \in {\mathbb {R}}, i=1,\ldots ,l\). We formulate it as a linear optimization problem with the cone of moments, then construct a hierarchy of semidefinite relaxations for solving it. If (1.3) is infeasible, we can get a certificate for that. If it is feasible, a CP decomposition of the optimal solution can also be obtained. We also discuss the best CP tensor approximation problem, which is to find a tensor in the intersection of a set of linear constraints and the cone of CP tensors such that it is close to a given tensor as much as possible. It is an extension of the CP-matrix approximation problem [10]. We transform the problem to a conic linear program over the cone of moments and the second-order cone. A hierarchical semidefinite relaxation algorithm is also presented for it.
The paper is organized as follows. In Sect. 2, we characterize the CP tensor as a moment sequence. In Sect. 3, we formulate the CP tensor program as a linear optimization problem with the cone of moments, then propose a smidefinite algorithm for solving it. The convergence properties of the algorithm are also studied. In Sect. 4, we show how to solve the best CP tensor approximation problem. Some computational experiments are presented in Sect. 5.
2 Moments and outer approximations
In this section, we characterize a CP tensor as a truncated moment sequence, then give some necessary and sufficient conditions for a symmetric tensor to be CP. A hierarchy of outer approximations of the cone of CP tensors is also introduced.
2.1 Moments and flat extension
As we know, a symmetric matrix can be identified by a vector that consists of its upper triangular entries. Similarly, a symmetric tensor \({\mathcal {A}}\in \mathrm {S}^m({\mathbb {R}}^{n})\) can also be identified by a vector \({\text {vech}}({\mathcal {A}})\) that consists of its upper triangular entries, i.e. the entries
Let \({\mathbb {N}}\) be the set of nonnegative integers. For \(\alpha = (\alpha _1,\ldots , \alpha _n) \in {\mathbb {N}}^n\), denote \(|\alpha | := \alpha _1+\cdots +\alpha _n\). Let
Then, there is a one-to-one correspondence between the index pair \((i_1,\ldots i_m)\) and the vector
where \(e_i\) is the i-th unit vector in \({\mathbb {R}}^n\). Hence, the identifying vector of \({\mathcal {A}}\in \mathrm {S}^m({\mathbb {R}}^{n})\) can also be written as
where \({\mathbb {R}}^{{\mathbb {T}}}\) denotes the space of real vectors indexed by \(\alpha \in {\mathbb {T}}\). We call \({\mathbf {a}}\) a \({\mathbb {T}}\)-truncated moment sequence (\({\mathbb {T}}\)-tms).
Let
Note that every nonnegative vector is a multiple of a vector in K. By (1.2), \({\mathcal {A}}\in CP_{m,n}\) if and only if there exist \(\lambda _1,\ldots ,\lambda _r >0\) and \(u^1,\ldots ,u^r \in K\) such that
We say a \({\mathbb {T}}\)-tms \({\mathbf {a}}\) admits a K-measure \(\mu \), i.e., a nonnegative Borel measure \(\mu \) supported in K such that
where \(x^{\alpha } := x^{\alpha _1}_1 \cdots x^{\alpha _n}_n\). A measure \(\mu \) satisfying the above is called a K-representing measure for \({\mathbf {a}}\). A measure is called finitely atomic if its support is a finite set, and is called r-atomic if its support consists of at most r distinct points. We refer to [22] for representing measures of truncated moment sequences.
Hence, by (2.3), a tensor \({\mathcal {A}}\in \mathrm {S}^m({\mathbb {R}}^{n})\), with the identifying vector \({\mathbf {a}} \in {\mathbb {R}}^{{\mathbb {T}}}\), is CP if and only if \({\mathbf {a}}\) admits an r-atomic K-measure, i.e.,
where each \(\lambda _i>0\), \(u^i \in K\) and
Denote
Then, \({\mathcal {M}}_{{\mathbb {T}},K}\) is the CP tensor cone. So,
Denote
For a polynomial \(p\in {\mathbb {R}}[x]_{{\mathbb {T}}}\), \(p|_K\ge 0\) (resp., \(p|_K> 0\)) denotes \(p\ge 0\) (resp., \(p> 0\)) on K. Then the dual cone of \({\mathcal {M}}_{{\mathbb {T}},K}\) is
which is the copositive tensor cone \(COP_{m,n}\) (cf. [20, 23]). We say \({\mathbb {R}}[x]_{{\mathbb {T}}}\) is K-full if there exists a polynomial \(p \in {\mathbb {R}}[x]_{{\mathbb {T}}}\) such that \(p|_{K} >0\) (cf. [11]). It is easy to check that \({\mathbb {R}}[x]_{{\mathbb {T}}}\) is K-full.
For a \({\mathbb {T}}\)-tms \({\mathbf {a}} \in {\mathbb {R}}^{{\mathbb {T}}}\), define a linear Riesz functional \({\mathcal {L}}_{{\mathbf {a}}}\) acting on \({\mathbb {R}}[x]_{{\mathbb {T}}}\) as
We also denote \(\langle p,{\mathbf {a}} \rangle := {\mathcal {L}}_{{\mathbf {a}}}(p)\) for convenience. Let
and
For \(s \in {\mathbb {R}}^{{\mathbb {N}}^n_{2k}}\) and \(q \in {\mathbb {R}}[x]_{2k}\), the k-th localizing matrix of q generated by s is the symmetric matrix \(L^{(k)}_q (s)\) satisfying
where \({\text {vec}}(p)\) denotes the coefficient vector of p in the graded lexicographical ordering, and \(\lceil t\rceil \) denotes the smallest integer that is not smaller than t. In particular, when \(q=1\), \(L^{(k)}_1 (s)\) is called a k-th order moment matrix and denoted as \(M_k(s)\).
In fact, we have
and
We refer to [19, 20, 22] for more details about localizing matrices and moment matrices.
Denote the polynomials:
Note that K given in (2.2) is nonempty compact. It can also be described equivalently as
where \(g(x)=(g_0(x), g_1(x),\ldots ,g_{n}(x))\). As shown in Nie [22], for \(s \in {\mathbb {R}}^{{\mathbb {N}}^n_{2k}}\), if
then s admits a unique K-measure, which is \(\text {rank} M_k(s)\)-atomic (cf. Curto and Fialkow [7]). We say s is flat if both (2.11) and (2.12) are satisfied.
Given two tms’ \(y \in {\mathbb {R}}^{{\mathbb {N}}^n_{d}}\) and \(z \in {\mathbb {R}}^{{\mathbb {N}}^n_{e}}\), we say z is an extension of y, if \(d\le e\) and \(y_{\alpha } = z_{\alpha }\) for all \(\alpha \in {\mathbb {N}}^n_{d}\). We denote by \(z|_{{\mathbb {T}}}\) the subvector of z, whose entries are indexed by \(\alpha \in {\mathbb {T}}\). For convenience, we denote by \(z|_{d}\) the subvector \(z |_{{\mathbb {N}}^n_{d}}\). If z is flat and extends y, we say z is a flat extension of y. It is shown in [22] that a \({\mathbb {T}}\)-tms \({\mathbf {a}} \in {\mathbb {R}}^{{\mathbb {T}}}\) admits a K-measure if and only if it is extendable to a flat tms \(z \in {\mathbb {R}}^{{\mathbb {N}}^n_{2k}}\) for some k. So, by (2.6),
2.2 Outer approximations
We call a subset \(I \subseteq {\mathbb {R}}[x]\) an ideal if \(I + I \subseteq I\) and \(I \cdot {\mathbb {R}}[x]\subseteq I\). For a tuple \(\chi = (\chi _1, \ldots , \chi _l)\) of polynomials in \({\mathbb {R}}[x]\), denoted by \(I(\chi )\) the ideal generated by \(\chi _1,\ldots , \chi _l\). The smallest ideal containing all \(\chi _i\) is the set \(\chi _1 {\mathbb {R}}[x] + \cdots + \chi _l {\mathbb {R}}[x]\). A polynomial \(f \in {\mathbb {R}}[x]\) is called a sum of squares (SOS) if there exist \(f_1, \ldots , f_t \in {\mathbb {R}}[x]\) such that \(f = f^2_1 + \cdots + f^2_t\).
Let h and g be as in (2.10). Denote
and
Then, \(I(h)=\bigcup _{k\in {\mathbb {N}}} I_{2k}(h)\) is the ideal generated by h, and \(Q(g)= \bigcup _{k\in {\mathbb {N}}} Q_k (g)\) is the quadratic module generated by g (cf. [23]). We say \(I(h)+ Q(g)\) is archimedean if there exists a number \(N> 0\) such that \(N- \Vert x\Vert ^2 \in I(h)+Q(g)\). Clearly, if \(f\in I(h) + Q(g)\), then \(f|_K\ge 0\). Conversely, if \(f|_K > 0\) and \(I(h) + Q(g)\) is archimedean, then \(f\in I(h) + Q(g)\). This is due to Putinar’s Positivstellensatz (cf. [29]).
For each \(k \in {\mathbb {N}}\) and \(k\ge \lceil m/2\rceil \), denote
Recall that \({\mathbb {R}}[x]_{{\mathbb {T}}}\) is K-full, \({\mathbb {T}}\) is finite and \(I(h)+ Q(g)\) is archimedean because \(1- \Vert x\Vert ^2 = -h(x) \in I(h)+Q(g)\). By Nie [23, Propositions 3.5], we have
Moreover,
where \(\text {cl}(\cdot )\) denotes the closure of a set.
Correspondingly, for each \(k \in {\mathbb {N}}\) and \(k\ge \lceil m/2\rceil \), denote
and
Since \({\mathbb {T}}\) is finite, \({\mathbb {R}}[x]_{{\mathbb {T}}}\) is K-full and \(I(h)+ Q(g)\) is archimedean, by Nie [23, Proposition 3.3], we have
and
Moreover, \(\Phi _k\) and \(\Omega _{k}\) are dual to each other.
As shown above, the hierarchy of \(\Phi _k\) provides inner approximations of \({\mathcal {P}}_{{\mathbb {T}},K}\) and converges monotonically and asymptotically to \({\mathcal {P}}_{{\mathbb {T}},K}\), while the hierarchy of \(\Omega _k\) provides outer approximations of \({\mathcal {M}}_{{\mathbb {T}},K}\) and converges monotonically and asymptotically to \({\mathcal {M}}_{{\mathbb {T}},K}\). So, \(\Phi _k\) and \(\Omega _k\) can approximate \({\mathcal {P}}_{{\mathbb {T}},K}\) and \({\mathcal {M}}_{{\mathbb {T}},K}\), arbitrarily well, respectively.
3 The CP tensor program
In this section, we reformulate the CP tensor program as a linear optimization problem with the cone of moments, then construct a hierarchy of semidefinite relaxations for solving it. The convergence properties of the semidefinite algorithm are also discussed.
3.1 Reformulation and semidefinite relaxations
Denote by
the identifying vector of \({\mathcal {X}}\). Denote by \({\mathbf {c}}\in {\mathbb {R}}^{\mathbb {T}}\) a constant vector with the element \({\mathbf {c}}_\alpha =\sqrt{\frac{m!}{\alpha _1 ! \cdots \alpha _n !}}\) for each \(\alpha \in {\mathbb {T}}\). For \({\mathbf {x}},{\mathbf {y}}\in {\mathbb {R}}^{\mathbb {T}}\), \({\mathbf {x}}\circ {\mathbf {y}}\) denotes their Hadamard product. For convenience, we denote \({\mathbf {c}} \circ {\mathbf {c}}\) by \({\mathbf {c}}^2\). Then, (1.3) can be formulated as the linear optimization problem:
where \({\mathcal {M}}_{{\mathbb {T}},K}\) is given by (2.5). The dual problem of (\(P_1\)) is
where \({\mathcal {P}}_{{\mathbb {T}},K}\) is given by (2.7). Denote by \({\mathcal {F}}(P)\) the feasible set of the problem (P). By weak duality, for all feasible points \({\mathbf {x}} \in {\mathcal {F}}(P_1)\) and \((\lambda ,{\mathbf {s}}) \in {\mathcal {F}}(D_1)\), we have \(\vartheta _{P_1}\ge \vartheta _{D_1}\).
By (2.19) and (2.20), for each \(k \in {\mathbb {N}}\) and \(k\ge \lceil m/2\rceil \), the k-th order relaxation of (\(P_1\)) can be defined as
The dual problem of \((P_1^k)\) is
Clearly, \(\vartheta _{P_1}^k\le \vartheta _{P_1}\) and \(\vartheta _{D_1}^k\le \vartheta _{D_1}\) for all k. Suppose that \(({\mathbf {x}}^{*,k}, \tilde{{\mathbf {x}}}^{*,k})\) is a minimizer of (\(P^k_1\)) and \((\lambda ^{*,k},{\mathbf {s}}^{*,k})\) is a maximizer of (\(D^k_1\)). If \({\mathbf {x}}^{*,k}=\tilde{{\mathbf {x}}}^{*,k}|_{{\mathbb {T}}}\in {\mathcal {M}}_{{\mathbb {T}},K}\), then \(\vartheta _{P_1}^k=\vartheta _{P_1}\) and \({\mathbf {x}}^{*,k}\) is a minimizer of (\(P_1\)), i.e., the relaxation \((P^k_1)\) is exact for solving (\(P_1\)). In this case, if \(\vartheta _{P_1}^k=\vartheta _{D_1}^k\), then \(\vartheta _{D_1}^k=\vartheta _{D_1}\) and \((\lambda ^{*,k},{\mathbf {s}}^{*,k})\) is a maximizer of (\(D_1\)). If the relaxation (\(P^k_1\)) is infeasible, then (\(P_1\)) is also infeasible.
3.2 A semidefinite algorithm
Based on the above, we propose a semidefinite algorithm for solving the CP tensor program (1.3).
Algorithm 3.1
-
Step 0. Input \({\mathcal {A}}, {\mathcal {A}}_i\in \mathrm {S}^m({\mathbb {R}}^{n}), b_i \in {\mathbb {R}}, i=1,\ldots ,l\) and K as (2.2). Let \(k := \lceil m/2\rceil \).
-
Step 1. Solve the relaxation problem (\(P^k_1\)). If (\(P^k_1\)) is infeasible, stop and output that (\(P_1\)) is infeasible; otherwise, compute an optimal solution \(({\mathbf {x}}^{*,k}, \tilde{{\mathbf {x}}}^{*,k})\) of (\(P^k_1\)). Let \(t :=1\).
-
Step 2. Let \(\hat{{\mathbf {x}}} :=\tilde{{\mathbf {x}}}^{*,k}|_{2t}\). If the rank condition (2.12) is not satisfied, go to Step 4.
-
Step 3. Compute \(\rho _i > 0\), \(v^i \in K\) and \(r = \text {rank} (M_t (\hat{{\mathbf {x}}}))\). Output the CP decomposition of \({\mathcal {X}}^*\) as
$$\begin{aligned} {\mathcal {X}}^* =\sum ^{r}_{i=1} \rho _i (v^i)^{\otimes m}. \end{aligned}$$ -
Step 4. If \(t < k\), set \(t := t+1\) and go to Step 2; otherwise, set \(k := k+1\) and go to Step 1.
We use Step 2 to check whether \(\tilde{{\mathbf {x}}}^{*,k}|_{2t}\) is flat or not. We use Henrion and Lasserre’s method in [12] to get an r-atomic K-measure for \(\hat{{\mathbf {x}}}\), which can further produce the CP decomposition of the optimal solution.
We say (\(D_1\)) has a relative interior, if there exists a feasible point pair \((\lambda ,{\mathbf {s}}) \in {\mathcal {F}}(D_1)\) such that \({\mathbf {s}}\in {\text {int}}({\mathcal {P}}_{{\mathbb {T}},K})\) and \(\lambda _i> 0, \ i=l_1+1,\ldots ,l\). The asymptotic convergence of Algorithm 3.1 is given as below.
Theorem 3.2
Let \({\mathbb {T}}\) and K be as in (2.1) and (2.10), respectively. Suppose (\(P_1\)) is feasible and (\(D_1\)) has a relative interior point. Algorithm 3.1 has the following properties:
-
(i)
For all k sufficiently large, (\(D^k_1\)) has a relative interior point and (\(P^k_1\)) has a minimizer \(({\mathbf {x}}^{*,k},\tilde{{\mathbf {x}}}^{*,k})\).
-
(ii)
The sequence \(\{{\mathbf {x}}^{*,k}\}\) is bounded, and each of its accumulation points is a minimizer of (\(P_1\)). The sequence \(\{\vartheta _{P_1}^k\}\) converges to the minimum of (\(P_1\)).
Proof
-
(i)
Let \((\lambda ^0,{\mathbf {s}}^0) \in {\mathcal {F}}(D_1)\) with \({\mathbf {s}}^0 \in {\text {int}}({\mathcal {P}}_{{\mathbb {T}},K})\) and \(\lambda ^0_i> 0(i=l_1+1,\ldots , l)\). Then, \({\mathbf {s}}^0|_K>0\) (cf. [23, Lemma 3.1]). Note that since K is compact, there exist \( \epsilon > 0\) and \( \delta > 0\) such that
$$\begin{aligned} {\mathbf {s}}|_K - \epsilon > \epsilon , \quad \forall \,{\mathbf {s}} \in B({\mathbf {s}}^0, \delta ). \end{aligned}$$By Nie [27, Theorem 6], there exists \(N_0 > 0\) such that
$$\begin{aligned} {\mathbf {s}} - \epsilon \in I_{2N_0} (h) + Q_{N_0}(g), \quad \forall \, {\mathbf {s}} \in B({\mathbf {s}}^0, \delta ). \end{aligned}$$So, (\(D^k_1\)) has a relative interior point for all \(k \ge N_0\). Thus, the strong duality holds for (\(P^k_1\)) and (\(D^k_1\)). As (\(P_1\)) is feasible, the relaxation problem (\(P^k_1\)) is also feasible.
So, (\(P^k_1\)) has a minimizer \(({\mathbf {x}}^{*,k}, \tilde{{\mathbf {x}}}^{*,k})\).
-
(ii)
We first show that \(\{{\mathbf {x}}^{*,k}\}\) is bounded. Let \((\lambda ^0, {\mathbf {s}}^0)\) and \(\epsilon \) be as in the proof of (i). The set \(I_{2N_0}(h) + Q_{N_0}(g)\) is dual to \(\Gamma _{N_0}\). For all \(k \ge N_0\), we have \(\tilde{{\mathbf {x}}}^{*,k} \in \Gamma _{N_0}\) and
$$\begin{aligned}&0 \le \langle {\mathbf {s}}^0 - \epsilon ,\tilde{{\mathbf {x}}}^{*,k}\rangle = \langle {\mathbf {s}}^0,\tilde{{\mathbf {x}}}^{*,k}\rangle - \epsilon \langle 1,\tilde{{\mathbf {x}}}^{*,k} \rangle , \\&\langle {\mathbf {s}}^0,{\mathbf {x}}^{*,k}\rangle \le \vartheta _{P_1}^k-b^T\lambda ^0. \end{aligned}$$
Since \(\vartheta _{P_1}^k\le \vartheta _{P_1}\) and \(\langle {\mathbf {s}}^0,\tilde{{\mathbf {x}}}^{*,k}\rangle =\langle {\mathbf {s}}^0, {\mathbf {x}}^{*,k}\rangle \), it holds that
So,
Note that \(I(h) + Q(g)\) is archimedean, following the line of the proof of [23, Theorem 4.3 (ii)], we can obtain that the sequence \(\{{\mathbf {x}}^{*,k}\}\) is bounded.
Suppose \({\mathbf {x}}^{*}\) is an accumulation point of \(\{{\mathbf {x}}^{*,k}\}\). Without loss of generality, we assume
Since \({\mathbf {x}}^{*,k}\in \Omega _{k} \), by (2.21) and (2.22), we have \({\mathbf {x}}^*\in \bigcap _{k=\lceil m/2\rceil }^{\infty } \Omega _k= {\mathcal {M}}_{{\mathbb {T}},K}\). Note that \({\mathbf {x}}^{*,k}\in {\mathcal {F}}(P^k_1)\), we further have \({\mathbf {x}}^{*}\in {\mathcal {F}}(P_1)\). Denote by \(\vartheta _{P_1}^*\) the objective function value of (\(P_1\)) at \({\mathbf {x}}^*\). Then,
Since (\(P^k_1\)) is a relaxation problem of (\(P_1\)) and \({\mathbf {x}}^{*,k}\) is a minimizer of (\(P^k_1\)), we have
Taking \(k\rightarrow \infty \), we get
This, together with (3.1), implies that
So, \({\mathbf {x}}^{*}\) is a minimizer of (\(P_1\)), and the sequence \(\{\vartheta _{P_1}^k\}\) converges to the minimum of (\(P_1\)). \(\square \)
Remark 3.3
By Theorem 3.2, we know that Algorithm 3.1 can solve the CP tensor program (1.3). If it is feasible, a CP decomposition of the optimal solution can also be obtained. In particularly, under some general conditions, Algorithm 3.1 converges within finitely many steps (cf. [23, 24]).
4 The best CP tensor approximation problem
In this section, we study the best CP tensor approximation problem, which approximates a given symmetric tensor by a CP tensor under linear constraints. We show that the problem can be transformed to a linear program over the cone of moments and the second order cone. A semidefinite algorithm is also presented for it.
Consider the best CP tensor approximation problem, stated as:
where \({\mathcal {A}}, {\mathcal {A}}_i\in \mathrm {S}^m({\mathbb {R}}^{n}), b_i \in {\mathbb {R}}, i=1,\ldots ,l\). Compared to the CP tensor program (1.3), the objective function of (4.1) is nonlinear on the variable \({\mathcal {X}}\).
If there are no linear constraints, (4.1) is reduced to the CP tensor projection problem:
Clearly, (4.2) is always feasible and has a solution. If the minimum is zero, then the projection of \({\mathcal {A}}\) onto \(CP_{m,n}\) is itself, which implies that \({\mathcal {A}}\) is CP. If the minimum is nonzero, then \({\mathcal {A}}\) is not CP. So, solving the CP tensor projection problem also provides a way to check whether a tensor is CP or not.
In the following, we show how to solve (4.1). Introducing a nonnegative variable \(\gamma \in {\mathbb {R}}\), we transform (4.1) to the following problem:
Let \({\mathcal {Y}}={\mathcal {X}}-{\mathcal {A}}\). Denote by
the identifying vectors of \({\mathcal {X}}\) and \({\mathcal {Y}}\), respectively. It is easy to check that \(\Vert {\mathcal {X}}-{\mathcal {A}}\Vert \le \gamma \) if and only if \(({\mathbf {c}}\circ {\mathbf {y}},\gamma )\in \mathcal {SOC}_{\mathbb {T}},\) where
is the second-order cone and is self-dual. Then, (4.3) can be formulated as the following conic linear program:
where \({\mathcal {M}}_{{\mathbb {T}},K}\) is given by (2.5). The dual problem of (\(P_2\)) is
where \({\mathcal {P}}_{{\mathbb {T}},K}\) is given by (2.7). By weak duality, for all \(({\mathbf {x}}, {\mathbf {y}},\gamma ) \in {\mathcal {F}}(P_2)\) and \((\lambda , {\mathbf {s}},{\mathbf {z}}) \in {\mathcal {F}}(D_2)\), we have \(\vartheta _{P_2}\ge \vartheta _{D_2}\).
By (2.19) and (2.20), for each \(k\ge \lceil m/2\rceil \), the k-th order relaxation of (\(P_2\)) can be defined as
The dual problem of \((P^k_2)\) is
Clearly, \(\vartheta _{P_2}^k\le \vartheta _{P_2}\) and \(\vartheta _{D_2}^k\le \vartheta _{D_2}\) for all k. Suppose that \(({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k},\tilde{{\mathbf {x}}}^{*,k})\) is a minimizer of (\(P^k_2\)) and \((\lambda ^{*,k}, {\mathbf {s}}^{*,k},{\mathbf {z}}^{*,k})\) is a maximizer of (\(D^k_2\)). If \({\mathbf {x}}^{*,k}=\tilde{{\mathbf {x}}}^{*,k}|_{\mathbb {T}}\in {\mathcal {M}}_{{\mathbb {T}},K}\), then \(\vartheta _{P_2}^k=\vartheta _{P_2}\) and \(({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k})\) is a minimizer of (\(P_2\)), i.e., the relaxation \((P^k_2)\) is exact for solving (\(P_2\)). In this case, if \(\vartheta _{P_2}^k=\vartheta _{D_2}^k\), then \(\vartheta _{D_2}^k=\vartheta _{D_2}\) and \((\lambda ^{*,k},{\mathbf {s}}^{*,k},{\mathbf {z}}^{*,k})\) is a maximizer of (\(D_2\)). If the relaxation (\(P^k_2\)) is infeasible, then (\(P_2\)) is also infeasible.
A semidefinite algorithm for solving the best CP tensor approximation problem (4.1) is presented as follows.
Algorithm 4.1
-
Step 0. Input \({\mathcal {A}}, {\mathcal {A}}_i\in \mathrm {S}^m({\mathbb {R}}^{n}), b_i \in {\mathbb {R}}, i=1,\ldots ,l\) and K as (2.2). Let \(k := \lceil m/2\rceil \).
-
Step 1. Solve the relaxation problem (\(P^k_2\)). If (\(P^k_2\)) is infeasible, stop and output that (\(P_2\)) is infeasible; otherwise, compute an optimal solution \(({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k},\tilde{{\mathbf {x}}}^{*,k})\) of (\(P^k_2\)). Let \(t :=1\).
-
Step 2. Let \(\hat{{\mathbf {x}}} :=\tilde{{\mathbf {x}}}^{*,k}|_{2t}\). If the rank condition (2.12) is not satisfied, go to Step 4.
-
Step 3. Compute \(\rho _i > 0\), \(v^i \in K\) and \(r = \text {rank} (M_t (\hat{{\mathbf {x}}}))\). Output the CP decomposition of \({\mathcal {X}}^*\) as
$$\begin{aligned} {\mathcal {X}}^* =\sum ^{r}_{i=1} \rho _i (v^i)^{\otimes m}. \end{aligned}$$ -
Step 4. If \(t < k\), set \(t := t+1\) and go to Step 2; otherwise, set \(k := k+1\) and go to Step 1.
We say (\(D_2\)) has a relative interior, if there exists a feasible point pair \((\lambda ,{\mathbf {s}},{\mathbf {z}}) \in {\mathcal {F}}(D_2)\) such that \(({\mathbf {s}},({\mathbf {c}}\circ {\mathbf {z}},1)) \in {\text {int}}({\mathcal {P}}_{{\mathbb {T}},K} \times \mathcal {SOC})\) and \(\lambda _i> 0, \ i=l_1+1,\ldots ,l\). The convergence properties of Algorithm 4.1 is similar to those of Algorithm 3.1.
Theorem 4.2
Let \({\mathbb {T}}\) and K be as in (2.1) and (2.10), respectively. Suppose (\(P_2\)) is feasible and (\(D_2\)) has a relative interior point. Algorithm 4.1 has the following properties:
-
(i)
For all k sufficiently large, (\(D^k_2\)) has a relative interior point and (\(P^k_2\)) has a minimizer \(({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k},\tilde{{\mathbf {x}}}^{*,k})\).
-
(ii)
The sequence \(\{({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k})\}\) is bounded, and each of its accumulation points is a minimizer of (\(P_2\)). The sequence \(\{\gamma ^{*,k}\}\) converges to the minimum of (4.1).
Proof
-
(i)
Let \((\lambda ^0, {\mathbf {s}}^0,{\mathbf {z}}^0)\in {\mathcal {F}}(D_2)\) with \(\lambda ^0_i> 0(i=l_1+1,\ldots , l)\), and \(({\mathbf {s}}^0,({\mathbf {c}}\circ {\mathbf {z}}^0,1)) \in {\text {int}}({\mathcal {P}}_{{\mathbb {T}},K} \times \mathcal {SOC})\). Then, \({\mathbf {s}}^0|_K>0\). Since K is compact, there exist \( \epsilon > 0\) and \( \delta > 0\) such that
$$\begin{aligned} {\mathbf {s}}|_K - \epsilon > \epsilon , \quad \forall \,{\mathbf {s}} \in B({\mathbf {s}}^0, \delta ). \end{aligned}$$By [27, Theorem 6], there exists \(N_0 > 0\) such that
$$\begin{aligned} {\mathbf {s}} - \epsilon \in I_{2N_0} (h) + Q_{N_0}(g), \quad \forall \, {\mathbf {s}} \in B({\mathbf {s}}^0, \delta ). \end{aligned}$$So, (\(D^k_2\)) has a relative interior point for all \(k \ge N_0\). Thus, the strong duality holds for (\(P^k_2\)) and (\(D^k_2\)). As (\(P_2\)) is feasible, the relaxation problem (\(P^k_2\)) is also feasible. So, (\(P^k_2\)) has a minimizer \(({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k},\tilde{{\mathbf {x}}}^{*,k})\).
-
(ii)
We first show \(\{({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k})\}\) is bounded. Let \((\lambda ^0, {\mathbf {s}}^0,{\mathbf {z}}^0)\) and \(\epsilon \) be as in the proof of (i). The set \(I_{2N_0}(h) + Q_{N_0}(g)\) is dual to \(\Gamma _{N_0}\). For all \(k \ge N_0\), we have \(\tilde{{\mathbf {x}}}^{*,k} \in \Gamma _{N_0}\) and
$$\begin{aligned}&0 \le \langle {\mathbf {s}}^0 - \epsilon ,\tilde{{\mathbf {x}}}^{*,k}\rangle = \langle {\mathbf {s}}^0,\tilde{{\mathbf {x}}}^{*,k}\rangle - \epsilon \langle 1,\tilde{{\mathbf {x}}}^{*,k} \rangle , \\&\langle ({\mathbf {s}}^0,({\mathbf {c}}\circ {\mathbf {z}}^0,1)),({\mathbf {x}}^{*,k}, ({\mathbf {c}}\circ {\mathbf {y}}^{*,k},\gamma ^{*,k}))\rangle \le \gamma ^{*,k}-b^T\lambda ^0 - ({\mathbf {c}}^2\circ {\mathbf {z}}^0)^T {\text {vech}}({\mathcal {A}}). \end{aligned}$$
Because \(\gamma ^{*,k}\le \vartheta _{P_2}\) and
we have
So,
Note that \(I(h) + Q(g)\) is archimedean, we can further prove that the sequence \(\{{\mathbf {x}}^{*,k}\}\) is bounded. By the definitions of \({\mathbf {x}}, {\mathbf {y}}\) and \(\gamma \), we know \(\{({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k})\}\) is bounded.
Suppose \(({\mathbf {x}}^{*}, {\mathbf {y}}^{*},\gamma ^{*})\) is an accumulation point of \(\{({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k})\}\). Without loss of generality, we assume
Since \({\mathbf {x}}^{*,k}\in \Omega _{k} \), by (2.21) and (2.22), we have \({\mathbf {x}}^*\in \bigcap _{k=\lceil m/2\rceil }^{\infty } \Omega _k= {\mathcal {M}}_{{\mathbb {T}},K}\). Note that \(({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k})\in {\mathcal {F}}(P^k_2)\), we further have \(({\mathbf {x}}^{*}, {\mathbf {y}}^{*},\gamma ^{*})\in {\mathcal {F}}(P_2)\). So,
Because (\(P^k_2\)) is a relaxation problem of (\(P_2\)) and \(({\mathbf {x}}^{*,k}, {\mathbf {y}}^{*,k},\gamma ^{*,k})\) is a minimizer of (\(P^k_2\)), we get
Taking \(k\rightarrow \infty \), we obtain
This, together with (4.4), implies that
So, \(({\mathbf {x}}^{*}, {\mathbf {y}}^{*},\gamma ^{*})\) is a minimizer of (\(P_2\)), and the sequence \(\{\gamma ^{*,k}\}\) converges to the minimum of (\(P_2\)). \(\square \)
Remark 4.3
It is worth to point out that, though the objective function of (4.1) is nonlinear, we can transform it to a linear optimization problem with the cone of moments and the second order cone. So, it is essentially a conic linear program.
If (\(P_2\)) is feasible, Algorithm 4.1 generally converges within finitely many steps (cf. [23, 24]).
5 Numerical experiments
In this section, we present some numerical experiments for solving the CP tensor program (1.3) and the best CP tensor approximation problem (4.1) by Algorithms 3.1 and 4.1, respectively. The CP decompositions of the optimal solutions are also given. We use the softwares GloptiPoly 3 [13] and SeDuMi [34] to solve the semidefinite relaxation problems (\(P^k_1\)) and (\(P^k_2\)). The experiments are implemented on a laptop with an Intel Core i5-2520M CPU (2.50 GHz) and 8 GB of RAM, using Matlab R2014b. We display 4 decimal digits for numerical numbers.
5.1 CP tensor programs
We use Algorithm 3.1 to solve the CP tensor programs (1.3).
Example 5.1
Consider the tensor \({\mathcal {A}} \in \mathrm {S}^3({\mathbb {R}}^{10})\) given as:
Case 1. Consider (1.3) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where
We apply Algorithm 3.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=2\) with \(\vartheta _{P_1}^{k}=-12.5084\). The optimal solution is \({\mathcal {X}}^* =\sum ^{2}_{i=1} \rho _i (v^i)^{\otimes 3}\), where
The computation takes about 7 s.
Case 2. Consider (1.3) with the CP cone and the linear constraints \( {\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_2\) are the same as in Case 1 and
We apply Algorithm 3.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=2\) with \(\vartheta _{P_1}^{k}=-16.2281\). The optimal solution is \({\mathcal {X}}^* =\sum ^{2}_{i=1} \rho _i (v^i)^{\otimes 3}\), where
It takes about 6 s.
Case 3. Consider (1.3) with the CP cone and the linear constraints \({\mathcal {A}}_1 \bullet {\mathcal {X}}\ge b_1, {\mathcal {A}}_2 \bullet {\mathcal {X}}= b_2 \), where \({\mathcal {A}}_i, b_i (i=1,2)\) are the same as in Case 2. We apply Algorithm 3.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=2\) with \(\vartheta _{P_1}^{k}=-32.9344\). The optimal solution is \({\mathcal {X}}^* = \rho (v)^{\otimes 3}\), where \(\rho =33.3333\) and
It takes about 5 s.
Case 4. Consider (1.3) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_1\) are the same as in Case 1 and
We apply Algorithm 3.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=2\) and the relaxation problem (\(P^k\)) is infeasible, so the problem (1.3) is infeasible. It takes about 5 s.
Example 5.2
Consider the tensor \({\mathcal {A}} \in \mathrm {S}^5({\mathbb {R}}^{5})\) given as:
Case 1. Consider (1.3) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where
We apply Algorithm 3.1 and choose \(d=6\) and \(k=3\) in Step 0. It terminates at \(k=3\) with \(\vartheta _{P_1}^{k}=-0.2992\). The optimal solution is \({\mathcal {X}}^* = \rho (v)^{\otimes 5}\), where \(\rho =0.3858\) and
It takes about 2 s.
Case 2. Consider (1.3) with the CP cone and the linear constraints \( {\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_1\) are the same as in Case 1 and
We apply Algorithm 3.1 and choose \(d=6\) and \(k=3\) in Step 0. It terminates at \(k=3\) with \(\vartheta _{P_1}^{k}=-0.3175\). The optimal solution is \({\mathcal {X}}^* =\sum ^{2}_{i=1} \rho _i (v^i)^{\otimes 5}\), where
It takes about 2 s.
Case 3. Consider (1.3) with the CP cone and the linear constraints \({\mathcal {A}}_1 \bullet {\mathcal {X}}\ge b_1, {\mathcal {A}}_2 \bullet {\mathcal {X}}= b_2 \), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_2\) are the same as in Case 1 and
We apply Algorithm 3.1 and choose \(d=6\) and \(k=3\) in Step 0. It terminates at \(k=3\) with \(\vartheta _{P_1}^{k}=-3.1521\). The optimal solution is \({\mathcal {X}}^* =\sum ^{2}_{i=1} \rho _i (v^i)^{\otimes 5}\), where
It takes about 2 s.
Case 4. Consider (1.3) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_2\) are the same as in Case 1 and
We apply Algorithm 3.1 and choose \(d=6\) and \(k=3\) in Step 0. It terminates at \(k=3\) and the relaxation problem (\(P^k\)) is infeasible, so the problem (1.3) is infeasible. It takes about 2 s.
5.2 CP projection problems
We compute the CP projection of a given tensor by applying Algorithm 4.1 to (4.2).
Example 5.3
Consider the tensor \({\mathcal {A}}\in \mathrm {S}^4({\mathbb {R}}^{5})\) given as:
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=4\) with \(\gamma ^{*,k}=17.3001\). The optimal solution is \({\mathcal {X}}^* =\sum ^{5}_{i=1} \rho _i (v^i)^{\otimes 5}\), where
It takes about 45 s.
Example 5.4
Consider the tensor \({\mathcal {A}}\in \mathrm {S}^5({\mathbb {R}}^{5})\) given as:
where \(u^k\) is generated randomly as follows:
We apply Algorithm 4.1 and choose \(d=6\) and \(k=3\) in Step 0. It terminates at \(k=5\) with \(\gamma ^{*,k}=2.1140\). The optimal solution is \({\mathcal {X}}^* =\sum ^{3}_{i=1} \rho _i (v^i)^{\otimes 5}\), where
It takes about 30 s.
Example 5.5
Consider the tensor \({\mathcal {A}}\in \mathrm {S}^3({\mathbb {R}}^{10})\) given as (cf. [30]):
and the other entries are zero, except permutations of the above indices. \({\mathcal {A}}\) is a strongly symmetric hierarchically dominated nonnegative tensor, so it is completely positive. A CP decomposition of length 15 of \({\mathcal {A}}\) is also given in [30].
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=3\) with \(\gamma ^{*,k}=3.2576 \times 10^{-7}\). Thus \({\mathcal {A}}\) is CP. The optimal solution is \({\mathcal {X}}^* =\sum ^{14}_{i=1} \rho _i (v^i)^{\otimes 3}\), where
We get a different CP decomposition from the one given in [30]. It takes about 4369 s.
5.3 Best CP approximation problems
In the following, we use Algorithm 4.1 to solve the best CP approximation problems (4.1).
Example 5.6
Consider the tensor \({\mathcal {A}} \in {\mathrm {S}^4({\mathbb {R}}^{5})}\) given as:
Case 1. Consider (4.1) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. \(k=4\) with \(\gamma ^{*,k}=108.2775\). The optimal solution is \({\mathcal {X}}^* =\sum ^{3}_{i=1} \rho _i (v^i)^{\otimes 4}\), where
It takes about 21 s.
Case 2. Consider (4.1) with the CP cone and the linear constraints \({\mathcal {A}}_1 \bullet {\mathcal {X}}\ge b_1, {\mathcal {A}}_2 \bullet {\mathcal {X}}= b_2\), where \({\mathcal {A}}_i, b_i, (i=1,2)\) are the same as in Case 1. We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=3\) with \(\gamma ^{*,k}=107.7770\). The optimal solution is \({\mathcal {X}}^* =\sum ^{3}_{i=1} \rho _i (v^i)^{\otimes 4}\), where
It takes about 18 s.
Case 3. Consider (4.1) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}\ge b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_2\) are the same as in Case 1 and
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=3\) with \(\gamma ^{*,k}=107.7815\). The optimal solution is \({\mathcal {X}}^* =\sum ^{3}_{i=1} \rho _i (v^i)^{\otimes 4}\), where
It takes about 3 s.
Case 4. Consider (4.1) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_2\) are the same as in Case 1 and
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=2\) and the relaxation problem (\(P^k\)) is infeasible, so the problem is infeasible. It takes about 2 s.
Example 5.7
Consider the tensor \({\mathcal {A}}\in \mathrm {S}^3({\mathbb {R}}^{6})\) given as:
where \(u^k\) is generated as:
Case 1. Consider (4.1) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=3\) with \(\gamma ^{*,k}=3.2829\). The optimal solution is \({\mathcal {X}}^* =\sum ^{9}_{i=1} \rho _i (v^i)^{\otimes 3}\), where
It takes about 8 s.
Case 2. Consider (4.1) with the CP cone and the linear constraints \( {\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_2\) are the same as in Case 1 and
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=3\) with \(\gamma ^{*,k}=2.8225\). The optimal solution is \({\mathcal {X}}^* =\sum ^{7}_{i=1} \rho _i (v^i)^{\otimes 3}\), where
It takes about 8 s.
Case 3. Consider (4.1) with the CP cone and the linear constraints \({\mathcal {A}}_1 \bullet {\mathcal {X}}\ge b_1, {\mathcal {A}}_2 \bullet {\mathcal {X}}= b_2 \), where \({\mathcal {A}}_i, b_i (i=1,2)\) are the same as in Case 2. We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=3\) with \(\gamma ^{*,k}=2.6451\). The optimal solution is \({\mathcal {X}}^* =\sum ^{6}_{i=1} \rho _i (v^i)^{\otimes 3}\), where
It takes about 8 s.
Case 4. Consider (4.1) with the CP cone and the linear constraints \({\mathcal {A}}_i \bullet {\mathcal {X}}= b_i (i=1,2)\), where \({\mathcal {A}}_1, {\mathcal {A}}_2, b_1\) are the same as in Case 2 and
We apply Algorithm 4.1 and choose \(d=4\) and \(k=2\) in Step 0. It terminates at \(k=2\) and the relaxation problem (\(P^k\)) is infeasible, so the problem is infeasible. It takes about 0.5 s.
References
Berman, A., Shaked-Monderer, N.: Completely Positive Matrices. World Scientific, Singapore (2003)
Burer, S.: On the copositive representation of binary and continuous nonconvex quadratic programs. Math. Program., Ser. A 120, 479–495 (2009)
Cichocki, A., Zdunek, R., Phan, A.H., Amari, S.: Nonnegative Matrix and Tensor Factorizations: Applications to Exploratory Multiway Data Analysis and Blind Source Separation. Wiley, New York (2009)
Comon, P.: Tensor Decompositions: State of the Art and Applications, Mathematics in Signal Processing, V (Coventry, 2000), Institute for Mathematics and its Apllications Conference Series, vol. 71, pp. 1–24. Oxford University Press, Oxford (2002)
Comon, P., Golub, G., Lim, L.H., Mourrain, B.: Symmetric tensors and symmetric tensor rank. SIAM J. Matrix Anal. Appl. 30, 1254–1279 (2008)
Cui, C., Dai, Y., Nie, J.: All real eigenvalues of symmetric tensors. SIAM J. Matrix Anal. Appl. 35, 1582–1601 (2014)
Curto, R., Fialkow, L.: Truncated K-moment problems in several variables. J. Oper. Theory 54, 189–226 (2005)
Dickinson, P.J., Gijben, L.: On the computational complexity of membership problems for the completely positive cone and its dual. Comput. Optim. Appl. 57, 403–415 (2014)
Fan, J., Zhou, A.: A semidefinite algorithm for completely positive tensor decomposition. Comput. Optim. Appl. (2016). https://doi.org/10.1007/s10589-016-9870-9
Fan, J., Zhou, A.: The CP-matrix approximation problem. SIAM J. Matrix Anal. Appl. 37, 171–194 (2016)
Fialkow, L., Nie, J.: The truncated moment problem via homogenization and flat extensions. J. Funct. Anal. 263, 1682–1700 (2012)
Henrion, D., Lasserre, J.: Detecting global optimality and extracting solutions in GloptiPoly, positive polynomials in control. In: Lecture Notes in Control and Information Sciences, vol. 312, pp. 293–310. Springer, Berlin (2005)
Henrion, D., Lasserre, J., Loefberg, J.: GloptiPoly 3: moments, optimization and semidefinite programming. Optim. Methods Softw. 24, 761–779 (2009)
Hu, Shenglong, Huang, Zheng-Hai, Qi, Liqun: Strictly nonnegative tensors and nonnegative tensor partition. Sci. China Math. 57, 181–195 (2014)
Kuang, X., Zuluaga, L.: Completely positive and completely positive semidefinite tensor relaxations for polynomial optimization. J. Glob. Optim. 70, 551–577 (2018)
Kolda, Tamara G., Bader, Brett W.: Tensor decompositions and applications. SIAM Rev. 51, 455–500 (2009)
Kolda, Tamara G., Mayo, J.R.: Shifted power method for computing tensor eigenpairs. SIAM J. Matrix Anal. Appl. 32, 1095–1124 (2011)
Kolda, Tamara G.: Numerical optimization for symmetric tensor decomposition. Math. Program., Ser. B 151, 225–248 (2015)
Lasserre, J.B.: Moments, Positive Polynomials and Their Applications. Imperial College Press, London (2009)
Laurent, M.: Sums of squares, moment matrices and optimization over polynomials. In: Putinar, M., Sullivant, S. (eds.) Emerging Applications of Algebraic Geometry, IMA Volumes in Mathematics and its Applications, vol. 149, pp. 157–270. Springer, Berlin (2009)
Luo, Z., Qi, L.: Completely positive tensors: properties, easily checkable subclasses, and tractable relaxations. SIAM J. Matrix Anal. Appl. 37, 1675–1698 (2016)
Nie, J.: The \(\cal{A}\)-truncated K-moment problem. Found. Comput. Math. 14, 1243–1276 (2014)
Nie, J.: Linear optimization with cones of moments and nonnegative polynomials. Math. Program., Ser. B 153, 247–274 (2015)
Nie, J.: Optimality conditions and finite convergence of Lasserre’s hierarchy. Math. Program., Ser. A 146, 97–121 (2014)
Nie, J., Wang, L.: Semidefinite relaxations for best rank-1 tensor approximations. SIAM J. Matrix Anal. Appl. 35, 1155–1179 (2014)
Nie, J.: Generating polynomials and symmetric tensor decompositions. Found. Comput. Math. 17, 423–465 (2017)
Nie, J., Schweighofer, M.: On the complexity of Putinar’s Positivstellensatz. J. Complex. 23, 135–150 (2007)
Peña, J., Vera, J.C., Zuluaga, L.F.: Completely positive reformulations for polynomial optimization. Math. Program., Ser. B 151, 405–431 (2015)
Putinar, M.: Positive polynomials on compact semi-algebraic sets. Indiana Univ. Math. J. 42, 969–984 (1993)
Qi, L., Xu, C., Xu, Y.: Nonnegative tensor factorization, completely positive tensors and an hierarchical elimination algorithm. SIAM J. Matrix Anal. Appl. 35, 1227–1241 (2014)
Qi, L.: Symmetric nonnegative tensors and copositive tensors. Linear Algebra Appl. 439, 228–238 (2013)
Shashua, A., Hazan, T.: Non-negative tensor factorization with applications to statistics and computer vision. In: ACM International Conference Proceeding Series: Proceedings of the 22nd International Conference on Machine Learning, vol. 119, pp. 792–799 (2005)
Shashua, A., Zass, R., Hazan, T.: Multiway clustering using supersymmetric nonnegative tensor factorization. In: Proceedings of the European Conference on Computer Vision, pp. 595–608 (2006)
Sturm, J .F.: SeDuMi 1.02: A MATLAB toolbox for optimization over symmetric cones. Optim. Methods Softw. 11 & 12, 625–653 (1999)
Xu, C., Luo, Z., Qi, L., Chen, Z.: \(\{0,1\}\) completely positive tensors and multi-hypergraphs. Linear Algebra Appl. 510, 110–123 (2016)
Zhou, A., Fan, J.: The CP-matrix completion problem. SIAM J. Matrix Anal. Appl. 35, 127–142 (2014)
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Anwa Zhou was partially supported by the NSFC Grant 11701356, the National Postdoctoral Program for Innovative Talents Grant BX201600097 and Project Funded by China Postdoctoral Science Foundation Grant 2016M601562. Jinyan Fan was partially supported by the NSFC Grant 11571234.
Rights and permissions
About this article
Cite this article
Zhou, A., Fan, J. A hierarchy of semidefinite relaxations for completely positive tensor optimization problems. J Glob Optim 75, 417–437 (2019). https://doi.org/10.1007/s10898-019-00751-8
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10898-019-00751-8
Keywords
- CP tensor program
- Best CP tensor approximation
- Linear optimization with moments
- Nonnegative decomposition
- Semidefinite program