Abstract
We prove the following Farkas’ Lemma for simultaneously diagonalizable bilinear forms: If \(A_1,\ldots ,A_k\), and \(B:\mathbb {R}^n \times \mathbb {R}^n \rightarrow \mathbb {R}\) are bilinear forms, then one—and only one—of the following holds:
-
(i)
\(B=a_1 A_1 + \cdots + a_k A_k,\) with non-negative \(a_i\text {'s}\),
-
(ii)
there exists (x, y) for which \(A_1(x,y) \ge 0 , \ldots , A_k(x,y) \ge 0\) and \(B(x,y) < 0\).
We study evaluation maps over the space of bilinear forms and consequently construct examples in which Farkas’ Lemma fails in the bilinear setting.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
There are many results relating the zero sets of linear or multilinear forms and their linear dependence. For example, it is well-known that if \(f_1, \ldots , f_k\) and g are linear forms such that
then \(g=a_1 f_1 + \cdots + a_k f_k\) for some scalars \(a_1, ..., a_k.\)
Also, if
then \(g=a_1 f_1 + \cdots + a_k f_k\), with \(a_i \ge 0\) for all i. This is known as Farkas’ Lemma [4]. Note that condition (2) above is stronger than condition (1), as can be checked by replacing x with \(-x\).
Published in 1902, Farkas’ Lemma leads to linear programming duality and has proved to be a central result in the development of linear and non-linear mathematical optimization. According to MathSciNet, over 250 papers cite Farkas’ Lemma (one such contains the particularly beautiful statement by J. Franklin [7] that “I hope to convince you that every mathematician should know the Farkas theorem and should know how to use it”). One good, fairly recent source for such papers can be found among the opening articles in [5]. However, very few of these papers deal with non-linear versions of the lemma which is our interest here. In the non-linear setting, a short article describing some open problems involving polynomial matrices is contained in the work of A. A. ten Dam and J. W. Nieuwenhuis [2].
In the multilinear setting, positive results are known only in the case of two m-linear forms A and B as follows:
then \(B=a A\) [1].
In addition, the following weak form of Farkas’ Lemma is given in [3]:
then \(B=a A\), with \(a\ge 0 .\)
On the other hand, in [1] the authors give an example of symmetric bilinear forms \(A_1, A_2\) and B such that
and in [3] the author shows (non-symmetric) bilinear forms \(A_1, A_2\) and B such that
with non-negative \(a_i\)’s.
Farkas’ Lemma also fails for 2-homogeneous polynomials. Consider \(P:\mathbb {R}^2 \rightarrow \mathbb {R}\) and \(Q:\mathbb {R}^2 \rightarrow \mathbb {R}\) given by \(P(x,y)=x^2\) and \(Q(x,y)=y^2\). Then \(P(x,y)\ge 0\) implies \(Q(x,y)\ge 0\), since both are always non-negative, but P and Q are independent.
In view of these examples, there appears to be little room for positive results. However in Sect. 1, below, we give a version of Farkas’ Lemma for simultaneously diagonalizable bilinear forms (see Theorem 1) as well as a result analogous to (1) above (see Theorem 2).
We note that a similar, but weaker, result:
(where \(C\succeq D\) means \(C-D\) is positive semidefinite), is known and can be obtained through the S-procedure as in [6]. We require, however, equality in (3) for our version of the Farkas’ Lemma.
Moreover, in Sect. 2, we take a closer look at the reason for the lack of positive results. Farkas’ Lemma may be viewed as an application of the Hahn-Banach Theorem (unavailable to Farkas in 1902): indeed, given linear forms \(f_1, \ldots , f_k\), their positive cone
is a convex set, so if \(g\not \in \mathcal {F}\) then it can be separated from \(\mathcal {F}\) by a linear functional on \({\mathbb {R}^n}^*\). But all linear functionals on \({\mathbb {R}^n}^*\) are given by “evaluation maps”. Thus there is a point \(x\in \mathbb {R}^n\) such that
and in particular, \(f_1(x)\ge 0, \ldots , f_k(x)\ge 0\) and \(g(x)<0\). This is the crucial point for the lack of Farkas type results in the multilinear (and other) settings: few linear functionals are “evaluation maps”. We study evaluation maps on spaces of real-valued bilinear forms on \(\mathbb {R}^n \ (n \ge 2)\) (see Propositions 3 and 4), and this enables us to produce new examples (see Examples 1 and 2) of bilinear forms \(A_1, A_2\) and B such that
with non-negative \(a_i\)’s.
2 Farkas’ Lemma for simultaneously diagonalizable bilinear forms
In this section, \(A_1,\ldots ,A_k\), and B will be bilinear forms on \(\mathbb {R}^n \times \mathbb {R}^n\). We will use the notation \(B_x(y)=B(x,y)\) (and similarly for the bilinear forms \(A_i\)). Note that a bilinear form A(x, y) can be written as \(x^t[A]y,\) where \([A]\in \mathbb {R}^{n \times n}\) is the coefficient matrix of A in the canonical basis: \([A]_{ij}=A(e_i,e_j)\). We shall prove Farkas’ Lemma in the following form:
Theorem 1
For \(j = 1, ..., k,\) let \(A_j:\mathbb {R}^n\times \mathbb {R}^n \rightarrow \mathbb {R}\) be simultaneously diagonalizable bilinear forms. Then
if and only if
Proof
Clearly \(\Rightarrow )\) is trivial, so we must see \(\Leftarrow )\).
Note that for any \(x\in \mathbb {R}^n\), we have
so that by the linear Farkas’ Lemma there are \(a_1(x), \ldots , a_k(x) \ge 0\) such that
The \(A_i\)’s are diagonalized simultaneously by a matrix U whose columns form a basis of eigenvectors \(\{u_1, \ldots , u_n\}\). Set
and
We have \(\langle v_i,u_j \rangle =\delta _{ij}\), and for each \(j=1,\ldots , k\),
Note that if \(r\not =s\), then for all j, \(A_j(v_r,u_s)=0.\) Therefore from \((*)\) we deduce that \(B(v_r,u_s)=0\) for \(r\not =s\) as well. Thus the same basis diagonalizes B. We have
Consider \({\textsc {1}}\!\!\!{{\textsc {1}}}=(1,\ldots ,1)\), and \(z=(z_1,\ldots ,z_n)\). Then for each \(j=1,\ldots , k\),
and similarly, \(B((U^{-1})^t {\textsc {1}}\!\!\!{{\textsc {1}}} , Uz)=B(v_1,u_1)z_1 + \cdots + B(v_n,u_n)z_n\). Now let \(a_1=a_1((U^{-1})^t {\textsc {1}}\!\!\!{{\textsc {1}}})\ge 0, \ldots , a_k=a_k((U^{-1})^t {\textsc {1}}\!\!\!{{\textsc {1}}})\ge 0\). Since, by \((**)\),
we have, for any \(z \in \mathbb {R}^n\),
Now, for \(r=1,\ldots ,n\), set \(z=e_r.\) We have
Thus
Therefore
and the proof is complete. \(\square \)
Note that by relaxing condition \((*)\) to
with the same proof but using (1) instead of the linear Farkas’ Lemma, we have the following.
Theorem 2
For \(j = 1, ..., k,\) let \(A_j:\mathbb {R}^n\times \mathbb {R}^n \rightarrow \mathbb {R}\) be simultaneously diagonalizable bilinear forms. Then
if and only if
Also, since any number of symmetric commuting matrices are simultaneously diagonalizable [8, Theorem 1.3.21], we have the following corollary.
Corollary 1
If \(A_1,\ldots ,A_k\), are symmetric bilinear forms on \(\mathbb {R}^n \times \mathbb {R}^n\) defined by commuting matrices, then
if and only if
Remark 1
We note that an analogous result can be obtained in the setting of a Hilbert space H as follows:
If \(A_1,\ldots ,A_k\), are symmetric bilinear forms on \(H \times H\) defined by \(A_i(x,y)=\langle y,T_i x \rangle \) where, for \(i=1,\ldots ,k\), the \(T_i'\hbox {s}\) are compact self-adjoint commuting operators, then
if and only if
3 Evaluation maps over spaces of bilinear forms
As we have mentioned above, linear forms on \({\mathbb {R}^n}^*\) identify with elements of \(\mathbb {R}^n\), and thus are all evaluation maps: \(\gamma \mapsto \gamma (x)\). This is far from the case in most function spaces. For example, continuous linear functionals on C[0, 1] are identified with regular Borel measures, while the evaluation maps are the deltas: \(\delta _x (f)=f(x)\). The two convex sets
can be separated by the measure \(\delta _x - \delta _y\) for any \(0 \le x \ne y \le 1,\) although they cannot be separated by any evaluation map \(\delta _x\).
We now characterize the evaluation maps on the space of bilinear forms \(\mathcal {L}^2 (\mathbb {R}^n)\) and on the space of symmetric bilinear forms \(\mathcal {L}_s^2 (\mathbb {R}^n)\) in order to construct examples of positive cones
and \(B\not \in \mathcal {F}\) which cannot be separated by evaluation maps.
We consider \(\mathcal {L}^2 (\mathbb {R}^n)\) as \(\mathbb {R}^{n \times n}\), the space of \(n \times n\) matrices with inner product:
where \(A=(a_{ij})\) and \(B=(b_{ij})\). For a matrix A, we denote by tr(A) the trace of A and by rk(A) the rank of A. Any \(\varphi \in \mathcal {L}^2 (\mathbb {R}^n)^*\) can be represented as \(\varphi (A)=tr(AN^t)\) for some \(n\times n\) matrix N. As we will see below, in the symmetric setting, any \(\varphi \in \mathcal {L}_s^2 (\mathbb {R}^n)^*\) can be represented as \(\varphi (A)=tr(AS)\), with a symmetric matrix S. We say that \(\varphi \) is an evaluation map if there are x and y in \(\mathbb {R}^n\) such that \(\varphi (A)=A(x,y)=x^tAy\).
We characterize the evaluation maps in \(\mathcal {L}^2 (\mathbb {R}^n)^*\) and \(\mathcal {L}_s^2 (\mathbb {R}^n)^*\) as matrices N and S as follows.
Proposition 3
\(\varphi \in \mathcal {L}^2 (\mathbb {R}^n)^*\) is an evaluation map if and only if \(\varphi (A)=tr(AN^t)\), where N has rank less than or equal to one.
Proof
\(x^tAy=\sum _i \sum _j a_{ij} x_i y_j\), so set \(n^t_{ij}=x_i y_j\) for the entries of \(N^t\). Thus
N has rank 0 if \(\varphi =0\) and rank one otherwise.
Conversely, if N has rank one, we can write
for some \(y \in \mathbb {R}^n,\) and so \(n^t_{ij}=x_i y_j\), and N produces an evaluation map. \(\square \)
We now discuss the symmetric setting. We note that we can associate \(\mathcal L_s^2(\mathbb {R}^n)\) with the space \(\mathbb {R}_s^{n \times n}\) of symmetric \(n\times n\) matrices. Also, any continuous linear functional \(\varphi :\mathcal L^2_s(\mathbb {R}^n) \rightarrow \mathbb {R}\) is given by \(\varphi (M) = tr(M[\varphi ]^t)\) for a suitable symmetric \(n \times n\) matrix \([\varphi ]\). Indeed, since M is symmetric, if \(\varphi (M)=tr(MN^t)\), we can replace N by a symmetric matrix as follows.
For each \(x, y \in \mathbb {R}^n,\) we will identify an evaluation functional E(x, y) on \(\mathbb {R}_s^{n\times n}\) by
We can associate E(x, y) to the symmetric \(n \times n\) matrix \([E(x,y)] \in \mathbb {R}_s^{n\times n}\) where
Observe that if a matrix \(N \in \mathbb {R}_s^{n \times n}\) represents an evaluation map, then so does the matrix \(UNU^t,\) where \(U \in \mathcal O(n)\) (the orthogonal group in \(\mathbb {R}^{n\times n}\)). To see this, if \(N = E(x,y)\) is an evaluation map, so that \(E(x,y)(M) = tr(MN^t),\) then for any such U,
where the second equality holds since N is symmetric.
Therefore, if N is a symmetric \(n \times n\) matrix, we can find an orthogonal matrix U so that \(UNU^t\) has a diagonal representation. In particular, for any such diagonal matrix \([E(x,y)] =\)
In fact, at most two of the diagonal entries must be non-zero. Indeed, suppose that there are three non-zero entries, \(\lambda _i = x_iy_i, \lambda _j = x_jy_j,\) and \(\lambda _k = x_ky_k.\) Since the off-diagonal entries are all 0, it follows that
As a consequence, \(\lambda _jx_i^2 + \lambda _ix_j^2 = 0, \lambda _kx_i^2 + \lambda _ix_k^2 = 0,\) and \(\lambda _kx_j^2 + \lambda _jx_k^2 = 0.\) Since none of the entries is 0, it follows that all three of \(\lambda _i, \lambda _j,\) and \(\lambda _k\) are of opposite sign, a clear impossibility. In addition, the argument shows that if there are just two non-zero diagonal entries, say with \(i = 1\) and \(j = 2,\) then \(\lambda _1\lambda _2 < 0.\)
Consequently, if \(\lambda _1 = \lambda _2 = 0,\) then the linear form \(\varphi = E(\vec {0},\vec {0}),\) and if only \(\lambda _1 \ne 0,\) then \(\varphi = E(\lambda _1 e_1, e_1).\) The third and last case occurs if \(\lambda _1 \ne 0 \ne \lambda _2.\) If say \(\lambda _1> 0 > \lambda _2,\) then by using the facts that \(\lambda _2\) is negative and that we are dealing with symmetric matrices it follows that \(\varphi = E(\sqrt{\frac{-\lambda _1}{\lambda _2}} \ e_1 + e_2, \sqrt{-\lambda _1 \lambda _2}\ e_1 +\lambda _2e_2).\)
Thus, given an evaluation map E(x, y), we have now seen that to the associated matrix \([E(x,y)] \in \mathbb {R}_s^{n\times n}\) there is an orthogonal matrix U such that \(U[E(x,y)]U^t\) is diagonal which also represents an evaluation map. Summarizing, we have the following.
Proposition 4
The square matrix \(S \in \mathbb {R}_s^{n\times n}\) represents an evaluation map if and only if S is one of the following:
-
(i)
\(S = 0,\) or
-
(ii)
S has only one non-zero eigenvalue, or
-
(iii)
S has exactly two non-zero eigenvalues which are of opposite sign.
We now construct counterexamples to Farkas’ Lemma for both cases, the four dimensional \(\mathcal {L}^2(\mathbb {R}^2)\) and the three dimensional \(\mathcal {L}_s^2(\mathbb {R}^2)\). Note that according to our characterizations, evaluation functionals over \(\mathcal {L}^2(\mathbb {R}^2)\) are of the form \(\varphi (A)=\langle A,N\rangle \) where \(\det N=0\), while evaluation functionals over \(\mathcal {L}_s^2(\mathbb {R}^2)\) are of the form \(\varphi (A)=\langle A,S\rangle \) where \(\det S\le 0\). We begin with counterexamples in \(\mathcal {L}^2(\mathbb {R}^2)\).
Example 1
Let \([I]^\perp \subset \mathcal {L}^2(\mathbb {R}^2)\) be the orthogonal complement of the line spanned by the identity matrix, and \(\{A_1, A_2, A_3 \}\) an orthonormal basis of \([I]^\perp \). We consider \(B=A_1+A_2+A_3-\frac{\varepsilon }{\sqrt{2}} I\) where \(\varepsilon > 0\) will be chosen later. Clearly \(B\not \in \mathcal {F} \equiv \{a_1 A_1+a_2 A_2+a_3 A_3 : a_i \ge 0 \}\). Let \(\varphi \) be a linear form separating B from \(\mathcal {F}\) chosen so that
\(\varphi \) is given by the matrix N: \(\varphi (X)=\langle X,N\rangle \). We may suppose N has norm one. We will show that if \(\varepsilon >0\) is chosen to be sufficiently small, then \(\varphi \) cannot be an evaluation map. We have
so
Thus
The first inequality follows from the fact that the right part is equal to the left part plus the double products of \(\langle A_i, N\rangle \langle A_j, N\rangle \), which are terms greater than or equal to zero since \(0 \le \varphi (X)\) for all \(X \in \mathcal {F}\).
Also,
so
Thus \(\sqrt{2}\sqrt{1-\varepsilon ^2} \ <\ \ \langle I,N\rangle \). Now by the parallelogram law we have
which can be made smaller than one for small \(\varepsilon \) because \(3 - 2\sqrt{2} < 0.18\). Thus, for such \(\varepsilon \), \(\Vert I-N \Vert \) is smaller than one, and N is invertible. By Proposition 3, \(\varphi \) cannot be an evaluation map.
For the symmetric case we have the following.
Example 2
Let \([I]^\perp \subset \mathcal {L}_s^2(\mathbb {R}^2)\) be the orthogonal complement of the line spanned by the identity matrix, and \(\{A_1, A_2 \}\) be an orthonormal basis of \([I]^\perp \). We consider \(B=A_1+A_2-\frac{\varepsilon }{\sqrt{2}} I\) where \(\varepsilon > 0\) will be chosen later. Clearly B is not in \(\mathcal {F} \equiv \{a_1 A_1+a_2 A_2 : a_i \ge 0 \}\). Let \(\varphi \) be such that
\(\varphi \) is given by the symmetric matrix S, which we may suppose has norm one: \(\varphi (X)=\langle X,S\rangle \). Picking a suitable small enough \(\varepsilon \) as we have done in the previous example, we have \(\Vert I-S \Vert < 1\). But then the determinant of S must be positive: if \(\det S \le 0\), the line segment joining I and S would, by the intermediate value theorem, contain a matrix X with \(\det X = 0\), at a distance smaller than one from the identity. By Proposition 4, \(\varphi \) cannot be an evaluation map.
Note that no basis of \([I]^\perp \) can be diagonalized simultaneously. Otherwise, all matrices would be simultaneously diagonalizable.
References
Aron, R., Downey, L., Maestre, M.: Zero sets and linear dependence of multilinear forms. Note Mat. 25(1), 49–54 (2005)
ten Dam, A.A., Nieuwenhuis, J.W.: A Farkas lemma for behavioral inequalities. In: Blondel, V.D., Megretski, A. (Eds.), Unsolved problems in mathematical systems and control theory, Princeton, 40–43 (2004)
Downey, L.: Farkas’ Lemma and multilinear forms. Missouri J. Math. Sci. 21(1), 65–67 (2009)
Farkas, J.: Theorie der einfachen Ungleichungen. J. Reine Angew. Math. 124, 1–27 (1902)
Farkas’ lemma: three decades of generalizations for mathematical optimization, TOP (An Official Journal of the Spanish Society of Statistics and Operations Research) 22 (2014)
Fradkov, A.L.: Duality theorems for certain nonconvex extremal problems. Siber. Math. J. 14, 247–264 (1973)
Franklin, J.: Mathematical methods of economics. Am. Math. Monthly 90, 229–244 (1983)
Horn, R.A., Johnson, C.R.: Matrix Analysis, 2nd edn., p. 643. Cambridge University Press, Cambridge (2013)
Acknowledgements
The first and second authors were supported by projects MTM2017-83262-C2-1-P/MCIN/AEI/10.13039/501100011033 (FEDER) and PID2021-122126NB-C33/MCIN/AEI/10.13039/501100011033 (FEDER). The second was also supported by PROMETEU/2021/070. The third and fourth authors were supported partially by PIP 112 201301 00422 CO (CONICET - Argentina).
Funding
Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Aron, R., García, D., Pinasco, D. et al. Farkas’ Lemma in the bilinear setting and evaluation functionals. Rev. Real Acad. Cienc. Exactas Fis. Nat. Ser. A-Mat. 117, 6 (2023). https://doi.org/10.1007/s13398-022-01337-y
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s13398-022-01337-y