Abstract
Feasible region is of a basic significance to optimization. Due do the linear structure of the standard LP problem (1.8), the feasible region P, defined by (1.9), has special features. In this chapter, we explore this theme from a geometric point of view.
Access provided by Autonomous University of Puebla. Download chapter PDF
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
Feasible region is of a basic significance to optimization. Due do the linear structure of the standard LP problem (1.8), the feasible region P, defined by (1.9), has special features. In this chapter, we explore this theme from a geometric point of view.
Denote the n-dimensional Euclidean space by \({\mathcal{R}}^{n}\), as usual. Points or column vectors are the basic geometrical elements. Denote a point, or the vector from the origin to this point by
whose components (coordinates) are x 1, …, x n . Thereafter, points and vectors will not be distinguished. Denote by e j , j = 1, …, n the jth coordinate vector, i.e., the unit vector with its jth component 1, and denote by e the vector of all ones. The reader is referred to related literature for basic concepts and operations in Euclidean space, such as linear dependency and independency, set of points and its boundedness and unboundedness, the inner product x T y of vectors \(x,y \in {\mathcal{R}}^{n}\), \(\cos < x,y >= {x}^{T}y/(\|x\|\|y\|)\) of the angle between them, their orthogonality x ⊥ y or x T y = 0, i.e., \(< x,y >=\pi /2\), the Euclidean module or norm \(\|x\| = \sqrt{{x}^{T } x}\) of vector x, and so on.
Involved in the standard LP problem, both c and x may be viewed as vectors in \({\mathcal{R}}^{n}\), columns a j , j = 1, …, n of A as vectors in \({\mathcal{R}}^{m}\), b as a vector in \({\mathcal{R}}^{m}\), and the feasible region P as a closed polyhedral in \({\mathcal{R}}^{n}\) in general (as will be clear a little later), though P could be degenerate, or even empty. The following lemma, proposed by Farkas (1902), renders a sufficient and necessary condition for nonempty P (the proof is delayed to at the end of Sect. 4.2).
FormalPara Lemma 2.1 (Farkas).Assume \(A \in {\mathcal{R}}^{m\times n}\) and \(b \in {\mathcal{R}}^{m}\) . The feasible region P is nonempty if and only if
Fig. 2.1 serves as a geometric explanation for the preceding Lemma, where \(a_{1},a_{2},a_{3}\) are columns of \(A \in {\mathcal{R}}^{2\times 3}\). Y is the set of all vectors which forms with every column of A an angle no more than π∕2 (corresponding to the shaded area between vectors v and w in the figure). b 1 forms with each y ∈ Y an angle no more than π∕2 but b 2 does not. Thereby, P is nonempty when b = b 1, whereas empty when b = b 2.
In addition to rank A = m, discussions in this chapter will be based on the following.
FormalPara Assumption.The feasible region P is nonempty and infinite.
1 Polyhedral Convex Set and the Feasible Region
For any given two points \(x,y \in {\mathcal{R}}^{n}\), set
is a straight line; if 0 < α < 1, it is an open segment with end points x and y, denoted by (x, y); if 0 ≤ α ≤ 1, it is a closed segment, denoted by [x, y]. Hereafter so-called “segment” will be all closed.
Definition 2.1.1.
Π is an affine set if, whenever it includes any two points, it includes the whole straight line passing through them. The smallest affine set including a set is the affine hull of the latter.
Straight lines in \({\mathcal{R}}^{2}\) and planes in \({\mathcal{R}}^{3}\) are instances of affine sets. The whole space \({\mathcal{R}}^{n}\) is an affine set. An empty set and a single point set are viewed as affine sets. It is clear that the intersection of affine sets is an affine set.
For any given α i , i = 1, …, k satisfying \(\sum _{i=1}^{k}\alpha _{i} = 1\), point
is called an affine combination of x 1, …, x k. It is easy to show that the set of all such affine combinations, i.e.,
is an affine set, called affine hull of these points. The two points in Definition 2.1.1 can be generalized to multiple points: it is easy to show that \(\Pi \) is an affine set if and only if the affine hull of any finitely many points within \(\Pi \) belongs to \(\Pi \).
Set L is a subspace of \({\mathcal{R}}^{n}\) if it is closed for all linear operations, that is, for any x, y ∈ L and \(\alpha,\beta \in \mathcal{R}\) it holds that α x +β y ∈ L. An affine set is an affine subspace if it is a subspace.
Theorem 2.1.1.
An affine set is an affine subspace if and only if it includes the origin.
Proof.
The necessity is clear. Sufficiency. Let \(\Pi \) be an affine set including the origin. Then for any \(x \in \Pi \) and \(\alpha \in \mathcal{R}\), it holds that
On the other hand, it holds for any \(x,y \in \Pi \) that
Therefore, \(\Pi \) is closed for linear operations, and is thus an affine subspace. □
Theorem 2.1.2.
For any nonempty affine set \(\Pi \) , there exists vector p so that
is an affine subspace, and such subspace is unique.
Proof.
According to Theorem 2.1.1, \(\Pi \) is an affine subspace if \(0 \in \Pi \). Note that \(L = \Pi \) corresponds to p = 0. Assume that \(0\not\in \Pi \). Since \(\Pi \neq \varnothing \), there exists \(0\neq y \in \Pi \). Letting \(p = -y\), it is clear that
is an affine set including the origin, and is hence an affine subspace.
Now let us show the uniqueness. Assume that L 1, L 2 are affine subspaces such that
It is clear that
If \(p = -p_{1} + p_{2}\), therefore, it holds that
from which it follows that x + p ∈ L 2 for any x ∈ L 1. Since 0 ∈ L 1, p ∈ L 2 holds. Further, since L 2 is a subspace, \(x = (x + p) - p \in L_{2}\) holds. Therefore, L 1 ⊂ L 2. L 2 ⊂ L 1 can be similarly derived. So it can be asserted that L 1 = L 2. □
Geometrically, the affine subspace L may be viewed as a parallelism of affine set \(\Pi \) along vector p. It is therefore called parallel subspace of \(\Pi \). The dimension of L is said to be that of \(\Pi \), which is equal to the number of independent components (coordinates) of elements in \(\Pi \). It is clear that an affine set with one or more than one dimension is unbounded.
Let a be a nonzero vector and let η be a real number. Set
is called superplane, whose normal vector is a (a ⊥ H); in fact, for any two points x, y ∈ H, it holds that
It is easy to show that any superplane is an affine set.
Any straight line in \({\mathcal{R}}^{2}\) and any plane in \({\mathcal{R}}^{3}\) are instances of superplane.
The “signed” distance from any point \(\bar{x}\) to superplane H is defined by \(r/\|a\|\), where r is the residual
If r = 0, point \(\bar{x}\) is within H. It might be well to assume η > 0. Then, if r < 0, the origin and \(\bar{x}\) are in the same side of H; if r > 0, the two points are in different sides of H.
Superplanes associated with the objective function are of significance to LP. Regarding objective value f as a parameter, the sets
are a family of contour surfaces of the objective function. The gradient ∇f = c of the objective function is the common normal vector of all the contour surfaces, pointing to the increasing side of objective value f.
The following gives a mathematical expression of an affine set.
Theorem 2.1.3.
Set \(\Pi \) is an affine set if and only if there exists \(W \in {\mathcal{R}}^{k\times n}\) and \(h \in {\mathcal{R}}^{k}\) such that
Proof.
It might be well to rule out the trivial case when \(\Pi \) is empty or the whole space.
Sufficiency. Let \(\Pi \) is defined by (2.1). For any \(x,y \in \Pi \) and α ∈ R 1, it holds that
leading to \(\alpha x + (1-\alpha )y \in \Pi \). Therefore, \(\Pi \) is an affine set.
Necessity. Let \(\Pi \) be an affine set. Assume that L is a parallel affine subspace, and w 1, …, w k are basis of the orthogonal complementary space of it. Then, it follows that
where rows of \(W \in {\mathcal{R}}^{k\times n}\) are \({({w}^{1})}^{T},\ldots,{({w}^{k})}^{T}\). Introduce notation h = Wp. Since L is a parallel subspace of \(\Pi \), there exists a vector p such that
Thus, for any \(x \in \Pi \), it holds that x − p ∈ L; and \(W(x - p) = 0\) means that x ∈ { x | Wx = h}. If x ∈ { x | Wx = h}, conversely, then \(Wx - Wp = W(x - p) = 0\), hence from x − p ∈ L it follows that \(x \in \Pi \). So, \(\Pi \) has expression (2.1). □
The preceding Theorem says that a set is affine set if and only if it is the intersection of finitely many superplanes. It is easy to show that the dimension of the affine set is equal to n − rank(W). In particular, the solution set
of the constraint system of the standard LP problem is an affine set. Since rank A = m, \(\Delta \neq \varnothing \) and \(\mathrm{dim}\ \Delta = n - m\).
Definition 2.1.2.
C is a convex set if it includes the whole segment whenever it includes its two end points. The smallest convex set including a set is the convex hull of the latter.
Any disks and the first quadrant in \({\mathcal{R}}^{2}\) and spheres in \({\mathcal{R}}^{3}\) are instances of convex sets. Clearly, segments and the whole space are convex sets too. Empty sets and single point sets are regarded as convex sets. Intersections of convex sets are convex. Any affine set is convex, but a convex set is not an affine set in general; e.g., disks and spheres are not affine sets. It is clear that any convex set has an affine hull. The dimension of the latter is said to be the dimension of the former. Hereafter, so-called “convex set” is a closed convex set.
For any α i ≥ 0, i = 1, …, k satisfying \(\sum _{i=1}\alpha _{i} = 1\), the point
is called convex combination of points x 1, …, x k. It is easy to show that the set of all such convex combinations, i.e.,
is the convex hull of these points. It is easy to show that C is a convex set if and only if the convex hull of any finitely many points within C belongs C.
Any superplane H divides the whole space to two closed half spaces, i.e.,
The intersection of infinitely many closed half spaces is called polyhedral, and bounded polyhedral called polyhedron. It could degenerate to a segment or point, or even empty set.
It is clear that a half space is convex. Therefore, a polyhedral or polyhedron is convex, as termed polyhedral convex set.
A convex set C is called polyhedral convex cone if α x ∈ C holds for any x ∈ C and α ≥ 0. It is easy to show that a set is a polyhedral convex cone if and if it is the intersection of finitely many closed half spaces passing through the origin, as expressed \(\{x \in {\mathcal{R}}^{n}\ \vert \ Ax \geq 0\}\).
The nonnegative constraints x ≥ 0 in the standard problem correspond to the positive octant, which is the intersection of the n closed half spaces with the coordinate planes as its boundary. Therefore, the feasible region P is the intersection of the affine set \(\Delta \) and the positive octant. As any superplane a T x = η may be viewed as the intersection of two closed half spaces (2.2), P may be also viewed as the intersection of finitely many closed half spaces. Therefore, we make the following statement.
Proposition 2.1.1.
The feasible region P is a polyhedral convex set.
Such a set could be degenerate, however. The following result concerns the dimension of P.
Proposition 2.1.2.
Define index set
where μ j , j ∈ A are nonnegative constants and denote by I J′ the coefficient matrix of system x j = μ j , j ∈ J′. Then it holds that
where
Proof.
Note that rank A = m and P is a nonempty infinite set, according to the basic assumption.
It is clear that | J′ | ≤ r and
IF | J′ | = r, then x j = μ j , j ∈ J′ is a canonical form of system
If | J′ | < r, besides x j j ∈ J′, there are additional r − | J′ | basic variables, and hence a canonical form of the preceding system. It is to say that there exists a canonical form, whose n − r nonbasic variables do not belong to J′. Therefore, it holds that \(\mathrm{dim}\ P = n - r\). Hence (2.4) follows from min{m + | J′ | , n} ≥ r ≥ max{m, | J′ | }. □
In particular, \(\mathrm{dim}\ P = n - m\) if J′ = ∅.
The special case of μ j = 0 is of significance to the standard LP problem. Introduce sets
where J is said to be index set of zero components.
According to the preceding definition, it is clear that P ⊂ P′, and in addition
If x ∈ P and x j > 0, then \(j \in \bar{ J}\).
If J ≠ ∅, the feasible region P has no interior point in the normal sense, as is often the case for real problems. For convenience of applications, the following concept is introduced instead.
Definition 2.1.3.
Assume \(\bar{x} \in P\). If there exists δ > 0 such that a neighborhood of \(\bar{x}\) is included in P, i.e.,
then \(\bar{x}\) is an interior point of P; otherwise, it is a boundary point.
The set of all interior points of P is called its interior, denoted by int P. P and its interior have the same dimension.
For sake of distinguishing, the point is said to be relative interior point when J ≠ ∅. The set of relative interior points is relative interior, while (strict) interior point or interior stands for the case of J = ∅.
Theorem 2.1.4.
Assume \(\bar{x} \in P\) . Then \(\bar{x} \in \mathrm{ int}\ P\) if and only if
Proof.
Sufficiency. Let point \(\bar{x} \in P\) satisfy (2.8). Using
and
then for any \(x \in \Omega (\delta )\), it holds that
Moreover, sine
it holds that
which and (2.9) together give
Thus x ∈ P. Therefore \(\Omega (\delta ) \subset P\), and \(\bar{x}\) is an interior point of P.
Necessity. Assuming \(\bar{x} \in \mathrm{ int}\ P\), then there is δ > 0 such that \(\Omega (\delta ) \subset P\); also there is \(p \in \bar{ J}\) such that \(\bar{x}_{p} = 0\). According to the definition of \(\bar{J}\), there is x′ ∈ P such that x′ p > 0. It is clear that for any α > 0 there exists
Hence, when α is sufficiently small, it holds that
In addition, it is clear that \(x^\prime_{j},\,\bar{x}_{j} = 0,\ j \in J\), and hence x j = 0, j ∈ J. Therefore \(x \in \Omega (\delta )\). On the other hand, from (2.10), \(\bar{x}_{p} = 0\) and α > 0, x′ p > 0 it follows that
which contradicts \(\Omega (\delta ) \subset P\). Therefore, (2.8) holds if \(\bar{x} \in \mathrm{ int}\ P\). □
Proposition 2.1.3.
If dim P ≥ 1, then P has a relative interior point.
Proof.
The assumption of the proposition implies \(\bar{J}\neq \varnothing \), because otherwise it holds that J = A and hence dim P = 0, leading to contradiction. On the other hand, for any \(j \in \bar{ J}\) there exists x ∈ P such that x j > 0; hence from the convexity of P, it follows that there is x ∈ P satisfying \(x_{j} > 0,\ j \in \bar{ J}\). According to Theorem 2.1.4, it is known that x ∈ int P ≠ ∅. □
Example 2.1.1.
Investigate the interior of the feasible region of the following problem:
Answer Adding the second constraint equality to the third gives
It is clear that the feasible region is nonempty, and the x 3 component of all feasible points equals 0. Eliminating x 3 from the problem leads to
the interior of the feasible region of which is clearly nonempty, corresponding to the relative interior of the feasible region of the original problem.
2 Geometric Structure of the Feasible Region
This section will discuss the geometric structure of the feasible region P. Actually, most results in this section are also valid for general polyhedral convex sets.
Definition 2.2.1.
Let P′ be a nonempty convex subset of P. It is a face of P if for any x ∈ P′, satisfying x ∈ (y, z) ⊂ P, it holds that y, z ⊂ P′.
The preceding means that a face includes the whole segment if it includes a interior point of any segment of P.
The following renders a mathematical expression of P.
Theorem 2.2.1.
A nonempty convex subset P(Q) of P is its face if and only if there exists index set Q ⊂ A such that
Proof.
Sufficiency. Let ∅ ≠ P(Q) ⊂ P be defined by (2.11). If v ∈ P(Q) is an interior point of segment (y, z) and y, z ∈ P, then there exists 0 < α < 1 such that
Hence from α > 0, 1 −α > 0, y, z ≥ 0 and
it follows that
Therefore, y, z ∈ P(Q), and hence P(Q) is a face of P.
Necessity. Let P(Q) ≠ ∅ be a face of P. Introduce notation
Now we show that P(Q) is equivalent to
It is clear that P(Q) ⊂ P(Q)′ ⊂ P.
If P(Q) includes the origin 0 only, it follows that b = 0 and Q = A. And P(Q)′ clearly includes 0, hence P(Q) = P(Q)′. Now assuming that P(Q) does not include 0, we will show P(Q)′ ⊂ P(Q).
Assume x ∈ P(Q)′. If x = 0 (b = 0), then for any 0 ≠ v ∈ P(Q) and
it holds that
which means that v ∈ (y, 0). Since P(Q) is a face and y, 0 ∈ P, we have x = 0 ∈ P(Q). On the other hand, if x ≠ 0, i.e.,
is nonempty, then from x ∈ P(Q)′ and (2.13), it is known that j ∉ Q for any j ∈ S. Therefore there exists u ∈ P(Q) such that u j > 0. If S includes two or more indices, then u, w ∈ P(Q) satisfy u i , w j > 0 for any i, j ∈ S, hence \(z = u/2 + w/2 \in P(Q)\) satisfies z i , z j > 0. This means that there exists v ∈ P(Q) such that v j > 0 for all j ∈ S. As for the relation between x and v, there are the following two cases only:
-
(i)
\(\{j \in S\ \vert \ x_{j} > v_{j}\} = \varnothing \). It is clear that
$$\displaystyle{z = 2v - x = v + (v - x) \in P.}$$Since P(Q) is a face of P, and \(v = z/2 + x/2 \in (x,z)\), it holds that x ∈ P(Q).
-
(ii)
\(\{j \in S\ \vert \ x_{j} > z_{j}\}\neq \varnothing \). Define
$$\displaystyle{z = x +\beta (v - x),\qquad \beta =\min \{ x_{j}/(x_{j} - v_{j})\ \vert \ x_{j} - v_{j} > 0,j \in S\} > 1.}$$It is easy to verify that z ∈ P, and
$$\displaystyle{v =\alpha z + (1-\alpha )x \in (x,z),\qquad 0 <\alpha = 1/\beta < 1.}$$In addition, P(Q) is a face of P, hence x ∈ P(Q). Therefore P(Q)′ ⊂ P(Q).
□
Clearly, P itself is a face (Q = ∅). If a face P(Q) ≠ P, it is called a proper face. It is easy to show that face P(Q) is a proper face if and only if dim P(Q) < dim P. From the proof of Proposition 2.1.2, it is know that dim P(Q) ≤ n − max{m, | Q | }. Face of face is a face.
If \(\mathrm{dim}\ P(Q) =\mathrm{ dim}\ P - 1\), face P(Q) is called facet of P. An 1-dimensional face is also called edge; an 0-dimensional face is called vertex or extreme point.
It is clear that the feasible region P has infinitely many faces. In fact, it is known that the number of \((n - m - k)\)-dimensional faces of P is no more than C n k (\(k = 1,\ldots,n - m\)); in particular, there exist, at most, an (n − m)-dimensional face (that is P itself), C n m+1 edges and \(C_{n}^{n-m} = C_{n}^{m}\) vertices.
“Vertex” can also be defined by the following alternatively.
Definition 2.2.2.
x is a vertex of P if x ∈ [y, z] leads to x = y or x = z for any y, z ∈ P.
The preceding implies that a vertex is not an interior point of any segment of P. It is clear that a vertex of face is a vertex of P, and that the origin is a vertex if it belongs to P.
Vertex has its distinctive algebraical attribute.
Lemma 2.2.1.
A point x ∈ P is a vertex if and only if columns of A corresponding to its positive components are linearly independent.
Proof.
It might be well to let the first s components of x be great than 0 and let the rest be 0. Assume that \(\bar{x}\) is the subvector consisting of the first s components of x and \(\bar{A}\) is the submatrix consisting of the first s columns of A. Then \(\bar{A}\bar{x} = b\).
Necessity. Let x be a vertex of P. If columns of \(\bar{A}\) are linearly dependent, there is a nonzero vector \(\bar{v}\) such that \(\bar{A}\bar{v} = 0\). Introduce notation
It is clear that for any real α it holds that
Take sufficiently small α > 0, such that \(\bar{y},\bar{z} \geq 0\). Construct vectors y and z, so that the first s components of them respectively constitute \(\bar{y}\) and \(\bar{z}\), and the others are 0. Then it is clear that y, z ∈ P and \(x = y/2 + z/2\). Thus x is not a vertex of P, as leads to a contradiction. Therefore, columns corresponding to all positive components of x are linear independent if x is a vertex of P.
Sufficiency. Assume that columns of \(\bar{A}\) are linearly independent. If x ∈ P is not a vertex, there are two points y, z ∈ P and a real α ∈ (0, 1) such that
from which it is known that the last n − s components of y and z are both 0. Therefore, \(v = x - y\neq 0\) and
which means that columns of \(\bar{A}\) are linearly dependent, as a contradiction. Therefore x is a vertex if columns of \(\bar{A}\) are linearly independent. □
In view of rank(A) = m, the preceding theorem implies that the number of positive components of a vertex of P is no more than m.
Lemma 2.2.2.
\(\bar{x}\) is a vertex of the feasible region if and only if it is a basic feasible solution.
Proof.
It is from Theorem 1.6.2 and Lemma 2.2.1. □
The preceding Lemma says that vertex and basic feasible solution of P are the same; thus, the two may be regarded as geometrical and algebraical names of the same mathematical item. It is usually difficult to determine if a point is a vertex based on the definition itself, while a basic feasible solution can be conveniently determined algebraically. Recall that every canonical form of the constraint system Ax = b corresponds to a basic solution, which is a basic feasible solution if it is nonnegative (Sect. 1.6).
Lemma 2.2.3.
Nonempty feasible region has a vertex.
Proof.
It is clearly the case when P is a single point set. Let P be infinite set, and \(\bar{x} \in P\). If \(\bar{x}\) is not a vertex, there are two distinct points y, z ∈ P and a real \(\bar{\alpha }\in (0,1)\) such that
Thus, a component of \(\bar{x}\) is 0 if and only if the corresponding components of both y, z are 0. Introduce
Then T ≠ ∅, since T = ∅ implies that \(\bar{x} = 0\) is a vertex. It might be well to assume that for some i ∈ { 1, …, n} it holds that z i > y i , and hence i ∈ T. This means that
It is easy to show that redefined
satisfies \(\bar{x} \in P\) and \(\bar{x}_{q} = 0\). Thus | T | for the new \(\bar{x}\) is less than that for the old by 1, at least. Repeating no more than n times, therefore, the preceding process terminates at a vertex. □
The above proof produces a series of feasible points, corresponding to faces, each of which is a proper face of its predecessor, until reaching a 0-dimensional face (vertex). Such a technique for shifting to faces of lower dimensions will often be used.
For any given point x and vector d ≠ 0, set {x +α d ∣ α ≥ 0} is said to be ray (or half-line), emanating from x along the direction of d. It is clear that a ray is an infinite set.
Definition 2.2.3.
A nonzero vector d is an unbounded direction if P includes rays, emanating from all x ∈ P along d.
Two unbounded directions having the same direction are regarded as the same.
Theorem 2.2.2.
Vector d is an unbounded direction of the nonempty feasible region if and only if
Proof.
Sufficiency. With the assumptions, it is easy to verify that for any given x ∈ P and α ≥ 0 it holds that x +α d ∈ P, therefore d ≠ 0 is an unbounded direction.
Necessity. Let d ≠ 0 be an unbounded direction. Thus, there is x ∈ P, satisfying x +α d ∈ P for any α ≥ 0. Hence, from Ax = b and \(A(x +\alpha d) = b\), it follows that Ad = 0. In addition, d ≥ 0 holds, because, otherwise, d has a negative component, and hence the corresponding component of x +α d is negative whenever α > 0 becomes sufficiently large, as contradicts x +α d ∈ P. □
Corollary 2.2.1.
A nonzero vector d is a unbounded direction if P includes the ray, emanating from some x ∈ P along d.
It is clear that any nonnegative linear combination of finitely many unbounded directions is an unbounded direction if the combination coefficients are not all zero. Note that “unbounded direction” is meaningless to an empty P.
Theorem 2.2.3.
The feasible region is unbounded if and only if it has an unbounded direction.
Proof.
Sufficiency is clear, it is only needed to show necessity.
If v ∈ P, then the translation of P, i.e.,
clearly includes the origin, and P is unbounded if and only if C is unbounded. It might be well to assume that 0 ∈ P.
Let S′ = { x k} ∈ P be an unbounded sequence of points. Without loss of generality, assume that
Then the sequence
on the unit sphere is bounded, hence has a cluster point. Letting x be its cluster point, then S″ includes a subsequence converging to x. It might be well to assume
Now it should be shown that x is an unbounded direction of P. Let M be a given positive number. Since \(\|{x}^{k}\| \rightarrow \infty \), there is a positive integer K such that \(\|{x}^{k}\| > M\) or \(M/\|{x}^{k}\| < 1\) when k ≥ K. Introduce
Since P is convex, we have 0, x k ∈ P and y k ∈ (0, x k), and hence y k ∈ P and y k → Mx when k ≥ K. As P is closed, it can be asserted that Mx ∈ P. □
Definition 2.2.4.
An unbounded direction of P is extreme direction if it can not be expressed by a positive linear combination of two distinct unbounded directions.
According to the preceding, that d is an extreme direction means that if there are unbounded directions d′, d″ and positive number σ 1, σ 2 > 0, satisfying \(d =\sigma _{1}d^\prime +\sigma _{2}d^{\prime\prime}\), then there must be d′ = σ d′, where σ > 0. Two extreme directions having the same direction are regarded as the same.
In Fig. 2.2, \({d}^{1},{d}^{2},{d}^{3}\) are unbounded directions of P. d 1 and d 2 are extreme directions, but d 3 is not.
Theorem 2.2.4.
An unbounded direction is extreme direction if and only if the rank of columns, corresponding to its positive components, is less than the number of columns by 1.
Proof.
It might be well to assume that k positive components of unbounded direction d correspond to the set of columns a 1, …, a k . The satisfaction of Ad = 0 implies that the columns are linearly dependent. Denote the rank of the set by is r, then it is clear that r < k. Without loss of generality, assume that the first r columns are linear independent. Introduce B = (a 1, …, a r ). It is clear that
Note that k ≥ 2, because if k = 1, otherwise, then from Ad = 0 it follows that a 1 = 0, as leads to a contradiction.
Necessity. Assume that d is an extreme direction, but r ≠ k − 1, that is, a 1, …, a k−1 are linearly dependent. Thus, there exists point
such that
Clearly, for sufficiently small δ > 0, it holds that
and \(Ad^\prime = Ad^{\prime\prime} = 0\). Therefore, d′, d″ are unbounded directions of P, and hence not of the same direction. But \(d = (d^\prime + d^{\prime\prime})/2\), as contradicts that d is an extreme direction. It therefore holds that \(r = k - 1\).
Sufficiency. Assume \(r = k - 1\). If there exist unbounded directions d′, d″ and σ 1, σ 2 > 0 such that
then zero components of d clearly correspond to zero components of d′ and d″. So, the last n − k components of d′ and d″ are all zero. In addition, since d′, d″ are unbounded directions, it holds that \(Ad^\prime = Ad^{\prime\prime} = 0\) (Theorem 2.2.2), and hence that
Note that d′ k , d″ k > 0; because if d′ k = 0, otherwise, then d′ B = 0, and hence d′ = 0, as is a contradiction. Premultiplying the two sides of the preceding two equalities by B T gives
from which it follows that
Therefore, it holds that
which means that unbounded directions d′ and d″ have the same direction, and hence d is an extreme direction. □
The extreme direction and 1-dimensional face (edge) have close relationship.
Theorem 2.2.5.
A vector is an extreme direction if and only if it is the unbounded direction of a edge.
Proof.
Necessity. Assume that d is an extreme direction. Based on Theorem 2.2.4, assume that columns corresponding to its positive components are
where the first r ≤ m columns are linearly independent. Thus
Since the rank of A is m, there are m − r columns which together with a 1, …, a r form a basis when r < m. Without loss of generality, assume that the first m columns of A constitute a basis, i.e.,
From (2.15) and Ad = 0, it is follows that
hence
Assume that basis B corresponds to the canonical form below:
Since its augmented matrix \((I\ \bar{N}\ \vert \ \bar{b})\) comes from (A | b) = (B N | b) by elementary transformations, and a m+1 is a linear combination of a 1, …, a r , the last m − r components of column \(\bar{a}_{m+1}\) in the canonical form are all zero.
Let \(\bar{x}\) belong to P. If \(\bar{x}_{N} = 0\), then \(\bar{x}\) is the basic feasible solution corresponding to basis B. Now assume, otherwise, that \(\bar{x}_{N}\neq 0\). We will create a new basis B associated with a basic feasible solution by a series of elementary transformations and some solution updating.
Assume that \(\bar{x}_{j} > 0\) holds for some j ∈ N, j ≠ m + 1. Reducing \(\bar{x}_{j}\) and keep the other nonbasic components unchanged, we determine the corresponding value of \(\bar{x}_{B}\) such that \(\bar{x}\) satisfies (2.17), resulting in a new solution \(\bar{x}\). For the column corresponding to \(\bar{x}_{j}\), of the canonical form (2.17), there are only two cases arising:
-
(i)
\(\bar{a}_{j} \geq 0\).
It is clear that \(\bar{x}_{j}\) may decrease to 0 and associated \(\bar{x}_{B}\) remains nonnegative. Thus, setting \(\bar{x}_{j} = 0\) gives a new feasible solution \(\bar{x}\).
-
(ii)
\(\bar{a}_{j}\not\geq 0\).
If the first r components of \(\bar{a}_{j}\) are nonnegative, one of the last m − r components of \(\bar{x}_{B}\) decreases to 0 first (so-called “blocking”) as \(\bar{x}_{j}\) decreases. Assume that the blocking is component i (r + 1 ≤ i ≤ m). Set \(\bar{x}_{j}\) to the according value, and interchange a j and a i to update B and N. By exchanging their indices at the same time, the nonbasic component \(\bar{x}_{j}\) of the new solution \(\bar{x}\) becomes 0.
If some of the first r components of \(\bar{a}_{j}\) are negative, then we can determine a σ > 0 such that the first r basic components of
$$\displaystyle{\bar{x}:=\bar{ x} +\sigma d}$$are sufficiently large and all the nonbasic components remain unchanged, except for \(\bar{x}_{m+1}\), such that no broking happens to the first r components of \(\bar{x}_{B}\) as \(\bar{x}_{j}\) decreases. If no broking happens to the last m − r components either, we set \(\bar{x}_{j} = 0\). If broking happens to component r + 1 ≤ i ≤ m, we set \(\bar{x}_{j}\) to the according value, and interchange a j and a i to updated B and N. Consequently, by exchanging their indices, the nonbasic component \(\bar{x}_{j}\) of the new solution \(\bar{x}\) is now equal to 0.
As such, we can transform all \(\bar{x}_{j},\ j \in N,\ j\neq m + 1\) to 0, without affecting the first r indices of B. Then, if \(\bar{x}_{m+1} = 0\), we are done.
If \(\bar{x}_{m+1} > 0\), we reduce it and keep the other nonbasic components unchanged. Since the last m − r components of \(\bar{a}_{m+1}\) are zero, the corresponding components of \(\bar{x}_{B}\) remain unchanged, and there will be only two cases arising:
-
(i)
The first r components of \(\bar{a}_{m+1}\) are all nonnegative. Then it is clear that \(\bar{x}_{m+1}\) can decrease to 0 and according \(\bar{x}_{B}\) remain nonnegative, thus we set \(\bar{x}_{m+1} = 0\).
-
(ii)
Some of the first r components of \(\bar{a}_{m+1}\) are negative. If the according \(\bar{x}_{B}\) remains nonnegative as \(\bar{x}_{m+1}\) decreases to 0, we set \(\bar{x}_{m+1} = 0\); otherwise, if broking happens for component 1 ≤ i ≤ r of \(\bar{x}_{B}\), we set \(\bar{x}_{m+1}\) to the associated value, and exchange a m+1 and a i to update B and N. Then, by exchanging their indices, the nonbasic component \(\bar{x}_{m+1}\) of the new \(\bar{x}\) is equal to zero.
Therefore, it might be well to assert that the basis B, defined by (2.16), corresponds to basic feasible solution \(\bar{x}\).
Consider the following face
It is clear that \(\bar{x} \in P^\prime\). Hence, from (2.15) and
it is known that
Therefore, d is a unbounded direction of P′. It is now only needed to show dim P′ = 1.
For any x′ ∈ P′ and \(x^\prime\neq \bar{x}\), introduce \(d^\prime = x^\prime -\bar{ x}\). It is clear that
and
From (2.19) and (2.21) it follows respectively that
Therefore, \(d^\prime = (d^\prime_{m+1}/d_{m+1})d\), where \(d^\prime_{m+1}/d_{m+1} > 0\). This implies that dim P′ = 1. Therefore, P′ is an 1-dimensional face or edge, and d is an unbounded direction of it.
Sufficiency. Assume that d is an unbounded direction of edge P′ (2.18), and hence satisfies (2.19). If d is a positive linear combination of unbounded directions d′, d″ of P, then there exists a correspondence between zero components of d and of d′, d″, and hence d′, d″ are also unbounded directions of P′. Since dim P′ = 1, in addition, d′ and d″ have the same direction. Therefore, d is an extreme direction. □
Lemma 2.2.4.
An unbounded direction that is not an extreme one is a positive linear combination of two unparallel unbounded directions.
Proof.
Without loss of generality, assume that columns a 1, …, a k correspond to positive components of an unbounded direction d, and the first r columns are linearly independent. As d is not an extreme direction, it holds by Theorem 2.2.4 that r < k − 1, or
Introduce matrix \(B_{1} = (a_{1},\ldots,a_{r})\). Within the set of a k+1, …, a n , determine m − r columns, which might be assumed to correspond to \(B_{1} = (a_{k+1},\ldots,a_{k+m-r})\), to construct basis B = (B 1, B 1) (it is clear that \(m - r \leq n - k\)). Then the nonbasic matrix is N = (N 1, N 2), where \(N_{1} = (a_{r+1},\ldots,a_{k}),\ N_{2} = (a_{k+m-r+1},\ldots,a_{n})\). Assume that B corresponds to the canonical form \(x_{B} =\bar{ b} -\bar{ N}x_{N}\). Then d satisfies
or equivalently,
Introduce \(e \in {\mathcal{R}}^{k-r}\) and
Letting \(\epsilon = {(\epsilon _{1},\ldots,\epsilon _{k-r})}^{T} > 0\), then vectors
satisfy \(d = d^\prime/2 + d^{\prime\prime}/2\). It is clear that d′, d″ ≠ 0 and \(Ad^\prime = 0,Ad^{\prime\prime} = 0\). As (2.22) holds, it is known that there exists a sufficiently small ε such that d′, d″ ≥ 0 are unparallel unbounded directions. □
Theorem 2.2.6.
If P has an unbounded direction, it has a extreme direction.
Proof.
Let d be an unbounded direction. Assume that it has k positive components and the rank of the corresponding columns is r ≤ k − 1. If \(r = k - 1\), then d is an extreme direction (Theorem 2.2.4). Otherwise, by Lemma 2.2.4, d can be expressed as
where d′, d″ are unparallel unbounded directions, and σ 1, σ 2 > 0. Without loss of generality, assume that d″ has at least one component greater than the corresponding component of d′. Thus, the following vector
is well-defined. Consider
where the first two equalities clearly hold. If the third does not hold, then \(d^{\prime\prime} -\alpha (d^{\prime\prime} - d^\prime) = 0\), as leads to
implies that d′, d″ are parallel, as a contradiction. Therefore, the third equality also holds, and hence \(\tilde{d}\) is an unbounded direction. In addition, it is known that the number of zero components of \(\tilde{d}\) is less than that of d by 1, at least. Then set \(d =\tilde{ d}\) and repeat the preceding steps. It is clear that such a process can only repeat finitely many times, and terminates at some extreme direction. □
According to Theorem 2.2.5, extreme directions and unbounded edges of P are 1-to-1 correspondent (extreme directions having the same direction are viewed as the same). As there are finitely many edges, the number of extreme directions are finite.
It is now time to lay a theoretical basis to Dantzig-Wolfe decomposition method for solving large-scale LP problems (Chap. 8).
Theorem 2.2.7 (Representation Theorem of the Feasible Region).
Let P be nonempty. Assume that {u 1 ,…,u s } is the vertex set and \(\{{v}^{1},\ldots,{v}^{t}\}\) the extreme direction set. Then x ∈ P if and only if
Proof.
When (2.23) holds, it is clear that x ∈ P. So it is only needed to show necessity. We use inductive method to dimensions.
If dim P = 0, P is a single point set, including a vertex. The conclusion holds clearly. Assume that it holds for dim P < k. We will show that it holds for dim P = k ≥ 1.
From Proposition 2.1.3, it follows that int P ≠ ∅. Assume x ∈ int P, and consider
Note that u 1 ≠ x. There are the following two cases arising:
-
(i)
u 1 − x ≰ 0.
Determine λ such that
Then, it is easy to verify that x′, defined by (2.24), satisfies x′ ∈ P and x′ q = 0. Therefore, x′ belongs to some proper face with its dimension less than k. According to the assumption of induction, x′ can be expressed as the sum of a convex combination of vertices and a nonnegative combination of extreme directions of the proper face. Since vertices and extreme directions of a face are also that of P, therefore, x′ can be expressed as the sum of a convex combination of vertices and a nonnegative combination of extreme directions of P, i.e.,
where u i and v j are vertices and extreme directions of P, respectively. Substituting the preceding into
which is equivalent to (2.24), leads to the expression of the form of (2.23), i.e.,
-
(ii)
u 1 − x ≤ 0.
Then x′, defined by (2.24), are all feasible points for any λ ≥ 0, hence \(-({u}^{1} - x)\) is an unbounded direction. According to Lemma 2.2.6, there is an extreme direction, say v 1. Now take a sufficiently large μ such that
has at least one component greater than the corresponding component of x. Consequently, the point defined by
is a feasible point, satisfying x′ q = 0. Therefore, this point belongs to some proper face with dimension less than k, and hence can be expressed as the sum of a convex combination of vertices and a nonnegative combination of extreme directions of P. As a result, an expression of the form (2.23) can be obtained in an analogous manner as case (i). □
The preceding Theorem indicates that a feasible point is the sum of a convex combination of vertices and a nonnegative combination of extreme directions, and vice versa. In particular, the following is a direct corollary.
Corollary 2.2.2.
Let the feasible region be bounded. A point is feasible if and only if it is a convex combination of vertices.
3 Optimal Face and Vertex
We describe a basic result without proof (see, e.g., Rockafellar 1997).
Theorem 2.3.1 (Partition Theorem).
Let \(\bar{x}\) be a boundary point of convex set S. Then there exists a superplane including \(\bar{x}\) and partitioning the total space to two half spaces, one of which includes S.
The superplane involved in the preceding Theorem is said to be supporting superplane of S. That is to say, there is a supporting superplane through every boundary point of a convex set. Although the result is applicable to any convex set, we are only concerned with the feasible region P, in particular.
Supporting superplane of P is closely related to its face, as the following reveals.
Lemma 2.3.1.
The intersection of P and a supporting superplane is a face.
Proof.
Assume that \(H =\{ x \in {\mathcal{R}}^{n}\ \vert \ {a}^{T}x =\eta \}\) is the superplane of P, and P′ = P ∩ H. Without loss of generality, let a T x ≤ η hold for all x ∈ P.
Assume that v ∈ P′ is an interior point of segment (y, z), and y, z ∈ P, i.e.,
where 0 < α < 1. Note that, P′ is a nonempty convex set.
It is only needed to show y, z ∈ H, as leads to y, z ∈ P′, hence P′ is a face of P.
Assume y, z ∉ H. Then it holds that
Multiplying the preceding two formulas respectively by α > 0 and 1 −α > 0, and then adding the results gives
combining which and (2.25) leads to v ∉ H, and hence v ∉ P′. This contradicts the assumption v ∈ P′. Therefore, at least one of y and z belongs to H.
Without loss of generality, assume z ∈ H. Then, it follows from (2.25) that
implying that y is in the straight line through v and z. Moreover, z, v ∈ H and H is a superplane, therefore y ∈ H. □
Lemma 2.3.2.
Let \(\bar{f}\) be the optimal value of the standard LP problem. Set F is the set of optimal solutions if and only if it is the intersection of P and the objective contour plane
Proof.
Assume \(F = P \cap \bar{ H}\). It is clear that any optimal solution \(\bar{x} \in P\) satisfies \({c}^{T}\bar{x} =\bar{ f}\), implying \(\bar{x} \in \bar{ H}\), hence \(\bar{x} \in F\). Therefore, F is the set of optimal solutions. If F is the set of optimal solutions, conversely, then \({c}^{T}\bar{x} =\bar{ f}\) holds for any \(\bar{x} \in F \subset P\). Therefore \(\bar{x} \in \bar{ H}\), and hence \(\bar{x} \subset P \cap \bar{ H}\). □
It is clear that \(\bar{H}\) is a supporting superplane of P, as is referred to as objective supporting superplane.
A face is optimal if its elements are all optimal solutions. A vertex is optimal if it is an optimal solution.
Lemma 2.3.3.
If there exists an optimal solution, then there exists an optimal face.
Proof.
According to Lemma 2.3.2, a nonempty set of optimal solutions is the intersection of feasible region P and objective contour plane \(\bar{H}\). Therefore it is an optimal face, according to Lemma 2.3.1 and the definition of an optimal face. □
Theorem 2.3.2.
If there exists a feasible solution, there exists a basic feasible solution. If there is an optimal solution, there is a basic optimal solution.
Proof.
By Lemmas 2.2.2 and 2.2.3, it is known that nonempty feasible region has a basic feasible solution. By Lemma 2.3.3, if there is an optimal solution, there is an optimal face, which is a nonempty polyhedral convex set, and hence having an optimal convex or basic optimal solution. □
In presence of optimal solution, there exists an optimal 0-dimensional face (or vertex). In general, there could exist optimal faces of higher dimensions. It is clear that the optimal face of the highest dimension is the set of all optimal solutions, as is referred to as optimal set. After a LP problem is solved by the simplex method, the optimal set can be obtained easily (Sect. 25.2).
3.1 Graphic Approach
A LP problem of 2-dimension can be solved via a graphic approach. To do so, let us return to Example 1.2.2. The shaded area enclosed by polygon OABCD in Fig. 1.1 is the feasible region (ignore the straight line \(x + 2y = 10\), at the moment). It is required to determine a point over the area such that the objective function reaches the highest value at the point (Fig. 2.3).
In the figure, the equation \(2x + 5y = 0\) of the contour line of the objection function corresponds to the dashed line OE, going through the origin, all points on which correspond to the same objective value 0. The line’s slope, i.e., the tangent of the angle between it and x axis, is \(-2/5 = -0.4\). Therefore, the corresponding contour line shifts parallel to the upper-right side as the objective value increases from 0. Points in the intersection of the line and the area of OABCD are all feasible points. The parallel shifting should be carried out as far as the intersection remains nonempty to attain the biggest possible objective value. It is seen from the figure that the contour line shifting the farthest is the dashed line BF though vertex B, that is, the “objective supporting plane”. The optimal set, i.e., the intersection of the line and the feasible region includes a single vertex B, corresponding to the basic optimal solution. Consequently, problem (1.2) is solved after measuring coordinates of point B, and calculating the associated objective value.
If the figure or measurement is not relatively accurate, however, the end solution would involve unacceptable errors. For a graphic approach, therefore, it would be better to use coordinate paper, with the help of algebraic calculating. Once B is known to be the optimal vertex, for instance, its coordinates can be obtained by solving the following system of equations:
from which the optimal basic solution \(\bar{x} = 1.5,\ \bar{y} = 3\) to problem (1.2) follows, with the optimal value f = 18. That is to say, the manufacturer should arrange production of 1,500 laths and 3,000 sheep piles daily, gaining 18,000 dollars profit.
The graphic approach is not suitable for cases of n ≥ 3, though it is simple. Even for case of n = 2, in fact, its application is rare seen. However, it still offers some inspiration, as is the topic of Sect. 2.5.
4 Feasible Direction and Active Constraint
Methods for solving mathematical problems fall into two categories: direct and iterative methods. The latter produces a sequence of points by iterations, offering an exact or approximate solution. Methods presented in this book belong to the iterative category.
Line search is the mostly used iterative approach in optimization. At a current point \(\bar{x}\), in each iteration, a new point \(\hat{x}\) is determined along a ray starting from \(\bar{x}\) along a nonzero vector d, i.e.,
where d is referred to as search direction, α > 0 as stepsize. Once the two are available, \(\hat{x}\) can be calculated and then one iteration is complete. Repeating this process yields a sequence of points, until a solution is reached. Formula (2.28) is referred to as line search or iterative scheme.
The determination of a search direction is crucial. In presence of constraints, d should be such that the intersection of the ray (2.28) and the feasible region is nonempty. More precisely, we introduce the following concept:
Definition 2.4.1.
Let P be the feasible region. Assume that \(\bar{x} \in P,\ d\neq 0\). If there exists \(\bar{\alpha }> 0\) such that
d is a feasible direction at \(\bar{x}\).
The preceding is relevant to general constrained optimization problems, including the LP problem. The following are some instances, in conjunction with feasible region P.
Example 2.4.1.
In Fig. 2.4, \(\bar{x}\) is an interior point of P, and hence any direction is feasible at it.
Example 2.4.2.
In Fig. 2.5, \(\bar{x}\) is a boundary point of P. It is seen that d 3, d 4 are feasible directions at \(\bar{x}\), but d 1, d 2 are not.
Example 2.4.3.
In Fig. 2.6, \(\bar{x}\) is a vertex of P. Any direction, e.g., d 4 or d 5, within the angle area between d 6 and d 7 (which are respectively along two sides of P) is feasible at \(\bar{x}\). Vectors \({d}^{1},{d}^{2},{d}^{3}\) are not feasible.
Let d be a feasible search direction. It is possible to maintain feasibility of some new iterate \(\hat{x}\). In order for \(\hat{x}\) to be close to an optimal solution, it is needed to take into account the objective function. To this end, it might be well to consider the following problem:
where m > n, and whose feasible region is
Definition 2.4.2.
Vector d satisfying condition c T d < 0 is a descent direction. If d is also a feasible direction at \(\bar{x} \in P\), it is a feasible descent direction at \(\bar{x}\).
It is clear that d is a feasible descent direction at \(\bar{x}\) if and only if it is a feasible direction and forms an obtuse angle with the objective gradient c. Once such a direction d is available, a stepsize α > 0 can be determined, and hence a new iterate \(\hat{x} \in P\), obtained by (2.28), corresponding to a smaller objective value. Then, one iteration is complete.
It is noticeable that not all constraints affect the determination of a feasible descent direction at a current point.
Definition 2.4.3.
A constraint which is violated, or satisfied as an equality, by the current point is an active constraint.
Aimed at a feasible point, the “active” constraint is usually defined as one satisfied as equality, or binding at the point. But it seems to be useful to include infeasible point by regarding a constraint violated as active.
Let us bring up (2.29) as an example. If \(a_{i}^{T}\bar{x} = b\), then a i T x ≥ b is active at \(\bar{x}\); if, otherwise, \(a_{i}^{T}\bar{x} > b\), then a i T x ≥ b is not active at that point. So, the current point is on a boundary of an active constraint. From the simple instances given in the preceding figures, it is seen that a feasible descent direction can be determined by taking into account active constraints only.
In practice, the preceding definition for active constraint does not come up to expectations, as a current point close to boundary could lead to too small stepsize, and hence insignificant progress. In view of this, Powell (1989) proposes the so-called “ε-active” constraint, where ε is a small positive number. For example, if \(a_{i}^{T}\bar{x} - b \leq \epsilon\), then a i T x ≥ b is an ε-active constraint at \(\bar{x}\); whereas it is not if \(a_{i}^{T}\bar{x} - b >\epsilon\).
The LP problem can be solved by the so-called “active set method”, which is usually used for solving nonlinear programming problems though some scholars prefer it to be a LP problem solver (e.g., Fletcher 1981; Hager 2002). We outline the method in conjunction with problem (2.29) in the remainder of this section.
Assume that the current vertex \(\bar{x}\) is an unique solution to the linear system below:
where \(\mathcal{A}\) is called the active set (of constraints), consisting of n indices of (total or part) active constraints, with linearly independent gradients a i . If \(\bar{x}\) is judged to be optimal under some criterion, we are done; otherwise, some index \(p \in \mathcal{A}\) is selected such that the n − 1 equalities
determines a descent edge (1-dimensional face).
In fact, since the rank of the coefficient matrix of (2.30) is n − 1, the associated homogeneous system
has a solution d such that
It is easily verified that for α ≥ 0, all points on the ray
satisfy (2.30). Under the condition that the objective value is lower bounded over the feasible region, the following stepsize is well defined:
In fact, such an α is the largest stepsize possible to maintain feasibility of \(\hat{x}\). Since \(a_{q}^{T}\hat{x} = b_{q}\), the \(a_{q}^{T}x \geq b_{q}\) is an active constraint at \(\hat{x}\).
Consequently, setting \(\bar{x} =\hat{ x}\) and redefining \(\mathcal{A} = \mathcal{A}\setminus \{p\} \cup \{ q\}\) completes an iteration of the active set method. If α > 0, the new vertex corresponds to a lower objective value. Note however that if there are multiple active constraints at the current vertex \(\bar{x}\), then α defined by (2.30) would vanish, so that the descent edge degenerates to a vertex. That is to say, the “new vertex” actually coincides with the old, though set \(\mathcal{A}\) changes. Such a vertex is called degenerate. In Fig. 2.6, e.g., \(\bar{x}\) is a degenerate vertex, at which the three edges along respective \({d}^{3},{d}^{6},{d}^{7}\) intersect.
The simplex method can be viewed as a special scheme of the active set method, as will be discussed in Sect. 3.9. In history, however, the former emerged before the latter from another path by fully taking advantage of the linear structure.
5 Heuristic Characteristic of Optimal Solution
From the graphic approach demonstrated in Fig. 2.3, it is seen that the optimal solution to the LP problem is attained at a vertex of the feasible region. It is imaginable that if the feasible region has a side going through the vertex, which is parallel to objective contour lines, then the whole side corresponds the optimal set, associated with the same optimal value.
So, the solution key lies on how to determine lines intersecting at an optimal vertex. In other words, it is only needed to know which inequalities are active at an optimal solution. Once these active inequalities are known, what left to do is just to solve an linear system; for the instance in Fig. 2.3, the optimal solution was quickly calculated through solving the system (2.27).
Thereby, we make the observation that normal directions (pointing to the interior of the feasible region) of lines AB and BC, intersecting at the optimal vertex B, form the largest angles with the parallel shifting direction of the contour line BF, among all the lines.
Now turn to the more general minimization problem
where \(A \in {\mathcal{R}}^{m\times n},\,c \in {\mathcal{R}}^{n},\,b \in {\mathcal{R}}^{m},\ m > 1\). Note that constraints here are all inequalities of “ ≥ ” type.
We could imagine analogously in the space of multiple dimensions. Now the constraint inequalities correspond to half spaces, and vertices correspond to intersection points formed by half spaces. What we should do is to examine angles between normal directions and parallel shifting direction of the contour plane. For the minimization problem, the shifting direction is the negative gradient direction of the objective function. This leads to the following plausible statement (Pan 1990).
Proposition 2.5.1 (Heuristic characteristic of optimal solution).
Gradients of active constraints at an optimal solution of a minimization problem tend to form the largest angles with the negative objective gradient.
It is now needed to quantize magnitude of the angles. If we denote the ith row vector of A by \(\bar{a}_{i}^{T}\), the cosine of the angle between the ith constraint gradient and negative objective gradient is then
For simplicity, we ignore the constant factor \(1/\|c\|\) and introduce the following
Definition 2.5.1.
The pivoting-index of the ith constraint is defined as
Then, we are able to compare the angles by pivoting-indices, and hence Proposition 2.5.1 may be reformulated as
Gradients of active constraints at an optimal solution tend to have the smallest pivoting-indices.
Example 2.5.1.
Investigate problem (1.2) via pivoting-indices:
Answer Calculate indices of the constraints, and put them in the following table in the order of increasing pivoting-indices:
Constraints | α i |
---|---|
\(-2x - 3y \geq -12\) | − 5. 26 |
\(-y \geq -3\) | − 5. 00 |
\(-x - y \geq -5\) | − 4. 95 |
x ≥ 0 | 2. 00 |
y ≥ 0 | 5. 00 |
From the preceding table, it is seen that \(-2x - 3y \geq -12\) and \(-y \geq -3\) are the two constraints with the smallest pivoting-indices. Thus, the two are active constraints at an optimal solution. This immediately leads to solving system (2.27), as coincides with the outcome from the graphic approach.
If Proposition 2.5.1 were true in general, solving the LP problem amounts to solving a system of linear equations by O(m 3) basic arithmetics, in contrast to existing iterative methods (see Sect. 3.8)! Unfortunately, the characteristic of optimal solution is only heuristic. Counterexamples are easily constructed. If adding a constraint x + 2y ≤ 10 to (1.2) (see Fig. 2.3), e.g., it is then clear that the added constraint is not active at the optimal solution, though its gradient forms the largest angle with the objective gradient (with pivoting-index − 5. 37); in fact, added is a redundant constraint, not affecting the feasible region at all. It is also not difficult to construct counterexamples without redundant constraint. Someday in the Fall of 1986 when the author was visiting the Mathematics Department of University of Washington, after he talked with Professor Rockafellar about his idea on the characteristic of an optimal solution, the latter quickly he showed a counterexample by sketching on a piece of paper, as is seen in Fig. 2.7.
For all that, Proposition 2.5.1 might still offer some clue toward optimal solution, shedding a light on LP research. Such a trick well be referred to as “the most-obtuse-angle heuristics” in this book.
In some cases, in fact, it is possible to detect unboundedness of a problem simply from signs of pivoting-indices.
Theorem 2.5.1.
Assume that the feasible region is nonempty. If pivoting-indices of constraints are all nonnegative, then problem (2.31) is unbounded.
Proof.
By contradiction. Assume that the problem is bounded under the assumptions. It follows that
In fact, there is a vector v such that
Thus, for any α ≥ 0 and feasible solution \(\bar{x}\), vector \(x =\bar{ x} +\alpha v\) satisfies Ax ≥ b, that is, v is an unbounded direction. Further, it is known from (2.35) that
which contradicts that (2.31) is upper bounded. Therefore, (2.34) holds.
Thus, according to Farkas’ Lemma 2.1, there is y ≥ 0 such that
premultiplying which by − c T gives
Hence Ac ≥ 0 follows from nonnegativeness of pivoting-indices. This implies that the right-hand side of the preceding is less than or equal to 0, as is a contradiction. Therefore, the problem is unbounded. □
An alternative proof of the preceding Theorem is via showing − c to be an unbounded direction of the feasible region. The following Corollary gives a necessary condition for the existence of an optimal solution.
Corollary 2.5.1.
If there is an optimal solution, then there is at least one constraint bearing negative pivoting-index.
Example 2.5.2.
Investigate the following LP problem by pivoting-indices:
Answer Calculate indices of the constraints, and fill in the right-hand side of the preceding table. According to Theorem 2.5.1, the problem is unbounded since all indices are nonnegative (see Fig. 2.8).
References
Abadie J, Corpentier J (1969) Generalization of the Wolfe reduced gradient method to the case of mon-linear constrained optimization. In: Fletcher R (ed) Optimization. Academic, London, pp 37–48
Abel P (1987) On the choice of the pivot columns of the simplex method: gradient criteria. Computing 38:13–21
Adler I, Megiddo N (1985) A simplex algorithm whose average number of steps is bounded between two quadratic functions of the smaller dimension. J ACM 32:871–895
Adler I, Resende MGC, Veige G, Karmarkar N (1989) An implementation of Karmarkar’s algorithm for linear programming. Math Program 44:297–335
Andersen E, Andersen K (1995) Presolving in linear programming. Math Program 71:221–245
Andersen ED, Gondzio J, Mészáros C, Xu X (1996) Implementation of interior-point methods for large scale linear programming. In: Terlaky T (ed) Interior point methods of mathematical programming. Kluwer, Dordrecht
Andrel N, Barbulescu M (1993) Balance constraints reduction of large-scale linear programming problems. Ann Oper Res 43:149–170
Anstreicher KM, Watteyne P (1993) A family of search directions for Karmarkar’s algorithm. Oper Res 41:759–767
Arrow KJ, Hurwicz L (1956) Reduction of constrained maxima to saddle-point problems. In: Neyman J (ed) Proceedings of the third Berkeley symposium on mathematical statistics and probability, vol 5. University of California Press, Berkeley, pp 1–26
Avis D, Chvatal V (1978) Notes on Bland’s pivoting rule. Math Program 8:24–34
Balas E (1965) An additive algorithm for solving linear programs with zero-one variables. Oper Res 13:517–546
Balinski ML, Gomory RE (1963) A mutual primal-dual simplex method. In: Graves RL, Wolfe P (eds) Recent advances in mathematical programming. McGraw-Hill, New York
Balinski ML, Tucker AW (1969) Duality theory of linear problems: a constructive approach with applications. SIAM Rev 11:347–377
Barnes ER (1986) A variation on Karmarkars algorithm for solving linear programming problems. Math Program 36:174–182
Bartels RH (1971) A stabilization of the simplex method. Numer Math 16:414–434
Bartels RH, Golub GH (1969) The simplex method of linear programming using LU decomposition. Commun ACM 12:266–268
Bartels RH, Stoer J, Zenger Ch (1971) A realization of the simplex method based on triangular decompositions. In: Wilkinson JH, Reinsch C (eds) Contributions I/II in handbook for automatic computation, volume II: linear algebra. Springer, Berlin/London
Bazaraa MS, Jarvis JJ, Sherali HD (1977) Linear programming and network flows, 2nd edn. Wiley, New York
Beale EML (1954) An alternative method for linear programming. Proc Camb Philos Soc 50:513–523
Beale EML (1955) Cycling in the dual simplex algorithm. Nav Res Logist Q 2:269–275
Beale E (1968) Mathematical programming in practice. Topics in operations research. Pitman & Sons, London
Benders JF (1962) Partitioning procedures for solving mixed-variables programming problems. Numer Math 4:238–252
Benichou MJ, Cautier J, Hentges G, Ribiere G (1977) The efficient solution of large scale linear programming problems. Math Program 13:280–322
Bixby RE (1992) Implementing the simplex method: the initial basis. ORSA J Comput 4:287–294
Bixby RE (1994) Progress in linear programming. ORSA J Comput 6:15–22
Bixby RE (2002) Solving real-world linear problems: a decade and more of progress. Oper Res l50:3–15
Bixby RE, Saltzman MJ (1992) Recovering an optimal LP basis from the interior point solution. Technical report 607, Dapartment of Mathematical Sciences, Clemson University, Clemson
Bixby RE, Wagner DK (1987) A note on detecting simple redundancies in linear systems. Oper Res Lett 6:15–17
Bixby RE, Gregory JW, Lustig IJ, Marsten RE, Shanno DF (1992) Very large-scale linear programming: a case study in combining interior point and simplex methods. Oper Res 40:885–897
Björck A, Plemmons RJ, Schneider H (1981) Large-scale matrix problems. North-Holland, Amsterdanm
Bland RG (1977) New finite pivoting rules for the simplex method. Math Oper Res 2:103–107
Borgwardt K-H (1982a) Some distribution-dependent results about the asymptotic order of the arerage number of pivot steps of the simplex method. Math Oper Res 7:441–462.
Borgwardt K-H (1982b) The average number of pivot steps required by the simplex method is polynomial. Z Oper Res 26:157–177
Botsaris CA (1974) Differential gradient methods. J Math Anal Appl 63:177–198
Breadley A, Mitra G, Williams HB (1975) Analysis of mathematica problems prior to applying the simplex algorithm. Math Program 8:54–83
Brown AA, Bartholomew-Biggs MC (1987) ODE vs SQP methods for constrained optimization. Technical report, 179, The Numerical Center, Hatfield Polytechnic
Carolan MJ, Hill JE, Kennington JL, Niemi S, Wichmann SJ (1990) An empirical evaluation of the KORBX algorithms for military airlift applications. Oper Res 38:240–248
Cavalier TM, Soyster AL (1985) Some computational experience and a modification of the Karmarkar algorithm. ISME working paper, The Pennsylvania State University, pp 85–105
Chan TF (1985) On the existence and computation of LU factorizations with small pivots. Math Comput 42:535–548
Chang YY (1979) Least index resolution of degeneracy in linear complementarity problems. Technical Report 79–14, Department of OR, Stanford University
Charnes A (1952) Optimality and degeneracy in linear programming. Econometrica 20:160–170
Cheng MC (1987) General criteria for redundant and nonredundant linear inequalities. J Optim Theory Appl 53:37–42
Chvatal V (1983) Linear programming. W.H. Freeman, New York
Cipra BA (2000) The best of the 20th century: editors name top 10 algorithms. SIAM News 33: 1–2
Coleman TF, Pothen A (1987) The null space problem II. Algorithms. SIAM J Algebra Discret Methods 8:544–562
Cook AS (1971) The complexity of theorem-proving procedure. In: Proceedings of third annual ACM symposium on theory of computing, Shaker Heights, 1971. ACM, New York, pp 151–158
Cottle RW, Johnson E, Wets R (2007) George B. Dantzig (1914–2005). Not AMS 54:344–362
CPLEX ILOG (2007) 11.0 User’s manual. ILOG SA, Gentilly, France
Curtis A, Reid J (1972) On the automatic scaling of mtrices for Gaussian elimination. J Inst Math Appl 10:118–124
Dantzig GB (1948) Programming in a linear structure, Comptroller, USAF, Washington, DC
Dantzig GB (1951a) Programming of interdependent activities, mathematical model. In: Koopmas TC (ed) Activity analysis of production and allocation. Wiley, New York, pp 19–32; Econometrica 17(3/4):200–211 (1949)
Dantzig GB (1951b) Maximization of a linear function of variables subject to linear inequalities. In: Koopmans TC (ed) Activity analysis of production and allocation. Wiley, New York, pp 339–347
Dantzig GB (1951c) A proof of the equivalence of the programming problem and the game problem. In: Koopmans T (ed) Activity analysis of production and allocation. Wiley, New York, pp 330–335
Dantzig GB (1963) Linear programming and extensions. Princeton University Press, Princeton
Dantzig GB (1991) linear programming. In: Lenstra JK, Rinnooy Kan AHG, Schrijver A (eds) History of mathematical programming. CWI, Amsterdam, pp 19–31
Dantzig GB, Ford LR, Fulkerson DR (1956) A primal-dual algorithm for linear programs. In: Kuhn HK, Tucker AW (eds) Linear inequalities and related systems. Princeton University Press, Princeton, pp 171–181
Dantzig GB, Orchard-Hays W (1953) Alternate algorithm for the revised simplex method using product form for the inverse. Notes on linear programming: part V, RM-1268. The RAND Corporation, Santa Monica
Dantzig GB, Orchard-Hayes W (1954) The product form for the inverse in the simplex method. Math Tables Other Aids Comput 8:64–67
Dantzig GB, Thapa MN (1997) Linear programming 1: introduction. Springer, New York
Dantzig GB, Thapa MN (2003) Linear programming 2: theory and extensions. Springer, New York
Dantzig GB, Wolfe P (1960) Decomposition principle for linear programs. Oper Res 8:101–111
Dantzig GB, Orden A, Wolfe P (1955) The generalized simplex method for minimizing a linear form under linear inequality constraints. Pac J Math 5:183–195
de Ghellinck G, Vial J-Ph, Polynomial (1986) Newton method for linear programming. Algorithnica 1:425–453. (Special issue)
Dikin I (1967) Iterative solution of problems of linear and quadratic programming. Sov Math Dokl 8:674–675
Dikin I (1974) On the speed of an iterative process. Upravlyaemye Sistemi 12:54–60
Dorfman R, Samuelson PA, Solow RM (1958) Linear programming and economic analysis. McGraw-Hill, New York
Duff IS, Erisman AM, Reid JK (1986) Direct methods for sparse matrices. Oxford University Press, Oxford
Evtushenko YG (1974) Two numerical methods for solving non-linear programming problems. Sov Math Dokl 15:420–423
Fang S-C (1993) Linear optimization and extensions: theory and algorithms. AT & T, Prentice-Hall, Englewood Cliffs
Farkas J (1902) Uber die Theorie der Einfachen Ungleichungen. Journal für die Reine und Angewandte Mathematik 124:1–27
Fiacco AV, Mccormick GP (1968) Nonlinear programming: sequential unconstrained minimization techniques. Wiley, New York
Fletcher R (1981) Practical methods of optimization. Volume 2: constrained optimization. Wiley, Chichester
Ford Jr LR, Fulkerson DR (1956) Maximal flow through a network. Can J Math 8:399–407
Forrest JJH, Goldfarb D (1992) Steepest edge simplex algorithm for linear programming. Math Program 57:341–374
Forrest J, Tomlin J (1972) Updating triangular factors of the basis to maintain sparsity in the product form simplex method. Math Program 2:263–278
Forsythe GE, Malcom MA, Moler CB (1977) Computer methods for mathematical computations. Pretice-Hall, EnglewoodCliffs
Fourer R (1979) Sparse Gaussian elimination of staircase linear systems. Tehchnical report ADA081856, Calif Systems Optimization LAB, Stanford University
Fourier JBJ (1823) Analyse des travaux de l’Academie Royale des Science, pendant l’Iannee, Partie mathematique, Histoire de l’Acanemie Royale des Sciences de l’nstitut de France 6 [1823] (1826), xxix–xli (partially reprinted as: Premier extrait, in Oeuvres de Fourier, Tome 11 (Darboux G, ed.), Gauthier-Villars, Paris, 1890, (reprinted: G. Olms, Hildesheim, 1970), pp 321–324
Frisch KR (1955) The logarithmic potentical method of convex programming. Memorandum, University Institute of Economics, Oslo
Fulkerson D, Wolfe P (1962) An algorithm for scaling matrices. SIAM Rev 4:142–146
Gale D, Kuhn HW, Tucker AW (1951) Linear programming and the theory of games. In: Koopmans T (ed) Activity analysis of production and allocation. Wiley, New York, pp 317–329
Gass SI (1985) Linear programming: methods and applications. McGraw-Hill, New York
Gass SI, Saaty T (1955) The computational algorithm for the parametric objective function. Nav Res Logist Q 2:39–45
Gay DM (1978) On combining the schemes of Reid and Saunders for sparse LP bases. In: Duff IS, Stewart GW (eds) Sparse matrix proceedings. SIAM, Philadelphia, pp 313–334
Gay D (1985) Electronic mail distribution of linear programming test problems. COAL Newsl 13:10–12
Gay DM (1987) A variant of Karmarkar’s linear programming algorithm for problems in standard form. Math Program 37:81–90
Geoffrion AM (1972) Generalized Benders decomposition. JOTA 10:137–154
George A, Liu W-H (1981) Computing solution of large sparse positive definite systems. Prentice-Hall, Engleewood Cliffs
Gill PE, Murray W (1973) A numerically stable form of the simplex algorithm. Linear Algebra Appl 7:99–138
Gill PE, Murray W, Saunders MA, Tomlin JA, Wright MH (1985) On projected Newton barrier methods for linear programming and an equivalence to Karmarkar’s projected method. Technical report SOL 85–11, Department of Operations Research, Stanford University
Gill PE, Murray W, Saunders MA, Tomlin JA, Wright MH (1986) On projected Newton methods for linear programming and an equivalence to Karmarkar’s projected method. Math Program 36:183–209
Gill PE, Murray W, Saunders MA, Wright MH (1987) Maintaining LU factors of a general sparse matrix. Linear Algebra Appl 88/89:239–270
Gill PE, Murray W, Saunders MA, Wright MH (1989) A practical anti-cycling procedure for linearly constrainted optimization. Math Program 45:437–474
Goldfarb D (1977) On the Bartels-Golub decomposition for linear programming bases. Math Program 13:272–279
Goldfarb D, Reid JK (1977) A practicable steepest-edge simplex algorithm. Math Program 12:361–371
Goldman AJ, Tucker AW (1956a) Polyhedral convex cones. In: Kuhn HW, Tucker AW (eds) Linear inequalities and related systems. Annals of mathmatical studies, vol 38. Princeton University Press, Princeton, pp 19–39
Goldman AJ, Tucker AW (1956b) Theory of linear programming. In: Kuhn HW, Tucker AW (eds) Linear inequalities and related systems. Annals of mathmatical studies, vol 38. Princeton University Press, Princeton, pp 53–97
Golub GH (1965) Numerical methods for solving linear least squares problems. Numer Math 7:206–216
Golub GH, Van Loan CF (1989) Matrix computations, 2edn. The Johns Hopkins University Press, Baltimore
Gomory AW (1958) Outline of an algorithm for integer solutions to linear programs. Bull Am Math Soc 64:275–278
Gonzaga CC (1987) An algorithm for solving linear programming problems in O(n 3 L) operations. Technical Report UCB/ERL M87/10, Electronics Research Laboratory, University of California, Berkeley
Gonzaga CC (1990) Convergence of the large step primal affine-scaling algorithm for primal non-degenerate linear problems. Technical report, Department of Systems Engineering and Computer Sciences, COPPE-Federal University of Riode Janeiro
Gould N, Reid J (1989) New crash procedure for large systems of linear constraints. Math Program 45:475–501
Greenberg HJ (1978) Pivot selection tactics. In: Greenberg HJ (ed) Design and implementation of optimization software. Sijthoff and Noordhoff, Alphen aan den Rijn, pp 109–143
Greenberg HJ, Kalan J (1975) An exact update for Harris’ tread. Math Program Study 4:26–29
Guerrero-Garcia P, Santos-Palomo A (2005) Phase I cycling under the most-obtuse-angle pivot rule. Eur J Oper Res 167:20–27
Guerrero-Garcia P, Santos-Palomo A (2009) A deficient-basis dual counterpart of Pararrizos, Samaras ans Stephanides’ primal-dual simplex-type algorithm. Optim Methods Softw 24:187–204
Güler O, Ye Y (1993) Convergence behavior of interior-point algorithms. Math Program 60:215–228
Hadley G (1972) Linear programming. Addison-Wesley, Reading
Hager WW (2002) The dual active set algorithm and its application to linear programming. Comput Optim Appl 21:263–275
Hall LA, Vanderbei RJ (1993) Two-third is sharp for affine scaling. Oper Res Lett 13:197–201
Hamming RW (1971) Introduction to applied numerical analysis. McGraw-Hill, New York
Harris PMJ (1973) Pivot selection methods of the Devex LP code. Math Program 5:1–28
Hattersley B, Wilson J (1988) A dual approach to primal degeneracy. Math Program 42:135–145
He X-C, Sun W-Y (1991) An intorduction to generalized inverse matrix (in Chinese). Jiangsu Science and Technology Press, Nanjing
Hellerman E, Rarick DC (1971) Reinversion with the preassigned pivot procedure. Math Program 1:195–216
Hellerman E, Rarick DC (1972) The partitioned preassigned pivot procedure. In: Rose DJ, Willouhby RA (eds) Sparse matrices and their applications. Plenum, New York, pp 68–76
Hertog DD, Roos C (1991) A survey of search directions in interior point methods for linear programming. Math Program 52:481–509
Hoffman AJ (1953) Cycling in the simplex algorithm. Technical report 2974, National Bureau of Standards
Hu J-F (2007) A note on “an improved initial basis for the simplex algorithm”. Comput Oper Res 34:3397–3401
Hu J-F, Pan P-Q (2006) A second note on ‘A method to solve the feasible basis of LP’ (in Chinese). Oper Res Manag Sci 15:13–15
Hu J-F, Pan P-Q (2008a) Fresh views on some recent developments in the simplex algorithm. J Southeast Univ 24:124–126
Hu J-F, Pan P-Q (2008b) An efficient approach to updating simplex multipliers in the simplex algorithm. Math Program Ser A 114:235–248
Jansen B, Terlakey T, Roos C (1994) The theory of linear programming: skew symmetric self-dual problems and the central path. Optimization 29:225–233
Jansen B, Roos C, Terlaky T (1996) Target-following methods for linear programming. In: Terlaky T (ed) Interior point methods of mathematical programming. Kluwer, Dordrecht
Jeroslow R (1973) The simplex algorithm with the pivot rule of maximizing criterion improvement. Discret Appl Math 4:367–377
Kalantari B (1990) Karmarkar’s algorithm with improved steps. Math Program 46:73–78
Kallio M, Porteus EL (1978) A class of methods for linear programming. Math Program 14:161–169
Kantorovich LV (1960) Mathematical methods in the organization and planning of production. Manag Sci 6:550–559. Original Russian version appeared in 1939
Karmarkar N (1984) A new polynomial time algorithm for linear programming. Combinatorica 4:373–395
Karmarkar N, Ramakrishnan K (1985) Further developments in the new polynomial-time algorithm for linear progrmming. In: Talk given at ORSA/TIMES national meeting, Boston, Apr 1985
Khachiyan L (1979) A polynomial algorithm in linear programming. Doklady Academiia Nauk SSSR 244:1093–1096
Kirillova FM, Gabasov R, Kostyukova OI (1979) A method of solving general linear programming problems. Doklady AN BSSR (in Russian) 23:197–200
Klee V (1965) A class of linear problemminng problems requiring a larger number of iterations. Numer Math 7:313–321
Klee V, Minty GJ (1972) How good is the simplex algorithm? In: Shisha O (ed) Inequalities-III. Academic, New York, pp 159–175
Koberstein A (2008) Progress in the dual simplex algorithm for solving large scale LP problems: techniques for a fast and stable implementation. Comput Optim Appl 41:185–204
Koberstein A, Suhl UH (2007) Progress in the dual simplex method for large scale LP problems: practical dual phase 1 algorithms. Comput Optim Appl 37:49–65
Kojima M, Mizuno S, Yoshise A (1989) A primal-dual interior point algorithm for linear programming. In: Megiddo N (ed) Progress in mathematical programming. Springer, New York, pp 29–47
Kojima M, Megiddo N, Mizuno S (1993) A primal-dual infeasible-interior-point algorithm for linear programming. Math Program 61:263–280
Kortanek KO, Shi M (1987) Convergence results and numerical experiments on a linear programming hybrid algorithm. Eur J Oper Res 32:47–61
Kostina E (2002) The long step rule in the bounded-variable dual simplex method: numerical experiments. Math Methods Oper Res 55:413–429
Kotiah TCT, Steinberg DI (1978) On the possibility of cycling with the simplex method. Oper Res 26:374–376
Kuhn HW, Quandt RE (1953) An experimental study of the simplex method. In: Metropolis NC et al (eds) Eperimental arithmetic, high-speed computing and mathematics. Proceedings of symposia in applied mathematics XV. American Mathematical Society, Providence, pp 107–124
Land AH, Doig AG (1960) An automatic method of solving discrete programming problems. Econometrica 28:497–520
Leichner SA, Dantzig GB, Davis JW (1993) A strictly, improving linear programming phase I algorithm. Ann Oper Res 47:409–430
Lemke CE (1954) The dual method of solving the linear programming problem. Nav Res Logist Q 1:36–47
Li W (2004) A note on two direct methods in linear programming. Eur J Oper Res 158:262–265
Li C, Pan P-Q, Li W (2002) A revised simplex algorithm based on partial pricing pivotrule (in Chinese). J Wenzhou Univ 15:53–55
Li W, Guerrero-Garcia P, Santos-Palomo A (2006a) A basis-deficiency-allowing primal phase-1 algorithm using the most-obtuse-angle column rule. Comput Math Appl 51:903–914
Li W, Pan P-Q, Chen G (2006b) A combined projected gradient algorithm for linear programming. Optim Methods Softw 21:541–550
Llewellyn RW (1964) Linear programming. Holt, Rinehart and Winston, New York
Luo Z-Q, Wu S (1994) A modified predictor-corrector method for linear programming. Comput Optim Appl 3:83–91
Lustig IJ (1990) Feasibility issues in a prinal-dual interior-point method for linear programming. Math Program 49:145–162
Lustig IJ, Marsten RE, Shanno DF (1991) Computational exeperience with a primal-dual interior point method for linear programming. Linear Algebra Appl 152:191–222
Lustig IJ, Marsten RE, Shanno DF (1992) On implementing Mehrotras’s predictor-corrector interior-point for linear programming. SIAM J Optim 2:435–449
Lustig IJ, Marsten R, Shanno D (1994) Interior point methods for linear programming: computational state of the art. ORSA J Comput 6:1–14
Markowitz HM (1957) The elimination form of the inverse and its application to linear programming. Manag Sci 3:255–269
Maros I (1986) A general phase-1 method in linear programming. Eur J Oper Res 23:64–77
Maros I (2003a) A generalied dual phase-2 simplex algorithm. Eur J Oper Res 149:1–16
Maros I (2003b) Computational techniques of the simplex method. International series in operations research and management, vol 61. Kluwer, Boston
Maros I, Khaliq M (2002) Advances in design and implementation of optimization software. Eur J Oper Res 140:322–337
Marshall KT, Suurballe JW (1969) A note on cycling in the simplex method. Nav Res Logist Q 16:121–137
Martin RK (1999) Large scale linear and integer optimization: a unified approach. Kluwer, Boston
Mascarenhas WF (1997) The affine scaling algorithm fails for λ = 0. 999. SIAM J Optim 7:34–46
McShane KA, Monma CL, Shanno DF (1989) An implementation of a primal-dual method for linear programming. ORSA J Comput 1:70–83
Megiddo N (1986a) Introduction: new approaches to linear programming. Algorithmca 1:387–394. (Special issue)
Megiddo N (1986b) A note on degenerach in linear programming. Math Program 35:365–367
Megiddo N (1989) Pathways to the optimal set in linear programming. In: Megiddo N (ed) Progress in mathematical programming. Springer, New York, pp 131–158
Megiddo N, Shub M (1989) Boundary behavior of interior point algorithm in linear programming. Math Oper Res 14:97–146
Mehrotra S (1991) On finding a vertex solution using interior point methods. Linear Algebra Appl 152:233–253
Mehrotra S (1992) On the implementation of a primal-dual interior point method. SIAM J Optim 2:575–601
Mizuno S, Todd MJ, Ye Y (1993) On adaptive-step primal-dual interior-point algorithms for linear programming. Math Oper Res 18:964–981
Monteiro RDC, Adler I (1989) Interior path following primal-dual algorithms: Part I: linear programming. Math Program 44:27–41
Murtagh BA (1981) Advances in linear programming: computation and practice. McGraw-Hill, New York/London
Murtagh BA, Saunders MA (1978) Large-scale linearly constrained optimization. Math Program 14:41–72
Murtagh BA, Saunders MA (1998) MINOS 5.5 user’s guid. Technical report SOL 83-20R, Department of Engineering Economics Systems & Operations Research, Stanford University, Stanford
Murty KG (1983) Linear programming. Wiley, New York
Nazareth JL (1987) Computer solutions of linear programs. Oxford University Press, Oxford
Nazareth JL (1996) The implementation of linear programming algorithms based on homotopies. Algorithmica 15:332–350
Nemhauser GL (1994) The age of optimizaiton: solving large-scale real-world problems. Oper Res 42:5–13
Nemhauser GL, Wolsey LA (1999) Integer and combinatorial optimization. Wiley, New York
Nocedal J, Wright SJ (1999) Numerical optimization. Springer, Berlin
Ogryczak W (1988) The simplex method is not always well behaved. Linear Algebra Appl 109:41–57
Orchard-Hays W (1954) Background development and extensions of the revised simplex method. Report RM 1433, The Rand Corporation, Santa Monica
Orchard-Hays W (1956) Evolution of computer codes for linear programming. Paper P-810, The RAND Corporation, p 2224
Orchard-Hays W (1971) Advanced linear programming computing techniques. McGraw-Hill, New York
Padberg MW (1995) Linear optimization and extensions. Springer, Berlin
Pan P-Q (1982) Differential equaltion methods for unconstarained optimization (in Chinese). Numer Math J Chin Univ 4:338–349
Pan P-Q (1990) Practical finite pivoting rules for the simplex method. OR Spektrum 12:219–225
Pan P-Q (1991) Simplex-like method with bisection for linear programming. Optimization 22:717–743
Pan P-Q (1992a) New ODE methods for equality constrained optimization (I) – equations. J Comput Math 10:77–92
Pan P-Q (1992b) New ODE methods for equality constrained optimization (II) – algorithms. J Comput Math 10:129–146
Pan P-Q (1992c) Modification of Bland’s pivoting rule (in Chinese). Numer Math 14:379–381
Pan P-Q (1994a) A variant of the dual pivot rule in linear programming. J Inf Optim Sci 15:405–413
Pan P-Q (1994b) Composite phase-1 methods without measuring infeasibility. In: Yue M-Y (ed) Theory of optimization and its applications. Xidian University Press, Xian, pp 359–364.
Pan P-Q (1994c) Ratio-test-free pivoting rules for the bisection simplex method. In: Proceedings of national conference on decision making science, Shangrao, pp 24–29
Pan P-Q (1994d) Ratio-test-free pivoting rules for a dual phase-1 method. In: Xiao S-T, Wu F (eds) Proceeding of the third conference of Chinese SIAM. Tsinghua University press, Beijing, pp 245–249
Pan P-Q (1995) New non-monotone procedures for achieving dual feasibility. J Nanjing Univ Math Biquarterly 12:155–162
Pan P-Q (1996a) A modified bisection simplex method for linear programming. J Comput Math 14:249–255
Pan P-Q (1996b) New pivot rules for achieving dual feasibility. In: Wei Z (ed) Theory and applications of OR. Proceedings of the fifth conference of Chinese OR society, Xian, 10–14 Oct 1996. Xidian University Press, Xian, pp 109–113
Pan P-Q (1996c) Solving linear programming problems via appending an elastic constraint. J Southeast Univ (English edn) 12:253–265
Pan P-Q (1997) The most-obtuse-angle row pivot rule for achieving dual feasibility in linear programming: a computational study. Eur J Oper Res 101:164–176
Pan P-Q (1998a) A dual projective simplex method for linear programming. Comput Math Appl 35:119–135
Pan P-Q (1998b) A basis-deficiency-allowing variation of the simplex method. Comput Math Appl 36:33–53
Pan P-Q (1999a) A new perturbation simplex algorithm for linear programming. J Comput Math 17:233–242
Pan P-Q (1999b) A projective simplex method for linear programming. Linear Algebra Appl 292:99–125
Pan P-Q (2000a) A projective simplex algorithm using LU decomposition. Comput Math Appl 39:187–208
Pan P-Q (2000b) Primal perturbation simplex algorithms for linear programming. J Comput Math 18:587–596
Pan P-Q (2000c) On developments of pivot algorithms for linear programming. In: Proceedings of the sixth national conference of operations research society of China, Changsha, 10–15 Oct 2000. Global-Link Publishing, Hong Kong, pp 120–129
Pan P-Q (2004) A dual projective pivot algorithm for linear programming. Comput Optim Appl 29:333–344
Pan P-Q (2005) A revised dual projective pivot algorithm for linear programming. SIAM J Optim 16:49–68
Pan P-Q (2008a) A largest-distance pivot rule for the simplex algorithm. Eur J Oper Res 187:393–402
Pan P-Q (2008b) A primal deficient-basis algorithm for linear programming. Appl Math Comput 198:898–912
Pan P-Q (2008c) Efficient nested pricing in the simplex algorithm. Oper Res Lett 38:309–313
Pan P-Q (2010) A fast simplex algorithm for linear programming. J Comput Math 28(6):837–847
Pan P-Q (2013) An affine-scaling pivot algorithm for linear programming. Optimization 62: 431–445
Pan P-Q, Li W (2003) A non-monotone phase-1 method in linear programming. J Southeast Univ (English edn) 19:293–296
Pan P-Q, Ouiang Z-X (1993) Two variants of the simplex algorithm (in Chinese). J Math Res Expo 13(2):274–275
Pan P-Q, Ouyang Z-X (1994) Moore-Penrose inverse simplex algorithms based on successive linear subprogramming approach. Numer Math 3:180–190
Pan P-Q, Pan Y-P (2001) A phase-1 approach to the generalized simplex algorithm. Comput Math Appl 42:1455–1464
Pan P-Q, Li W, Wang Y (2004) A phase-1 algorithm using the most-obtuse-angle rule for the basis-deficiency-allowing dual simplex method. OR Trans 8:88–96
Pan P-Q, Li W, Cao J (2006a) Partial pricing rule simplex method with deficient basis. Numer Math J Chin Univ (English series) 15:23–30
Pan P-Q, Hu J-F, Li C (2006b) Feasible region contraction interior point algorithm. Appl Math Comput 182:1361–1368
Papadimitriou CH, Steiglitz K (1982) Combinatorial optimization: algorithms and complexity. Prentice-Hall, New Jersey
Perold AF (1980) A degeneracy exploiting LU factorization for the simplex method. Math Program 19:239–254
Powell MJD (1989) A tolerant algorithm for linearly constrained optimization calculations. Math Program 45:547–566
Reid JK (1982) A sparsity-exploiting variant of the Bartels-Golub decomposition for linear programming bases. Math Program 24:55–69
Rockafellar RT (1997) Convext analysis. Princeton University Press, Princeton
Roos C (1990) An exponential example for Terlaky’s pivoting rule for the criss-cross simplex method. Math Program 46:79–84
Roos C, Vial J-Ph (1992) A polynomial method of approximate centers for linear programming. Math Program 54:295–305
Roos C, Terlaky T, Vial J-P (1997) Theory and algorithms for linear programming. Wiley, Chichester
Rothenberg RI (1979) Linear programming. North-Holland, New York
Ryan D, Osborne M (1988) On the solution of highly degenerate linear problems. Math Program 41:385–392
Saigal R (1995) Linear programming. Kluwer, Boston
Santos-Palomo A (2004) The sagitta method for solving linear problems. Eur J Oper Res 157:527–539
Saunders MA (1972) Large scale linear programming using the Cholesky factorization. Technical report STAN-CS-72-152, Stanford University
Saunders MA (1973) The complexity of LU updating in the simplex method. In: Andersen R, Brent R (eds) The complexity of computational problem solving. University Press, St. Lucia, pp 214–230
Saunders MA (1976) A fast and stable implementation of the simplex method using Bartels-Golub updating. In: Bunch J, Rose D (eds) Sparse matrix computation. Academic, New York, pp 213–226
Schrijver A (1986) Theory of linear and integer programming. Wiley, Chichester
Shanno DF, Bagchi A (1990) A unified view of interior point methods for linear programming. Ann Oper Res 22:55–70
Shen Y, Pan P-Q (2006) Dual besiction simplex algorithm (in Chinese). In: Preceedings of the national conference of operatios research society of China, Shenzhen. Globa-Link Informatics, Hong Kong, pp 168–174
Shi Y, Pan P-Q (2011) Higher order iteration schemes for unconstrained optimization. Am J Oper Res 1:73–83
Smale S (1983a) On the average number of steps of the simplex method of linear programming. Math Program 27:241–262
Smale S (1983b) The problem of the average speed of the simplex method. In: Bachem A, Grotschel M, Korte B, (eds) Mathematical programming, the state of the art. Springer, Berlin, pp 530–539
Srinath LS (1982) Linear programming: principles and applications. Affiliated East-West Press, New Delhi
Suhl UH (1994) Mathematical optimization system. Eur J Oper Res 72:312–322
Suhl LM, Suhl UH (1990) Computing sparse LU factorization for large-scale linear programming bases. ORSA J Comput 2:325–335
Suhl LM, Suhl UH (1993) A fast LU-update for linear programming. Ann Oper Res 43:33–47
Sun W, Yuan Y-X (2006) Optimization theory an methods: nonlinear programming. Springer, New York
Swietanowaki A (1998) A new steepest edge approximation for the simplex method for linear programming. Comput Optim Appl 10:271–281
Taha H (1975) Integer programming: theory, applications, and computations. Academic, Orlando
Talacko J V, Rockefeller RT (1960) A Compact Simplex Algorithm and a Symmetric Algorithm for General Linear Programs. Unpublished paper, Marquette University
Tanabe K (1977) A geometric method in non-linear programming. Technical Report 23343-AMD780, Brookhaven National Laboratory, New York
Tanabe K (1990) Centered Newton method for linear prgramming: interior and ‘exterior’ point method, (in Japannese). In: Tone K (ed) New methods for linear programming 3. The Institute of Statistical Mathematics, Tokeo, pp 98–100
Tapia RA, Zhang Y (1991) An optimal-basis identification technique for interior-point linear programming algorithms. Linear Algebra Appl 152:343–363
Terlaky T (1985) A covergent criss-cross method. Math Oper Stat Ser Optim 16:683–690
Terlaky T (1993) Pivot rules for linear programming: a survey on recent theoretical developments. Ann Oper Res 46:203–233
Terlaky T (ed) (1996) Interior point methods of mathematical programming. Kluwer, Dordrecht
Todd MJ (1982) An implementation of the simplex method for linear programming problems with variable upper bounds. Math Program 23:23–49
Todd MJ (1983) Large scale linear programming: geometry, working bases and factorizations. Math Program 26:1–23
Tomlin JA (1972) Modifying trangular factors of the basis in the simplex method. In: Rose DJ, Willoughby RA (eds) Sparse matrices and applications. Plenum, New York
Tomlin JA (1974) On pricing and backward transformation in linear programming. Math Program 6:42–47
Tomlin JA (1975) On scaling linear programming problems. Math Program Study 4:146–166
Tomlin JA (1987) An experimental approach to Karmarkar’s projective method, for linear programming. Math Program 31:175–191
Tsuchiya T (1992) Global convergence property of the affine scaling method for primal degenerate linear programming problems. Math Oper Res 17:527–557
Tsuchiya T, Muramatsu M (1995) Global convergence of a long-step affine-scaling algorithm for degenrate linear programming problems. SIAM J Optim 5:525–551
Tucker AW (1956) Dual systems of homegeneous linear relations. In: Kuhn HW, Tucker AW, Dantzig GB (eds) Linear inequalities and related systems. Princeton University Press, Princeton, pp 3–18
Turner K (1991) Computing projections for the Karmarkar algorithtm. Linear Algebra Appl 152:141–154
Vanderbei RJ, Lagarias JC (1990) I.I. Dikin’s convergence result for the affine-scaling algorithm. Contemp Math 114:109–119
Vanderbei RJ, Meketon M, Freedman B (1986) A modification of Karmarkars linear programming algorithm. Algorithmica 1:395–407
Vemuganti RR (2004) On gradient simplex methods for linear programs. J Appl Math Decis Sci 8:107–129
Wang Z (1987) A conformal elimination-free algorithm for oriented matroid programming. Chin Ann Math 8(B1):16–25
Wilkinson JH (1971) Moden error analysis. SIAM Rev 13:548–568
Wolfe P (1963) A technique for resolving degeneracy in linear programming. J Oper Res Soc 11:205–211
Wolfe P (1965) The composite simplex algorithm. SIAM Rev 7:42–54
Wolsey L (1998) Integer programming. Wiley, New York
Wright SJ (1997) Primal-dual interior-point methods. SIAM, Philadelphia
Xu X, Ye Y (1995) A generalized homogeneous and self-dual algorithm for linear prgramming. Oper Res Lett 17:181–190
Xu X, Hung P-F, Ye Y (1996) A simplified homogeneous and self-dual linear programming algorithm and its implementation. Ann Oper Res 62:151–171
Yan W-L, Pan P-Q (2001) Improvement of the subproblem in the bisection simplex algorithm (in Chinese). J Southeast Univ 31:324–241
Yan A, Pan P-Q (2005) Variation of the conventional pivot rule and the application in the deficient basis algorithm (in Chinese). Oper Res Manag Sci 14:28–33
Yan H-Y, Pan P-Q (2009) Most-obtuse-angle criss-cross algorithm for linear programming (in Chinese). Numer Math J Chin Univ 31:209–215
Yang X-Y, Pan P-Q (2006) Most-obtuse-angle dual relaxation algorithm (in Chinese). In: Preceedings of the national conference of operatios research society of China, Shenzhen. Globa-Link Informatics, Hong Kong, pp 150–155
Ye Y (1987) Eliminating columns in the simplex method for linear programming. Technical report SOL 87–14, Department of Operations Research, Stanford Univesity, Stanford
Ye Y (1990) A ‘build-down’ scheme for linear probleming. Math Program 46:61–72
Ye Y (1997) Interior point algorithms: theory and analysis. Wiley, New York
Ye Y, Todd MJ, Mizuno S (1994) An \(o(\sqrt{n}l)\)-iteration homogeneous and selfdual linear programming algorithm. Math Oper Res 19:53–67
Zhang H, Pan P-Q (2008) An interior-point algorithm for linear programming (in Chinese). In: Preceedings of the national conference of operatios research society of China, Nanjing. Globa-Link Informatics, Hong Kong, pp 183–187
Zhang J-Z, Xu S-J (1997) Linear programming (in Chinese). Science Press, Bejing
Zhang L-H, Yang W-H, Liao L-Z (2013) On an efficient implementation of the face algorithm for linear programming. J Comp Math 31:335–354
Zhou Z-J, Pan P-Q, Chen S-F (2009) Most-obtuse-angle relaxation algorithm (in Chinese). Oper Res Manag Sci 18:7–10
Zionts S (1969) The criss-cross method for solving linear programming problems. Manag Sci 15:420–445
Zlatev Z (1980) On some pivotal strategies in Gaussian elimination by sparse technique. SIAM J Numer Anal 17:12–30
Zörnig P (2006) Systematic construction of examples for cycling in the simplex method. Comput Oper Res 33:2247–2262
Zoutendijk G (1960) Methods of feasible directions. Elsevier, Amsterdam
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
PAN, PQ. (2014). Geometry of the Feasible Region. In: Linear Programming Computation. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-40754-3_2
Download citation
DOI: https://doi.org/10.1007/978-3-642-40754-3_2
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-40753-6
Online ISBN: 978-3-642-40754-3
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)