Abstract
The tableau simplex algorithm requires a feasible simplex tableau to get itself started, whereas the revised simplex algorithm starts from a feasible basis or its inverse. The task of a Phase-I method is to provide such a starting point to Phase-II for achieving optimality. Such a two-Phase methodology has been successful in practice, compared with one-Phase, like the big-M.
Access provided by Autonomous University of Puebla. Download chapter PDF
Keywords
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.
The tableau simplex algorithm requires a feasible simplex tableau to get itself started, whereas the revised simplex algorithm starts from a feasible basis or its inverse. The task of a Phase-I method is to provide such a starting point to Phase-II for achieving optimality. Such a two-Phase methodology has been successful in practice, compared with one-Phase, like the big-M.
Not have been used in practice, the artificial-variable method, presented in Sect. 3.3, is usually seen in textbooks only, because it is so rigid that there is no much choice for an initial basis, as the basis has to involve all the artificial variables, let alone expending problem’s scale. Moreover, some cumbersome treatment has to be carried out subsequently after zero optimal value of the auxiliary program is attained but there are still artificial variables remanning basic.
This chapter will focus on practicable Phase-I methods, which are artificial variable free, or involves a single artificial variable only (for further references, see Pan 1994b; Pan and Li 2003; Pan et al. 2004).
For simplicity of exposition, the conventional pivot rule will be utilized in this chapter. Nevertheless, it is preferable to use the nested largest-distance rule instead, although all rules presented in Chap. 11 are applicable for Phase-I.
1 Infeasibility-Sum Method
The auxiliary objective function of this method only involves variables, associated with negative components of the current basic solution. The method is widely used in practice, as it is artificial variable free, and performs remarkably.
Assume that an initial simplex tableau of form (3.18) is available, associated with basic solution
Introduce notation
Assume that \(I\neq \varnothing\). Construct the following auxiliary program:
where the objective function, called “infeasibility-sum”, is the negative sum of variables, associated with negative components of the basic solution. Note that the constraint system of the auxiliary program is the same as that of the original problem, except the nonnegativity restriction is only imposed on variables, associated with nonnegative components of the basic solution.
It is clear that solution (13.1) is feasible to the auxiliary program. The associated feasible simplex tableau can be obtained by eliminating nonzero entries in the objective row by relevant elementary transformations. Consequently, the reduced objective function is
where
The positiveness of objective value w 0 comes from (13.2).
Theorem 13.1.1 (Infeasibility test).
If reduced costs of an auxiliary simplex tableau are all nonnegative, the original problem is infeasible.
Proof.
Assume that \(\bar{z}_{N} \geq 0\), and \(\tilde{x} \geq 0\) is a feasible solution to the original problem. Substitute \(\tilde{x}\) to (13.4) results in a system, whose left-hand side is clearly less than or equal to zero, and, by w 0 > 0 and \(\bar{z}_{N} \geq 0\), the right-hand side is strictly greater than zero, as is a contradiction. Therefore, the statement is valid. □
If there is a negative reduced cost, then any existing column rule applies. Assume that a column index q was determined such that
As there is no nonnegativity restriction on current infeasible basic variables, the auxiliary program could be unbounded even if the original problem itself bounded. In fact, it is the case when \(\bar{I} = \varnothing\) or the following condition holds:
Thus, the conventional row rule is not applicable. To overcome this difficulty, it is modified as follows.
Rule 13.1.1 (Auxiliary row rule)
Determine a row index p such that
It will be seen a little later the bonus, yielding from use of “max” in the second expression of (13.7).
Lemma 13.1.2.
Assume that the selected column index q satisfies (13.5). Then rule 13.1.1 determines a row index p such that the simplex tableau, resulting from the according basis change, is feasible to the auxiliary program.
Proof.
First, it can be asserted that there is a row index t ∈ I such that \(\bar{a}_{tq} <0\); because otherwise from
it follows that
which contradicts the assumption (13.5). Therefore, the second expression of (13.7) is always well-defined, hence so is the row rule.
Take \(\bar{a}_{pq}\) as the pivot. Executing according elementary transformations leads to a new simplex tableau, whose right-hand side can be expressed in terms of its predecessor’s entries (see (3.14)), i.e.,
where α is determined by (13.7). It is easy to verify that these components are all nonnegative, hence the resulting simplex tableau is again feasible to the auxiliary program. □
Theorem 13.1.3.
Assume feasibility of the original problem and nondegeneracy throughout solution process. The number of basic infeasible variables strictly decreases in finitely many iterations.
Proof.
It is seen from the proof of Lemma 13.1.2 that the second expression of (13.7) is still well-defined if “\(\max\)” there replaced by “\(\min\)”, i.e.,
If, in some iteration, the stepsize α is determined by the first expression of (13.7), and it holds that
then the component, associated with row index s, of the new solution satisfies
which implies that a basic infeasible variable becomes feasible, at least.
If, in some iteration, the second expression of (13.7) is used, then, for the new solution, not only the basic components, associated with all row indices \(i \in \bar{ I}\), are feasible, but so are basic components, associated with row indices
This is of course favorable.
Assume that the number of basic infeasible variables does not strictly decreases forever. In each iteration, there is always a negative cost since the original problem is feasible (Theorem 13.1.1), and the stepsize α is determined by the first expression of (13.7), which does not satisfy (13.9). But the number of basis is finite, hence there are some bases appear infinitely, as cycling occurs. This is impossible because the objective value of the auxiliary program strictly decreases under the nondegeneracy assumption. Therefore the number of basic infeasible variables strictly decreases in finitely many iterations. □
When, after some iteration, the number of basic infeasible variables strictly decreases, but does not vanish, a new auxiliary program is constructed, whose objective function (infeasibility-sum) involves less variables than its predecessor, and iterations are carried out, until infeasibility of the original problem is detected or a feasible simplex tableau is reached. Involving a series of infeasibility-sums as objective functions in general, this method is sometimes referred to as method with piecewise sum of infeasibilities.
The overall steps is put in the following algorithm.
Algorithm 13.1.1 (Tableau Phase-I: infeasibility-sum).
Initial: simplex tableau of form (3.18). This algorithm finds a feasible simplex tableau by handling auxiliary programs of form (13.3).
-
1.
Stop if \(\bar{b}_{i} \geq 0,\forall i = 1,\cdots \,,m\) (a feasible simplex tableau reached).
-
2.
For all \(j \in N\), compute \(\bar{z}_{j} =\sum _{i\in I}\bar{a}_{i,j}\), where \(I =\{ i = 1,\cdots \,,m\ \vert \ \bar{b}_{i} <0\}\).
-
3.
Stop if \(\bar{z}_{N} \geq 0\) (the original problem is infeasible).
-
4.
Form a simplex tableau to the auxiliary program (13.3).
-
5.
Carry out a single iteration of simplex Algorithm 3.2.1, in which Rule 13.1.1 is used for row pivoting instead.
-
6.
Go to step 1.
Example 13.1.1.
Solve the following problem by Algorithm 13.1.1 in Phase-I:
Answer Phase-I. Multiply the second equation by − 1 to turn the coefficient of x 5 to 1. Setting \(I =\{ 1,2\}\), construct auxiliary program
Its initial tableau is
Respectively add the first and the second row to the objective row to eliminate the nonzero basic entries, yielding simplex tableau
Iteration 1:
\(q = 3;\min \{8/1\} = 8,\ p = 3\). So 1 in the x 3 column and the third row is the pivot. Add 2, 4, 5 times of row 3 to rows 1,2,4, respectively.
The right-hand side of the preceding tableau is nonnegative, hence Phase-I is finished. To go to Phase-II, use coefficients of the original objective function to cover the bottom (objective) row.
Add 2 times of row 3 to the bottom row to eliminate nonzero basic entries, obtaining feasible simplex tableau
Carry out Phase-II from the preceding tableau. By the conventional rule, x 1 column is selected as the pivot column. All components of this column are nonpositive, as detects unboundedness of the original problem.
In practice, used is usually the following revised version of Algorithm 13.1.1.
Algorithm 13.1.2 (Phase-I: infeasibility-sum).
Initial: \((B,N),{B}^{-1},\bar{x}_{B}={B}^{-1}b\). This algorithm finds a basic feasible solution to the standard LP problem by handling auxiliary programs of form (13.3).
-
1.
Stop if \(\bar{x}_{B} \geq 0\) (producing a basic feasible solution).
-
2.
Construct c B :
$$\displaystyle{c_{j_{i}} = \left \{\begin{array}{ll} - 1,&\mbox{ If}\ \bar{x}_{j_{i}} <0, \\ 0, &\mbox{ If}\bar{x}_{j_{i}} \geq 0,\end{array} \right.i = 1,\cdots \,,m.}$$ -
3.
Carry out a single iteration of Algorithm 3.5.1 (or 3.5.2), in which Rule 13.1.1 is used for row pivoting.
-
4.
Stop if it terminates as step 3 (or step 2) (the original problem is infeasible).
-
5.
Go to step 1.
2 Single-Artificial-Variable Method
Although the infeasible-sum method is artificial variable free, the row pivot rule used in it is cumbersome, compared with the normal row rule. This section describes a method, which uses the normal row rule and only involves a single artificial variable.
Let (B, N) be basis and nonbasis of the standard problem (1.8). There are two schemes for the method.
Scheme 1. The auxiliary program is based on the original data of the problem.
Given m-dimensional vector h ≥ 0 such that b − Bh ≠ 0. Introducing artificial variable x n+1 and normalized vector
construct the following auxiliary program
It is easy to verify that
is a feasible solution to the program, though not necessarily a basic solution. For instance, setting h = e > 0 seems to be helpful to avoid zero stepsize, though according a n+1 would loose sparsity.
The constraint system of the auxiliary program is equivalent to
where
If \(\bar{b} \geq 0\), then x n+1 is clearly allowed to be zero. Thus, erasing the artificial variable leads to a basic feasible solution to the original problem; otherwise, the following applies.
Proposition 13.2.1.
If \(\bar{b}\not\geq 0\) , row index set \(\{i = 1,\cdots \,,m\ \vert \ \bar{a}_{i,n+1} <0,\bar{b}_{i} <0\}\) is nonempty.
Proof.
We show that if \(\bar{b}_{i} <0\) for some \(i \in \{ 1,\cdots \,,m\}\), then \(\bar{a}_{i,n+1} <0\). Assume it is not the case, i.e.,
Then from the preceding, (13.10) and (13.13) it is known that
hence \(\bar{b}_{i} \geq h_{i} \geq 0\), as a contradiction. □
Now assume \(\bar{b}\not\geq 0\). If \(\bar{x}_{N} = 0\) is fixed and the value of \(\bar{x}_{n+1}\) decreases from \(\|b - Bh\|\), then \(\bar{x}_{B}\) changes from h accordingly. If \(\bar{b}_{i} <0\) for some \(i \in \{ 1,\cdots \,,m\}\), the according inequality constraint may block-up the value of \(\bar{x}_{n+1}\) from decreasing to zero. In fact, the smallest x n+1 value, satisfying inequalities
is
As the value of \(\bar{x}_{n+1}\) decreases, the according value of \(\bar{x}_{j_{r}}\) hits 0 first. Let \(x_{j\,r}\) leave and \(x_{n+1}\) enter the basis, then the resulting basis corresponds to a basic feasible solution to the auxiliary program, with its objective value strictly decreases. As it is lower bounded, solving the auxiliary program will render its optimal solution.
It is not difficult to show that a basic feasible solution or basis to the original problem can be obtained if the optimal value of the auxiliary program is equal to zero (x n+1 leaves the basis). The original problems is infeasible if the optimal value of the auxiliary program is positive.
Scheme 2. The auxiliary program is based on a canonical form of the constraint system.
Assume that the canonical form, associated with basis B, is available and that \(\bar{b} = {B}^{-1}b\not\geq 0\). Giving m-dimensional vector h ≥ 0, and defining
construct the following auxiliary program
It is clear that there is a feasible solution to the program, i.e.,
Thus, determining pivot column index \(q = n + 1\) and row index p satisfying (13.14), one can obtain a feasible basis by performing according basis change. It is noticed that Proposition 13.2.1 is still valid in this case.
As for how to determine h, it is only required that h ≥ 0, in principle. With respect to practice, however, it is not proper for the associated \(\|\bar{a}_{n+1}\|\) being too large. A simple choice is h = e > 0. Another choice is setting
and taking
where \(\delta _{i} \geq 0,i = 1,\cdots \,,m\) are a set of constants, which are equal or close to zero. Consequently, (13.15) becomes
We prefer the latter because it is better for maintaining sparsity of \(\bar{a}_{n+1}\).
More specifically, assume availability of a simplex tableau, say (3.18), of the original problem. Inserting the artificial x n+1 column, we obtain the auxiliary simplex tableau below:
where the original objective row is excluded. Determine row index r by (13.14). Then drop \(x_{j_{r}}\) from and enter x n+1 to the basis by elementary transformations, so that x n+1 becomes the rth basic variable. Taking the rth row as the objective row, the auxiliary program can be solved by the simplex method. If the optimal value vanishes and x n+1 has left the basis, a feasible simplex tableau to the original problem is obtained. If the optimal value is nonzero, the original problem is infeasible.
The steps can be summarized to the following algorithm.
Algorithm 13.2.1 (Tableau Phase-I: single-artificial-variable).
Initial: a simplex tableau of form (13.18). This algorithm finds a feasible simplex tableau to the standard LP problem.
-
1.
Select a row index \(r \in \arg \min \{\bar{ b}_{i}\ \vert \ i = 1,\cdots \,,m\}\).
-
2.
Stop if \(\bar{b}_{r} \geq 0\) (infeasible problem).
-
3.
Determine row index \(r \in \arg \min \{\bar{ b}_{i}/\bar{a}_{i,n+1}\vert \bar{a}_{i,n+1} <0,\bar{b}_{i} <0,i = 1,\cdots \,,m\}\).
-
4.
Carry out elementary transformation to convert \(\bar{a}_{r,\,n+1}\) to 1, and eliminate all other nonzeros of the column.
-
5.
Determine column index \(q \in \arg \max _{j\in N}\bar{a}_{rj}\).
-
6.
Stop if \(\bar{a}_{rj} \leq 0\).
-
7.
Determine row index \(p \in I =\arg \min \{\bar{ b}_{i}/\bar{a}_{i\,q}\vert \bar{a}_{i\,q}> 0;\ i = 1,\cdots \,,m\}\).
-
8.
Set p = r if r ∈ I.
-
9.
Carry out elementary transformations to convert \(\bar{a}_{p\,q}\) to 1, and eliminate all other nonzeros of the column.
-
10.
Stop if p = r (feasibility achieved).
-
11.
Go to step 5.
It is not difficult to transform Algorithm 13.2.1 to a revised version, as is omitted here. We will see that the single-variable auxiliary program can be handled in a relevant and favorable manner in the so-called “reduced simplex” context (Sect. 15.3).
Example 13.2.1.
Find a feasible simplex tableau to the following problem by Algorithm 13.2.1:
Answer The constraint system is of a canonical form. Set \(\delta _{1} =\delta _{2} = 0\). Introducing artificial variable x 7, construct an auxiliary program of form (13.16):
Its initial tableau is (we put the objective row of the original problem at the bottom to turn to Phase-II conveniently when Phase-I is finished).
Iteration 1:
-
1.
\(\min \{-1,2,-4\} = -4 <0\).
-
3.
\(\max \{-1/ - 1,-4/ - 4\} = 1,r = 1\).
-
4.
Multiply row 1 by − 1, then add 4 times of row 1 to row 3:
-
5.
\(\max \{1,-1,1,-1\} = 1> 0,\ q = 1\).
-
7.
\(\min \{1/1,0/5\} = 0,\ p = 3\).
-
8.
p ≠ 1.
-
9.
Multiply row 3 by 1∕5, then add \(-1,3,-1\) times of row 3 to rows 1,2,4, respectively:
Iteration 2:
-
5.
\(\max \{2/5,4/5,-1/5,1/5\} = 4/5> 0,\ q = 3\).
-
7.
\(\min \{1/(4/5),2/(13/5),0/(1/5)\} = 0,\ p = 3\).
-
8.
p ≠ 1.
-
9.
Multiply row 3 by 5, then add \(-4/5,-13/5,11/5\) times of row 3 to rows 1,2,4, respectively:
Iteration 3:
-
5.
\(\max \{-4,6,-1,3\} = 6> 0,\ q = 2\).
-
7.
\(\min \{1/6,2/15\} = 2/15,\ p = 2\).
-
8.
p ≠ 1.
-
9.
Multiply row 2 by 1∕15, then add − 6, 7, 11 times of row 2 to rows 1,3,4, respectively:
Iteration 4:
-
5.
\(\max \{5/6,-1/5,-1/5,-2/5\} = 5/6> 0,q = 1\).
-
7.
\(\min \{(1/5)/(11/5)\} = 1/11,\ p = 1\).
-
8.
p = 1.
-
9.
Multiply row 1 by 5∕6, then add \(13/15,16/15,-6/5,-22/15\) times of row 1 to rows 2,3,4, respectively:
As the artificial variable x 7 has already left the basis, the resulting is a feasible simplex tableau to the original problem.
It is noted that δ 1 and δ 2 above were all set to 0, consequently the first 3 iterations made no progress due to degeneracy. Although parameters \(\delta _{i},i = 1,\cdots \,,m\) are required to be nonnegative theoretically, therefore, they should be set to different positive values practically. Moreover, it seems to be relevant to set δ i ≪ 1 such that \(\|\bar{a}_{n+1}\|\) is not too large.
3 The Most-Obtuse-Angle Column Rule
Surprisingly enough, the conventional pivot rule used in the dual simplex algorithm was found to be useful, with favorable numerical results, for achieving primal feasibility (Pan 1994a); it is to say that the rule is itself applicable even when reduced costs are not nonnegative. This leads to a simple artificial-variable-free variant involving no reduced cost, as turns out to be a counterpart of the most-obtuse-angle row rule (Sect. 14.3).
Assume that a simplex tableau, say (3.18), is infeasible, and a pivot row index p has been selected by the conventional rule
So it holds that \(\bar{b}_{p} <0\).
The key of the presented Phase-I method is to use the following column rule.
Rule 13.3.1 (Most-obtuse-angle column rule)
Determine pivot column index
In case of \(\bar{a}_{p\,q} <0\), after the basis change, the basic infeasible variable \(x_{j_{p}}\) becomes nonbasic (hence feasible), and the pth component of the right-hand side of the tableau changes from negative to positive, that is, \(\hat{b}_{p} =\bar{ b}_{p}/\bar{a}_{p\,q}> 0\).
The steps are put in the following algorithm. A generalized version of it will be derived alternatively in Sect. 20.7.
Algorithm 13.3.1 (Tableau Phase-I: the most-obtuse-angle column rule).
Initial: simplex tableau of form (3.18). This algorithm finds a feasible simplex tableau to the standard LP problem.
-
1.
Select row index \(p \in \arg \min \{\bar{ b}_{i}\ \vert \ i = 1,\cdots \,,m\}\).
-
2.
Stop if \(\bar{b}_{p} \geq 0\).
-
3.
Determine column index \(q \in \arg \min _{j\in N}\bar{a}_{p\,j}\).
-
4.
Stop if \(\bar{a}_{p\,q} \geq 0\).
-
5.
Carry out elementary transformations to convert \(\bar{a}_{p\,q}\) to 1, and eliminate all other nonzeros of the column.
-
6.
Go to step 1.
It is noted that only the right-hand side and the pivot row of the simplex tableau are used in each iteration of the preceding Algorithm.
Theorem 13.3.1.
Assume termination of Algorithm 13.3.1 . It terminates either at
-
(i)
Step 2, obtaining a basic feasible solution; or at
-
(ii)
Step 4, detecting infeasibility of the problem.
Proof.
When it terminates at step 2, it is clear that \(\bar{b} \geq 0\), hence a feasible simplex tableau is achieved. In case of termination at step 4, it holds that \(\bar{b}_{p} <0\) and that \(\bar{a}_{p\,j} \geq 0,\forall j \in N\). By Lemma 3.3.1, there is no feasible solution to the problem. □
The algorithm is not a monotone one, that is, the objective value would not monotonically change in solution process. Finiteness of it is hence not guaranteed even if nondegeneracy is assumed. Although Guerrero-Garcia and Santos-Palomo (2005) offer a cycling instance for this algorithm, it may be still expected that cycling hardly happens in practice.
Geometric meaning of the most-obtuse-angle column rule may be revealed by investigating in the dual space. Let \((\bar{y},\bar{z})\) be the current dual basic solution. The (m + n)-dimensional vector
is an uphill direction. In fact, it is verified that d is in the null space of coefficient matrix of the dual constraint equalities, satisfying
and forms an acute angle with the dual objective gradient, i.e.,
The negative σ q implies that d forms the largest possible obtuse angle with the gradient e m+q of the dual nonnegative constraint z q ≥ 0. Therefore, if the uphill direction is close to the dual objective gradient, then e m+q tends to form the most obtuse angle with the dual objective gradient. According to the heuristic characteristic of optimal basis (Sect. 2.5), it is favorable to let the dual constraint z q ≥ 0 be satisfied as equality, accordingly entering x q to the basis.
We stress that the use of the conventional dual rule in step 1 is just for simplicity, does not mean itself the best choice. In fact, it should be much better to use rules, presented in Chap. 12 instead. From a geometric point of view, in fact, the most-obtuse-angle column rule should be best matched by the dual steepest-edge rule II.
Nevertheless, Rule 13.3.1 does not employ information associated with the objective function at all. Taking into account the extent to which the current vertex violates dual constraints, we suggest the following variant.
Rule 13.3.2 (Variant of most-obtuse-angle column rule)
Given constant 0 < τ ≤ 1. Select pivot column index
In order to widen the range of choice, τ should be close to 1, so that in the case when \(\bar{z}_{N}\) has negative components, the dual constraint inequality, that is violated the most, will be satisfied as equality. It might be suitable to set σ = 0. 95, or so.
An advantage of this artificial-variable-free Phase-I method lie in its good numerical stability and remarkable simplicity. Even though there are no related numerical results available at present, it seems to be promising, as its dual counterpart performs remarkably for solving large and sparse problems in computational experiments (see Sect. 14.3).
Example 13.3.1.
Find a feasible simplex tableau to the following problem by Algorithm 13.3.1:
Answer Initial: Adding the second equality constraint to the third to eliminate x 4. Such doing results in the following simplex tableau.
Iteration 1:
-
1.
\(\min \{-2,3,-1\} = -2 <0,\ p = 1\).
-
3.
\(\min \{2,-1\} = -1 <0,q = 2\).
-
5.
Multiply row 1 by − 1, then add \(-2,-3,1\) times of row 1 to rows 2,3,4, respectively:
Iteration 2:
-
1.
\(\min \{2,-1,-7\} = -7 <0,\ p = 3\).
-
3.
\(\min \{-1,3\} = -1 <0,\ q = 1\).
-
5.
Add 2, 5 times of row 3 to rows 1,2, respectively:
The right-hand side of the above tableau is nonnegative, hence the obtained is a feasible simplex tableau, from which Phase-II can get itself started.
In practice, usually used is the following revised version of the preceding algorithm.
Algorithm 13.3.2 (Phase-I: most-obtuse-angle column rule).
Initial: \((B,N),{B}^{-1},\bar{x}_{B} = {B}^{-1}b\). This algorithm finds a basic feasible solution to the standard LP problem.
-
1.
Select pivot row index \(p \in \arg \min \{\bar{ x}_{j_{i}}\ \vert \ i = 1,\cdots \,,m\}\).
-
2.
Stop if \(\bar{x}_{j_{p}} \geq 0\) (feasibility achieved).
-
3.
Compute \(h = {B}^{\mathrm{-T}}e_{p},\,\sigma _{N} = {N}^{\mathrm{T}}h\).
-
4.
Determine pivot column index \(q \in =\arg \min _{j\in N}\sigma _{j}\).
-
5.
Stop if \(\sigma _{q} \geq 0\) (infeasible problem).
-
6.
Compute \(\bar{a}_{q} = {B}^{-1}a_{q}\) and \(\alpha =\bar{ x}_{j_{p}}/\sigma _{q}\).
-
7.
Set \(\bar{x}_{q} =\alpha\), and update:\(\bar{x}_{B} =\bar{ x}_{B} -\alpha \bar{ a}_{q}\) if \(\alpha \neq 0\).
-
8.
Update \({B}^{-1}\) by (3.23).
-
9.
Update \((B,N)\) by exchanging \(j_{p}\) and \(q\).
-
10.
Go to step 1.
4 Perturbation of Reduced Costs
As a result of not involving reduced costs, the feasible solution, generated by Algorithm 13.3.2, could be far from dual feasibility, and hence the number of iterations subsequently required by Phase-II would be large. In this aspect, the Phase-I method, presented in this section, seems to be advantageous (Pan 2000b,c).
It solves an auxiliary perturbed program. Let (3.18) be an infeasible simplex tableau and let \(\delta _{j} \geq 0,\ j \in N\) be predetermined perturbation parameters. The auxiliary program results from perturbing reduced costs to nonnegative values. More precisely, introduce index set
Replacing reduced costs \(\bar{z}_{j}\) in the simplex tableau by
leads to a dual feasible tableau, associated with a perturbed program. As a result, the dual simplex algorithm can get itself started, until optimality of the perturbed program is achieved, or dual unboundedness of it is detected.
Theorem 13.4.1.
If the auxiliary perturbed program is dual unbounded, the original problem is infeasible.
Proof.
The perturbed program has the same constraints as the original problem. Unboundedness of the former implies that there is no feasible solution to former. and hence so to the latter. □
Assume that optimality of the perturbed program is achieved. It might be well to assume that (3.18) is again the final tableau, associated with basis B. The right-hand side of it is now nonnegative. Consequently, computing \(\hat{z}_{N} = c_{N} - {N}^{\mathrm{T}}{B}^{-1}c_{B}\) (restoring reduced costs of the original problem) and covering \(\bar{z}_{N}\) with \(\hat{z}_{N}\) leads to a feasible simplex tableau to the original problem.
The overall steps are put into the following algorithm.
Algorithm 13.4.1 (Tableau Phase-I: reduced cost perturbation).
Given perturbation parameters δ j ≥ 0, j ∈ N. Initial: simplex tableau of form (3.18). This algorithm finds a feasible simplex tableau to the standard LP problem.
-
1.
Perturb: \(\bar{z}_{j} =\delta _{j}\), \(\forall j \in \{ j \in N\ \vert \ \bar{z}_{j} <\delta _{j}\}\).
-
2.
Call the dual simplex Algorithm 4.4.1.
-
3.
Stop if it terminates at step 3 (infeasible problem).
-
4.
If it terminates at step 2, compute \(\bar{z}_{N} = c_{N} - {N}^{\mathrm{T}}{B}^{-1}c_{B}\).
-
5.
Stop (feasibility achieved).
Note There exists a perturbation-free variant of the preceding algorithm: in each iteration, one use \(\bar{z}{^\prime}_{N}\) determined by (13.23) in place of \(\bar{z}_{N}\) for minimum-ratio test, therefore saving computations for restoring reduced costs.
Algorithm 13.4.1 performed very well in preliminary computational experiments (Pan 2000b). A reason seems to be that not all perturbed reduced costs affect the final outcome. Sometimes, reduced costs restored in step 4 of Algorithm 13.4.1 are themselves nonnegative, and hence there is no need for carrying out Phase-II. In fact, it is verified that manipulations in step 2 amount to solving the following problem
where
That is to say, perturbing \(\bar{z}_{j} <\delta _{j}\) to \(\bar{z}_{j} =\delta _{j}\) amounts to slackening dual constraint \(a_{j}^{\mathrm{T}}y \leq c_{j}\) to
If slackened constraints are all inactive for dual optimal solutions, or in other words, nonnegative constraints of primal variables, associated with perturbed reduced costs, are active for primal optimal solutions, the primal and dual optimal solutions will not be affected by such perturbations at all (for detailed analysis, see Pan 2000b).
Example 13.4.1.
Solve the following problem using Algorithm 13.4.1 in Phase-I:
Answer
Phase-I.
Initial: There is an available simplex tableau (\(B =\{ 5,6,4\}\)) to this problem. Perturb negative reduced cost − 3 to 0, so that it becomes dual feasible. As an illustration for implementation, we add the perturbed reduced costs at the bottom (δ 3 = 0).
Taking the bottom row as objective row, execute the dual simplex Algorithm 4.4.1. Note that the second (objective) bottom row does not take a part in pivoting.
Iteration 1:
-
1.
\(\min \{-4,5,-3\} = -4 <0,\ p = 1\).
-
3.
\(J =\{ 1,2\}\neq \varnothing\).
-
4.
\(\min \{-1/(-2),-1/(-1)\} = 1/2,\ q = 1\).
-
5.
Multiply row 1 by \(-1/2\), then add \(-1,1,-1,-1\) times of row 1 to rows 2,3,4,5, respectively:
Iteration 2:
-
1.
\(\min \{2,3,-1\} = -1 <0,\ p = 3\).
-
3.
\(J =\{ 5\}\neq \varnothing\).
-
4.
\(\min \{-(1/2)/(-1/2)\} = 1,\ q = 5\).
-
5.
Multiply row 3 by − 2, then add \(1/2,-1/2,-1/2,-1/2\) times of row 3 to rows 1,2,4,5, respectively:
Optimality of the perturbed tableau is achieved. Deleting the perturbed (bottom) row gives a feasible simplex tableau of the original problem.
Phase-II.
Iteration 3: call the simplex Algorithm 3.2.1.
Select x 3 column as pivot column and the second row as pivot row. Multiply row 2 by 1∕3, then add 2, 1, −2 times of row 2 to rows 1,4,5, respectively:
The reduced costs are all nonnegative, optimality is hence achieved. Basic optimal solution and associated objective value:
Although the perturbation parameter was set to 0 in the preceding example, positive parameter value should be practically favorable for the sake of anti-degeneracy. It seems to be suitable to use
where ε is the feasibility tolerance (\(\epsilon = 1{0}^{-6}\), or so). Computational experiences indicate that the perturbation method is not sensitive to the magnitude of the parameters though exceeding 10−1 seems to be inadvisable.
References
Abadie J, Corpentier J (1969) Generalization of the Wolfe reduced gradient method to the case of mon-linear constrained optimization. In: Fletcher R (ed) Optimization. Academic, London, pp 37–48
Abel P (1987) On the choice of the pivot columns of the simplex method: gradient criteria. Computing 38:13–21
Adler I, Megiddo N (1985) A simplex algorithm whose average number of steps is bounded between two quadratic functions of the smaller dimension. J ACM 32:871–895
Adler I, Resende MGC, Veige G, Karmarkar N (1989) An implementation of Karmarkar’s algorithm for linear programming. Math Program 44:297–335
Andersen E, Andersen K (1995) Presolving in linear programming. Math Program 71:221–245
Andersen ED, Gondzio J, Mészáros C, Xu X (1996) Implementation of interior-point methods for large scale linear programming. In: Terlaky T (ed) Interior point methods of mathematical programming. Kluwer, Dordrecht
Andrel N, Barbulescu M (1993) Balance constraints reduction of large-scale linear programming problems. Ann Oper Res 43:149–170
Anstreicher KM, Watteyne P (1993) A family of search directions for Karmarkar’s algorithm. Oper Res 41:759–767
Arrow KJ, Hurwicz L (1956) Reduction of constrained maxima to saddle-point problems. In: Neyman J (ed) Proceedings of the third Berkeley symposium on mathematical statistics and probability, vol 5. University of California Press, Berkeley, pp 1–26
Avis D, Chvatal V (1978) Notes on Bland’s pivoting rule. Math Program 8:24–34
Balas E (1965) An additive algorithm for solving linear programs with zero-one variables. Oper Res 13:517–546
Balinski ML, Gomory RE (1963) A mutual primal-dual simplex method. In: Graves RL, Wolfe P (eds) Recent advances in mathematical programming. McGraw-Hill, New York
Balinski ML, Tucker AW (1969) Duality theory of linear problems: a constructive approach with applications. SIAM Rev 11:347–377
Barnes ER (1986) A variation on Karmarkars algorithm for solving linear programming problems. Math Program 36:174–182
Bartels RH (1971) A stabilization of the simplex method. Numer Math 16:414–434
Bartels RH, Golub GH (1969) The simplex method of linear programming using LU decomposition. Commun ACM 12:266–268
Bartels RH, Stoer J, Zenger Ch (1971) A realization of the simplex method based on triangular decompositions. In: Wilkinson JH, Reinsch C (eds) Contributions I/II in handbook for automatic computation, volume II: linear algebra. Springer, Berlin/London
Bazaraa MS, Jarvis JJ, Sherali HD (1977) Linear programming and network flows, 2nd edn. Wiley, New York
Beale EML (1954) An alternative method for linear programming. Proc Camb Philos Soc 50:513–523
Beale EML (1955) Cycling in the dual simplex algorithm. Nav Res Logist Q 2:269–275
Beale E (1968) Mathematical programming in practice. Topics in operations research. Pitman & Sons, London
Benders JF (1962) Partitioning procedures for solving mixed-variables programming problems. Numer Math 4:238–252
Benichou MJ, Cautier J, Hentges G, Ribiere G (1977) The efficient solution of large scale linear programming problems. Math Program 13:280–322
Bixby RE (1992) Implementing the simplex method: the initial basis. ORSA J Comput 4:287–294
Bixby RE (1994) Progress in linear programming. ORSA J Comput 6:15–22
Bixby RE (2002) Solving real-world linear problems: a decade and more of progress. Oper Res l50:3–15
Bixby RE, Saltzman MJ (1992) Recovering an optimal LP basis from the interior point solution. Technical report 607, Dapartment of Mathematical Sciences, Clemson University, Clemson
Bixby RE, Wagner DK (1987) A note on detecting simple redundancies in linear systems. Oper Res Lett 6:15–17
Bixby RE, Gregory JW, Lustig IJ, Marsten RE, Shanno DF (1992) Very large-scale linear programming: a case study in combining interior point and simplex methods. Oper Res 40:885–897
Björck A, Plemmons RJ, Schneider H (1981) Large-scale matrix problems. North-Holland, Amsterdanm
Bland RG (1977) New finite pivoting rules for the simplex method. Math Oper Res 2:103–107
Borgwardt K-H (1982a) Some distribution-dependent results about the asymptotic order of the arerage number of pivot steps of the simplex method. Math Oper Res 7:441–462.
Borgwardt K-H (1982b) The average number of pivot steps required by the simplex method is polynomial. Z Oper Res 26:157–177
Botsaris CA (1974) Differential gradient methods. J Math Anal Appl 63:177–198
Breadley A, Mitra G, Williams HB (1975) Analysis of mathematica problems prior to applying the simplex algorithm. Math Program 8:54–83
Brown AA, Bartholomew-Biggs MC (1987) ODE vs SQP methods for constrained optimization. Technical report, 179, The Numerical Center, Hatfield Polytechnic
Carolan MJ, Hill JE, Kennington JL, Niemi S, Wichmann SJ (1990) An empirical evaluation of the KORBX algorithms for military airlift applications. Oper Res 38:240–248
Cavalier TM, Soyster AL (1985) Some computational experience and a modification of the Karmarkar algorithm. ISME working paper, The Pennsylvania State University, pp 85–105
Chan TF (1985) On the existence and computation of LU factorizations with small pivots. Math Comput 42:535–548
Chang YY (1979) Least index resolution of degeneracy in linear complementarity problems. Technical Report 79–14, Department of OR, Stanford University
Charnes A (1952) Optimality and degeneracy in linear programming. Econometrica 20:160–170
Cheng MC (1987) General criteria for redundant and nonredundant linear inequalities. J Optim Theory Appl 53:37–42
Chvatal V (1983) Linear programming. W.H. Freeman, New York
Cipra BA (2000) The best of the 20th century: editors name top 10 algorithms. SIAM News 33: 1–2
Coleman TF, Pothen A (1987) The null space problem II. Algorithms. SIAM J Algebra Discret Methods 8:544–562
Cook AS (1971) The complexity of theorem-proving procedure. In: Proceedings of third annual ACM symposium on theory of computing, Shaker Heights, 1971. ACM, New York, pp 151–158
Cottle RW, Johnson E, Wets R (2007) George B. Dantzig (1914–2005). Not AMS 54:344–362
CPLEX ILOG (2007) 11.0 User’s manual. ILOG SA, Gentilly, France
Curtis A, Reid J (1972) On the automatic scaling of mtrices for Gaussian elimination. J Inst Math Appl 10:118–124
Dantzig GB (1948) Programming in a linear structure, Comptroller, USAF, Washington, DC
Dantzig GB (1951a) Programming of interdependent activities, mathematical model. In: Koopmas TC (ed) Activity analysis of production and allocation. Wiley, New York, pp 19–32; Econometrica 17(3/4):200–211 (1949)
Dantzig GB (1951b) Maximization of a linear function of variables subject to linear inequalities. In: Koopmans TC (ed) Activity analysis of production and allocation. Wiley, New York, pp 339–347
Dantzig GB (1951c) A proof of the equivalence of the programming problem and the game problem. In: Koopmans T (ed) Activity analysis of production and allocation. Wiley, New York, pp 330–335
Dantzig GB (1963) Linear programming and extensions. Princeton University Press, Princeton
Dantzig GB (1991) linear programming. In: Lenstra JK, Rinnooy Kan AHG, Schrijver A (eds) History of mathematical programming. CWI, Amsterdam, pp 19–31
Dantzig GB, Ford LR, Fulkerson DR (1956) A primal-dual algorithm for linear programs. In: Kuhn HK, Tucker AW (eds) Linear inequalities and related systems. Princeton University Press, Princeton, pp 171–181
Dantzig GB, Orchard-Hays W (1953) Alternate algorithm for the revised simplex method using product form for the inverse. Notes on linear programming: part V, RM-1268. The RAND Corporation, Santa Monica
Dantzig GB, Orchard-Hayes W (1954) The product form for the inverse in the simplex method. Math Tables Other Aids Comput 8:64–67
Dantzig GB, Thapa MN (1997) Linear programming 1: introduction. Springer, New York
Dantzig GB, Thapa MN (2003) Linear programming 2: theory and extensions. Springer, New York
Dantzig GB, Wolfe P (1960) Decomposition principle for linear programs. Oper Res 8:101–111
Dantzig GB, Orden A, Wolfe P (1955) The generalized simplex method for minimizing a linear form under linear inequality constraints. Pac J Math 5:183–195
de Ghellinck G, Vial J-Ph, Polynomial (1986) Newton method for linear programming. Algorithnica 1:425–453. (Special issue)
Dikin I (1967) Iterative solution of problems of linear and quadratic programming. Sov Math Dokl 8:674–675
Dikin I (1974) On the speed of an iterative process. Upravlyaemye Sistemi 12:54–60
Dorfman R, Samuelson PA, Solow RM (1958) Linear programming and economic analysis. McGraw-Hill, New York
Duff IS, Erisman AM, Reid JK (1986) Direct methods for sparse matrices. Oxford University Press, Oxford
Evtushenko YG (1974) Two numerical methods for solving non-linear programming problems. Sov Math Dokl 15:420–423
Fang S-C (1993) Linear optimization and extensions: theory and algorithms. AT & T, Prentice-Hall, Englewood Cliffs
Farkas J (1902) Uber die Theorie der Einfachen Ungleichungen. Journal für die Reine und Angewandte Mathematik 124:1–27
Fiacco AV, Mccormick GP (1968) Nonlinear programming: sequential unconstrained minimization techniques. Wiley, New York
Fletcher R (1981) Practical methods of optimization. Volume 2: constrained optimization. Wiley, Chichester
Ford Jr LR, Fulkerson DR (1956) Maximal flow through a network. Can J Math 8:399–407
Forrest JJH, Goldfarb D (1992) Steepest edge simplex algorithm for linear programming. Math Program 57:341–374
Forrest J, Tomlin J (1972) Updating triangular factors of the basis to maintain sparsity in the product form simplex method. Math Program 2:263–278
Forsythe GE, Malcom MA, Moler CB (1977) Computer methods for mathematical computations. Pretice-Hall, EnglewoodCliffs
Fourer R (1979) Sparse Gaussian elimination of staircase linear systems. Tehchnical report ADA081856, Calif Systems Optimization LAB, Stanford University
Fourier JBJ (1823) Analyse des travaux de l’Academie Royale des Science, pendant l’Iannee, Partie mathematique, Histoire de l’Acanemie Royale des Sciences de l’nstitut de France 6 [1823] (1826), xxix–xli (partially reprinted as: Premier extrait, in Oeuvres de Fourier, Tome 11 (Darboux G, ed.), Gauthier-Villars, Paris, 1890, (reprinted: G. Olms, Hildesheim, 1970), pp 321–324
Frisch KR (1955) The logarithmic potentical method of convex programming. Memorandum, University Institute of Economics, Oslo
Fulkerson D, Wolfe P (1962) An algorithm for scaling matrices. SIAM Rev 4:142–146
Gale D, Kuhn HW, Tucker AW (1951) Linear programming and the theory of games. In: Koopmans T (ed) Activity analysis of production and allocation. Wiley, New York, pp 317–329
Gass SI (1985) Linear programming: methods and applications. McGraw-Hill, New York
Gass SI, Saaty T (1955) The computational algorithm for the parametric objective function. Nav Res Logist Q 2:39–45
Gay DM (1978) On combining the schemes of Reid and Saunders for sparse LP bases. In: Duff IS, Stewart GW (eds) Sparse matrix proceedings. SIAM, Philadelphia, pp 313–334
Gay D (1985) Electronic mail distribution of linear programming test problems. COAL Newsl 13:10–12
Gay DM (1987) A variant of Karmarkar’s linear programming algorithm for problems in standard form. Math Program 37:81–90
Geoffrion AM (1972) Generalized Benders decomposition. JOTA 10:137–154
George A, Liu W-H (1981) Computing solution of large sparse positive definite systems. Prentice-Hall, Engleewood Cliffs
Gill PE, Murray W (1973) A numerically stable form of the simplex algorithm. Linear Algebra Appl 7:99–138
Gill PE, Murray W, Saunders MA, Tomlin JA, Wright MH (1985) On projected Newton barrier methods for linear programming and an equivalence to Karmarkar’s projected method. Technical report SOL 85–11, Department of Operations Research, Stanford University
Gill PE, Murray W, Saunders MA, Tomlin JA, Wright MH (1986) On projected Newton methods for linear programming and an equivalence to Karmarkar’s projected method. Math Program 36:183–209
Gill PE, Murray W, Saunders MA, Wright MH (1987) Maintaining LU factors of a general sparse matrix. Linear Algebra Appl 88/89:239–270
Gill PE, Murray W, Saunders MA, Wright MH (1989) A practical anti-cycling procedure for linearly constrainted optimization. Math Program 45:437–474
Goldfarb D (1977) On the Bartels-Golub decomposition for linear programming bases. Math Program 13:272–279
Goldfarb D, Reid JK (1977) A practicable steepest-edge simplex algorithm. Math Program 12:361–371
Goldman AJ, Tucker AW (1956a) Polyhedral convex cones. In: Kuhn HW, Tucker AW (eds) Linear inequalities and related systems. Annals of mathmatical studies, vol 38. Princeton University Press, Princeton, pp 19–39
Goldman AJ, Tucker AW (1956b) Theory of linear programming. In: Kuhn HW, Tucker AW (eds) Linear inequalities and related systems. Annals of mathmatical studies, vol 38. Princeton University Press, Princeton, pp 53–97
Golub GH (1965) Numerical methods for solving linear least squares problems. Numer Math 7:206–216
Golub GH, Van Loan CF (1989) Matrix computations, 2edn. The Johns Hopkins University Press, Baltimore
Gomory AW (1958) Outline of an algorithm for integer solutions to linear programs. Bull Am Math Soc 64:275–278
Gonzaga CC (1987) An algorithm for solving linear programming problems in O(n 3 L) operations. Technical Report UCB/ERL M87/10, Electronics Research Laboratory, University of California, Berkeley
Gonzaga CC (1990) Convergence of the large step primal affine-scaling algorithm for primal non-degenerate linear problems. Technical report, Department of Systems Engineering and Computer Sciences, COPPE-Federal University of Riode Janeiro
Gould N, Reid J (1989) New crash procedure for large systems of linear constraints. Math Program 45:475–501
Greenberg HJ (1978) Pivot selection tactics. In: Greenberg HJ (ed) Design and implementation of optimization software. Sijthoff and Noordhoff, Alphen aan den Rijn, pp 109–143
Greenberg HJ, Kalan J (1975) An exact update for Harris’ tread. Math Program Study 4:26–29
Guerrero-Garcia P, Santos-Palomo A (2005) Phase I cycling under the most-obtuse-angle pivot rule. Eur J Oper Res 167:20–27
Guerrero-Garcia P, Santos-Palomo A (2009) A deficient-basis dual counterpart of Pararrizos, Samaras ans Stephanides’ primal-dual simplex-type algorithm. Optim Methods Softw 24:187–204
Güler O, Ye Y (1993) Convergence behavior of interior-point algorithms. Math Program 60:215–228
Hadley G (1972) Linear programming. Addison-Wesley, Reading
Hager WW (2002) The dual active set algorithm and its application to linear programming. Comput Optim Appl 21:263–275
Hall LA, Vanderbei RJ (1993) Two-third is sharp for affine scaling. Oper Res Lett 13:197–201
Hamming RW (1971) Introduction to applied numerical analysis. McGraw-Hill, New York
Harris PMJ (1973) Pivot selection methods of the Devex LP code. Math Program 5:1–28
Hattersley B, Wilson J (1988) A dual approach to primal degeneracy. Math Program 42:135–145
He X-C, Sun W-Y (1991) An intorduction to generalized inverse matrix (in Chinese). Jiangsu Science and Technology Press, Nanjing
Hellerman E, Rarick DC (1971) Reinversion with the preassigned pivot procedure. Math Program 1:195–216
Hellerman E, Rarick DC (1972) The partitioned preassigned pivot procedure. In: Rose DJ, Willouhby RA (eds) Sparse matrices and their applications. Plenum, New York, pp 68–76
Hertog DD, Roos C (1991) A survey of search directions in interior point methods for linear programming. Math Program 52:481–509
Hoffman AJ (1953) Cycling in the simplex algorithm. Technical report 2974, National Bureau of Standards
Hu J-F (2007) A note on “an improved initial basis for the simplex algorithm”. Comput Oper Res 34:3397–3401
Hu J-F, Pan P-Q (2006) A second note on ‘A method to solve the feasible basis of LP’ (in Chinese). Oper Res Manag Sci 15:13–15
Hu J-F, Pan P-Q (2008a) Fresh views on some recent developments in the simplex algorithm. J Southeast Univ 24:124–126
Hu J-F, Pan P-Q (2008b) An efficient approach to updating simplex multipliers in the simplex algorithm. Math Program Ser A 114:235–248
Jansen B, Terlakey T, Roos C (1994) The theory of linear programming: skew symmetric self-dual problems and the central path. Optimization 29:225–233
Jansen B, Roos C, Terlaky T (1996) Target-following methods for linear programming. In: Terlaky T (ed) Interior point methods of mathematical programming. Kluwer, Dordrecht
Jeroslow R (1973) The simplex algorithm with the pivot rule of maximizing criterion improvement. Discret Appl Math 4:367–377
Kalantari B (1990) Karmarkar’s algorithm with improved steps. Math Program 46:73–78
Kallio M, Porteus EL (1978) A class of methods for linear programming. Math Program 14:161–169
Kantorovich LV (1960) Mathematical methods in the organization and planning of production. Manag Sci 6:550–559. Original Russian version appeared in 1939
Karmarkar N (1984) A new polynomial time algorithm for linear programming. Combinatorica 4:373–395
Karmarkar N, Ramakrishnan K (1985) Further developments in the new polynomial-time algorithm for linear progrmming. In: Talk given at ORSA/TIMES national meeting, Boston, Apr 1985
Khachiyan L (1979) A polynomial algorithm in linear programming. Doklady Academiia Nauk SSSR 244:1093–1096
Kirillova FM, Gabasov R, Kostyukova OI (1979) A method of solving general linear programming problems. Doklady AN BSSR (in Russian) 23:197–200
Klee V (1965) A class of linear problemminng problems requiring a larger number of iterations. Numer Math 7:313–321
Klee V, Minty GJ (1972) How good is the simplex algorithm? In: Shisha O (ed) Inequalities-III. Academic, New York, pp 159–175
Koberstein A (2008) Progress in the dual simplex algorithm for solving large scale LP problems: techniques for a fast and stable implementation. Comput Optim Appl 41:185–204
Koberstein A, Suhl UH (2007) Progress in the dual simplex method for large scale LP problems: practical dual phase 1 algorithms. Comput Optim Appl 37:49–65
Kojima M, Mizuno S, Yoshise A (1989) A primal-dual interior point algorithm for linear programming. In: Megiddo N (ed) Progress in mathematical programming. Springer, New York, pp 29–47
Kojima M, Megiddo N, Mizuno S (1993) A primal-dual infeasible-interior-point algorithm for linear programming. Math Program 61:263–280
Kortanek KO, Shi M (1987) Convergence results and numerical experiments on a linear programming hybrid algorithm. Eur J Oper Res 32:47–61
Kostina E (2002) The long step rule in the bounded-variable dual simplex method: numerical experiments. Math Methods Oper Res 55:413–429
Kotiah TCT, Steinberg DI (1978) On the possibility of cycling with the simplex method. Oper Res 26:374–376
Kuhn HW, Quandt RE (1953) An experimental study of the simplex method. In: Metropolis NC et al (eds) Eperimental arithmetic, high-speed computing and mathematics. Proceedings of symposia in applied mathematics XV. American Mathematical Society, Providence, pp 107–124
Land AH, Doig AG (1960) An automatic method of solving discrete programming problems. Econometrica 28:497–520
Leichner SA, Dantzig GB, Davis JW (1993) A strictly, improving linear programming phase I algorithm. Ann Oper Res 47:409–430
Lemke CE (1954) The dual method of solving the linear programming problem. Nav Res Logist Q 1:36–47
Li W (2004) A note on two direct methods in linear programming. Eur J Oper Res 158:262–265
Li C, Pan P-Q, Li W (2002) A revised simplex algorithm based on partial pricing pivotrule (in Chinese). J Wenzhou Univ 15:53–55
Li W, Guerrero-Garcia P, Santos-Palomo A (2006a) A basis-deficiency-allowing primal phase-1 algorithm using the most-obtuse-angle column rule. Comput Math Appl 51:903–914
Li W, Pan P-Q, Chen G (2006b) A combined projected gradient algorithm for linear programming. Optim Methods Softw 21:541–550
Llewellyn RW (1964) Linear programming. Holt, Rinehart and Winston, New York
Luo Z-Q, Wu S (1994) A modified predictor-corrector method for linear programming. Comput Optim Appl 3:83–91
Lustig IJ (1990) Feasibility issues in a prinal-dual interior-point method for linear programming. Math Program 49:145–162
Lustig IJ, Marsten RE, Shanno DF (1991) Computational exeperience with a primal-dual interior point method for linear programming. Linear Algebra Appl 152:191–222
Lustig IJ, Marsten RE, Shanno DF (1992) On implementing Mehrotras’s predictor-corrector interior-point for linear programming. SIAM J Optim 2:435–449
Lustig IJ, Marsten R, Shanno D (1994) Interior point methods for linear programming: computational state of the art. ORSA J Comput 6:1–14
Markowitz HM (1957) The elimination form of the inverse and its application to linear programming. Manag Sci 3:255–269
Maros I (1986) A general phase-1 method in linear programming. Eur J Oper Res 23:64–77
Maros I (2003a) A generalied dual phase-2 simplex algorithm. Eur J Oper Res 149:1–16
Maros I (2003b) Computational techniques of the simplex method. International series in operations research and management, vol 61. Kluwer, Boston
Maros I, Khaliq M (2002) Advances in design and implementation of optimization software. Eur J Oper Res 140:322–337
Marshall KT, Suurballe JW (1969) A note on cycling in the simplex method. Nav Res Logist Q 16:121–137
Martin RK (1999) Large scale linear and integer optimization: a unified approach. Kluwer, Boston
Mascarenhas WF (1997) The affine scaling algorithm fails for λ = 0. 999. SIAM J Optim 7:34–46
McShane KA, Monma CL, Shanno DF (1989) An implementation of a primal-dual method for linear programming. ORSA J Comput 1:70–83
Megiddo N (1986a) Introduction: new approaches to linear programming. Algorithmca 1:387–394. (Special issue)
Megiddo N (1986b) A note on degenerach in linear programming. Math Program 35:365–367
Megiddo N (1989) Pathways to the optimal set in linear programming. In: Megiddo N (ed) Progress in mathematical programming. Springer, New York, pp 131–158
Megiddo N, Shub M (1989) Boundary behavior of interior point algorithm in linear programming. Math Oper Res 14:97–146
Mehrotra S (1991) On finding a vertex solution using interior point methods. Linear Algebra Appl 152:233–253
Mehrotra S (1992) On the implementation of a primal-dual interior point method. SIAM J Optim 2:575–601
Mizuno S, Todd MJ, Ye Y (1993) On adaptive-step primal-dual interior-point algorithms for linear programming. Math Oper Res 18:964–981
Monteiro RDC, Adler I (1989) Interior path following primal-dual algorithms: Part I: linear programming. Math Program 44:27–41
Murtagh BA (1981) Advances in linear programming: computation and practice. McGraw-Hill, New York/London
Murtagh BA, Saunders MA (1978) Large-scale linearly constrained optimization. Math Program 14:41–72
Murtagh BA, Saunders MA (1998) MINOS 5.5 user’s guid. Technical report SOL 83-20R, Department of Engineering Economics Systems & Operations Research, Stanford University, Stanford
Murty KG (1983) Linear programming. Wiley, New York
Nazareth JL (1987) Computer solutions of linear programs. Oxford University Press, Oxford
Nazareth JL (1996) The implementation of linear programming algorithms based on homotopies. Algorithmica 15:332–350
Nemhauser GL (1994) The age of optimizaiton: solving large-scale real-world problems. Oper Res 42:5–13
Nemhauser GL, Wolsey LA (1999) Integer and combinatorial optimization. Wiley, New York
Nocedal J, Wright SJ (1999) Numerical optimization. Springer, Berlin
Ogryczak W (1988) The simplex method is not always well behaved. Linear Algebra Appl 109:41–57
Orchard-Hays W (1954) Background development and extensions of the revised simplex method. Report RM 1433, The Rand Corporation, Santa Monica
Orchard-Hays W (1956) Evolution of computer codes for linear programming. Paper P-810, The RAND Corporation, p 2224
Orchard-Hays W (1971) Advanced linear programming computing techniques. McGraw-Hill, New York
Padberg MW (1995) Linear optimization and extensions. Springer, Berlin
Pan P-Q (1982) Differential equaltion methods for unconstarained optimization (in Chinese). Numer Math J Chin Univ 4:338–349
Pan P-Q (1990) Practical finite pivoting rules for the simplex method. OR Spektrum 12:219–225
Pan P-Q (1991) Simplex-like method with bisection for linear programming. Optimization 22:717–743
Pan P-Q (1992a) New ODE methods for equality constrained optimization (I) – equations. J Comput Math 10:77–92
Pan P-Q (1992b) New ODE methods for equality constrained optimization (II) – algorithms. J Comput Math 10:129–146
Pan P-Q (1992c) Modification of Bland’s pivoting rule (in Chinese). Numer Math 14:379–381
Pan P-Q (1994a) A variant of the dual pivot rule in linear programming. J Inf Optim Sci 15:405–413
Pan P-Q (1994b) Composite phase-1 methods without measuring infeasibility. In: Yue M-Y (ed) Theory of optimization and its applications. Xidian University Press, Xian, pp 359–364.
Pan P-Q (1994c) Ratio-test-free pivoting rules for the bisection simplex method. In: Proceedings of national conference on decision making science, Shangrao, pp 24–29
Pan P-Q (1994d) Ratio-test-free pivoting rules for a dual phase-1 method. In: Xiao S-T, Wu F (eds) Proceeding of the third conference of Chinese SIAM. Tsinghua University press, Beijing, pp 245–249
Pan P-Q (1995) New non-monotone procedures for achieving dual feasibility. J Nanjing Univ Math Biquarterly 12:155–162
Pan P-Q (1996a) A modified bisection simplex method for linear programming. J Comput Math 14:249–255
Pan P-Q (1996b) New pivot rules for achieving dual feasibility. In: Wei Z (ed) Theory and applications of OR. Proceedings of the fifth conference of Chinese OR society, Xian, 10–14 Oct 1996. Xidian University Press, Xian, pp 109–113
Pan P-Q (1996c) Solving linear programming problems via appending an elastic constraint. J Southeast Univ (English edn) 12:253–265
Pan P-Q (1997) The most-obtuse-angle row pivot rule for achieving dual feasibility in linear programming: a computational study. Eur J Oper Res 101:164–176
Pan P-Q (1998a) A dual projective simplex method for linear programming. Comput Math Appl 35:119–135
Pan P-Q (1998b) A basis-deficiency-allowing variation of the simplex method. Comput Math Appl 36:33–53
Pan P-Q (1999a) A new perturbation simplex algorithm for linear programming. J Comput Math 17:233–242
Pan P-Q (1999b) A projective simplex method for linear programming. Linear Algebra Appl 292:99–125
Pan P-Q (2000a) A projective simplex algorithm using LU decomposition. Comput Math Appl 39:187–208
Pan P-Q (2000b) Primal perturbation simplex algorithms for linear programming. J Comput Math 18:587–596
Pan P-Q (2000c) On developments of pivot algorithms for linear programming. In: Proceedings of the sixth national conference of operations research society of China, Changsha, 10–15 Oct 2000. Global-Link Publishing, Hong Kong, pp 120–129
Pan P-Q (2004) A dual projective pivot algorithm for linear programming. Comput Optim Appl 29:333–344
Pan P-Q (2005) A revised dual projective pivot algorithm for linear programming. SIAM J Optim 16:49–68
Pan P-Q (2008a) A largest-distance pivot rule for the simplex algorithm. Eur J Oper Res 187:393–402
Pan P-Q (2008b) A primal deficient-basis algorithm for linear programming. Appl Math Comput 198:898–912
Pan P-Q (2008c) Efficient nested pricing in the simplex algorithm. Oper Res Lett 38:309–313
Pan P-Q (2010) A fast simplex algorithm for linear programming. J Comput Math 28(6):837–847
Pan P-Q (2013) An affine-scaling pivot algorithm for linear programming. Optimization 62: 431–445
Pan P-Q, Li W (2003) A non-monotone phase-1 method in linear programming. J Southeast Univ (English edn) 19:293–296
Pan P-Q, Ouiang Z-X (1993) Two variants of the simplex algorithm (in Chinese). J Math Res Expo 13(2):274–275
Pan P-Q, Ouyang Z-X (1994) Moore-Penrose inverse simplex algorithms based on successive linear subprogramming approach. Numer Math 3:180–190
Pan P-Q, Pan Y-P (2001) A phase-1 approach to the generalized simplex algorithm. Comput Math Appl 42:1455–1464
Pan P-Q, Li W, Wang Y (2004) A phase-1 algorithm using the most-obtuse-angle rule for the basis-deficiency-allowing dual simplex method. OR Trans 8:88–96
Pan P-Q, Li W, Cao J (2006a) Partial pricing rule simplex method with deficient basis. Numer Math J Chin Univ (English series) 15:23–30
Pan P-Q, Hu J-F, Li C (2006b) Feasible region contraction interior point algorithm. Appl Math Comput 182:1361–1368
Papadimitriou CH, Steiglitz K (1982) Combinatorial optimization: algorithms and complexity. Prentice-Hall, New Jersey
Perold AF (1980) A degeneracy exploiting LU factorization for the simplex method. Math Program 19:239–254
Powell MJD (1989) A tolerant algorithm for linearly constrained optimization calculations. Math Program 45:547–566
Reid JK (1982) A sparsity-exploiting variant of the Bartels-Golub decomposition for linear programming bases. Math Program 24:55–69
Rockafellar RT (1997) Convext analysis. Princeton University Press, Princeton
Roos C (1990) An exponential example for Terlaky’s pivoting rule for the criss-cross simplex method. Math Program 46:79–84
Roos C, Vial J-Ph (1992) A polynomial method of approximate centers for linear programming. Math Program 54:295–305
Roos C, Terlaky T, Vial J-P (1997) Theory and algorithms for linear programming. Wiley, Chichester
Rothenberg RI (1979) Linear programming. North-Holland, New York
Ryan D, Osborne M (1988) On the solution of highly degenerate linear problems. Math Program 41:385–392
Saigal R (1995) Linear programming. Kluwer, Boston
Santos-Palomo A (2004) The sagitta method for solving linear problems. Eur J Oper Res 157:527–539
Saunders MA (1972) Large scale linear programming using the Cholesky factorization. Technical report STAN-CS-72-152, Stanford University
Saunders MA (1973) The complexity of LU updating in the simplex method. In: Andersen R, Brent R (eds) The complexity of computational problem solving. University Press, St. Lucia, pp 214–230
Saunders MA (1976) A fast and stable implementation of the simplex method using Bartels-Golub updating. In: Bunch J, Rose D (eds) Sparse matrix computation. Academic, New York, pp 213–226
Schrijver A (1986) Theory of linear and integer programming. Wiley, Chichester
Shanno DF, Bagchi A (1990) A unified view of interior point methods for linear programming. Ann Oper Res 22:55–70
Shen Y, Pan P-Q (2006) Dual besiction simplex algorithm (in Chinese). In: Preceedings of the national conference of operatios research society of China, Shenzhen. Globa-Link Informatics, Hong Kong, pp 168–174
Shi Y, Pan P-Q (2011) Higher order iteration schemes for unconstrained optimization. Am J Oper Res 1:73–83
Smale S (1983a) On the average number of steps of the simplex method of linear programming. Math Program 27:241–262
Smale S (1983b) The problem of the average speed of the simplex method. In: Bachem A, Grotschel M, Korte B, (eds) Mathematical programming, the state of the art. Springer, Berlin, pp 530–539
Srinath LS (1982) Linear programming: principles and applications. Affiliated East-West Press, New Delhi
Suhl UH (1994) Mathematical optimization system. Eur J Oper Res 72:312–322
Suhl LM, Suhl UH (1990) Computing sparse LU factorization for large-scale linear programming bases. ORSA J Comput 2:325–335
Suhl LM, Suhl UH (1993) A fast LU-update for linear programming. Ann Oper Res 43:33–47
Sun W, Yuan Y-X (2006) Optimization theory an methods: nonlinear programming. Springer, New York
Swietanowaki A (1998) A new steepest edge approximation for the simplex method for linear programming. Comput Optim Appl 10:271–281
Taha H (1975) Integer programming: theory, applications, and computations. Academic, Orlando
Talacko J V, Rockefeller RT (1960) A Compact Simplex Algorithm and a Symmetric Algorithm for General Linear Programs. Unpublished paper, Marquette University
Tanabe K (1977) A geometric method in non-linear programming. Technical Report 23343-AMD780, Brookhaven National Laboratory, New York
Tanabe K (1990) Centered Newton method for linear prgramming: interior and ‘exterior’ point method, (in Japannese). In: Tone K (ed) New methods for linear programming 3. The Institute of Statistical Mathematics, Tokeo, pp 98–100
Tapia RA, Zhang Y (1991) An optimal-basis identification technique for interior-point linear programming algorithms. Linear Algebra Appl 152:343–363
Terlaky T (1985) A covergent criss-cross method. Math Oper Stat Ser Optim 16:683–690
Terlaky T (1993) Pivot rules for linear programming: a survey on recent theoretical developments. Ann Oper Res 46:203–233
Terlaky T (ed) (1996) Interior point methods of mathematical programming. Kluwer, Dordrecht
Todd MJ (1982) An implementation of the simplex method for linear programming problems with variable upper bounds. Math Program 23:23–49
Todd MJ (1983) Large scale linear programming: geometry, working bases and factorizations. Math Program 26:1–23
Tomlin JA (1972) Modifying trangular factors of the basis in the simplex method. In: Rose DJ, Willoughby RA (eds) Sparse matrices and applications. Plenum, New York
Tomlin JA (1974) On pricing and backward transformation in linear programming. Math Program 6:42–47
Tomlin JA (1975) On scaling linear programming problems. Math Program Study 4:146–166
Tomlin JA (1987) An experimental approach to Karmarkar’s projective method, for linear programming. Math Program 31:175–191
Tsuchiya T (1992) Global convergence property of the affine scaling method for primal degenerate linear programming problems. Math Oper Res 17:527–557
Tsuchiya T, Muramatsu M (1995) Global convergence of a long-step affine-scaling algorithm for degenrate linear programming problems. SIAM J Optim 5:525–551
Tucker AW (1956) Dual systems of homegeneous linear relations. In: Kuhn HW, Tucker AW, Dantzig GB (eds) Linear inequalities and related systems. Princeton University Press, Princeton, pp 3–18
Turner K (1991) Computing projections for the Karmarkar algorithtm. Linear Algebra Appl 152:141–154
Vanderbei RJ, Lagarias JC (1990) I.I. Dikin’s convergence result for the affine-scaling algorithm. Contemp Math 114:109–119
Vanderbei RJ, Meketon M, Freedman B (1986) A modification of Karmarkars linear programming algorithm. Algorithmica 1:395–407
Vemuganti RR (2004) On gradient simplex methods for linear programs. J Appl Math Decis Sci 8:107–129
Wang Z (1987) A conformal elimination-free algorithm for oriented matroid programming. Chin Ann Math 8(B1):16–25
Wilkinson JH (1971) Moden error analysis. SIAM Rev 13:548–568
Wolfe P (1963) A technique for resolving degeneracy in linear programming. J Oper Res Soc 11:205–211
Wolfe P (1965) The composite simplex algorithm. SIAM Rev 7:42–54
Wolsey L (1998) Integer programming. Wiley, New York
Wright SJ (1997) Primal-dual interior-point methods. SIAM, Philadelphia
Xu X, Ye Y (1995) A generalized homogeneous and self-dual algorithm for linear prgramming. Oper Res Lett 17:181–190
Xu X, Hung P-F, Ye Y (1996) A simplified homogeneous and self-dual linear programming algorithm and its implementation. Ann Oper Res 62:151–171
Yan W-L, Pan P-Q (2001) Improvement of the subproblem in the bisection simplex algorithm (in Chinese). J Southeast Univ 31:324–241
Yan A, Pan P-Q (2005) Variation of the conventional pivot rule and the application in the deficient basis algorithm (in Chinese). Oper Res Manag Sci 14:28–33
Yan H-Y, Pan P-Q (2009) Most-obtuse-angle criss-cross algorithm for linear programming (in Chinese). Numer Math J Chin Univ 31:209–215
Yang X-Y, Pan P-Q (2006) Most-obtuse-angle dual relaxation algorithm (in Chinese). In: Preceedings of the national conference of operatios research society of China, Shenzhen. Globa-Link Informatics, Hong Kong, pp 150–155
Ye Y (1987) Eliminating columns in the simplex method for linear programming. Technical report SOL 87–14, Department of Operations Research, Stanford Univesity, Stanford
Ye Y (1990) A ‘build-down’ scheme for linear probleming. Math Program 46:61–72
Ye Y (1997) Interior point algorithms: theory and analysis. Wiley, New York
Ye Y, Todd MJ, Mizuno S (1994) An \(o(\sqrt{n}l)\)-iteration homogeneous and selfdual linear programming algorithm. Math Oper Res 19:53–67
Zhang H, Pan P-Q (2008) An interior-point algorithm for linear programming (in Chinese). In: Preceedings of the national conference of operatios research society of China, Nanjing. Globa-Link Informatics, Hong Kong, pp 183–187
Zhang J-Z, Xu S-J (1997) Linear programming (in Chinese). Science Press, Bejing
Zhang L-H, Yang W-H, Liao L-Z (2013) On an efficient implementation of the face algorithm for linear programming. J Comp Math 31:335–354
Zhou Z-J, Pan P-Q, Chen S-F (2009) Most-obtuse-angle relaxation algorithm (in Chinese). Oper Res Manag Sci 18:7–10
Zionts S (1969) The criss-cross method for solving linear programming problems. Manag Sci 15:420–445
Zlatev Z (1980) On some pivotal strategies in Gaussian elimination by sparse technique. SIAM J Numer Anal 17:12–30
Zörnig P (2006) Systematic construction of examples for cycling in the simplex method. Comput Oper Res 33:2247–2262
Zoutendijk G (1960) Methods of feasible directions. Elsevier, Amsterdam
Author information
Authors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer-Verlag Berlin Heidelberg
About this chapter
Cite this chapter
PAN, PQ. (2014). Simplex Phase-I Method. In: Linear Programming Computation. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-40754-3_13
Download citation
DOI: https://doi.org/10.1007/978-3-642-40754-3_13
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-40753-6
Online ISBN: 978-3-642-40754-3
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)