A standard linear programing refers to an LP in the form:
A nonstandard LP can always be transformed into an equivalent standard LP, by adding new variables. For example, is equivalent to with .
Polyhedral set
A feasible region of an LP problem is a polyhedral set (polyhedron), that is, a set of the form .
We can find the corner points of a polyhedral set in the following way. First impose the equality constraints to ensure feasibility, and then require that enough additional constraints be active, so that we get a total of linearly independent active constraints and a unique is therefore determined. We can call such an be a basic solution. If no inactive constraint is violated, it is called a basic feasible solution.
Definition: Consider a polyhedral set and a point . Then is a basic solution if: all equality constraints are active and; out of the constraints that are active at , there are of them that are linearly independent.
For a polyhedral set, the concept of vertex (the usual geometry definition, i.e., the point on one side of a hyperplane which meets the polyhedron only at that point), extreme point (the point that cannot be represented as a convex combination of any other two points in the polyhedron), basic feasible solution (the point in the set where the equality constraints are active and there are of the active constraints that are linearly independent) are all the same.
Consider the polyhedral set : (we should assume that has full row rank ). Denote , where is the column vector of .
Theorem 1: Characterization of basic solution. is a basic solution iff , and there exists indices such that the columns are linear independent and .
According to this theorem, we have:
Procedure for constructing basic solutions: Choose independent columns of , set for the column indexes not selected, and then solve .
Theorem 1': is a basic feasible solution iff are linear independent, where are some indexes for which is positive.
Optimal condition
Theorem 2: For a standard LP, if is bounded, then at least one of its vertices is an optimal point of the LP problem.
Theorem 2': Suppose has at least one vertex and there exists an optimal solution. Then there exists an optimal solution which is a vertex of .
Theorem 2'': Suppose has at least one extreme point. Then, either the optimal cost is , or there exists a vertex which is optimal.
Corollary: Consider LP of minimizing over a nonempty polyhedral set. Then either the optimal cost is or there exists an optimal solution.
We call a nonzero vector a direction of a convex set if . If , then it is clear is a direction of iff
The extreme points of are called the extreme direction of . The polyhedral set can only have finitely many extreme directions.
Theorem 3: Consider standard LP. Denote the extreme directions of as . If , then at least one extreme point of is an optimal solution. Otherwise, the optimal objective value is .
Simplex method
Introduction of simplex method
Based on Theorem 2'', the simplex method searches for an optimal solution by moving from one basic feasible solution to another, along the edges of the feasible set, always in a cost reducing direction. Eventually, a basic feasible solution is reached at which none of the available edges leads to a cost reduction; such a basic feasible solution is optimal and the algorithm terminates.
Why does this method work? Many optimization algorithms are structured as follows: given a feasible solution, we search its neighborhood to find a nearby feasible solution with lower cost. If no nearby feasible solution leads to a cost improvement, the algorithm terminates and we have a locally optimal solution. For general optimization problems, a locally optimal solution need not be (globally) optimal. Fortunately, in linear programming, local optimality implies global optimality; this is because we are minimizing a convex function over a convex set.
Equivalent LP
Consider standard LP:
Let be a basis matrix corresponding to , i.e.,
and be the non basis matrix. The LP can be written as
We obtain
Hence:
- If , then is a feasible solution of the above LP.
- If in addition , then is an optimal solution of the above LP.
The feasibility and optimality condition: if , then the corresponding basic solution is feasible. If in addition , then the corresponding basic solution is an optimal solution.
If there is a non basis index such that , then increasing the value of can reduce the objective value.
Simplex tableau
Solution | |||
---|---|---|---|
This table is the simplex tableau of a standard LP, which encodes the following equations:
In practice we prefer to write out all components for the convenience of calculation and implementation of algorithm.
Solution | |||
---|---|---|---|
0 | |||
Simplex method: algorithm for a standard LP
- If , then the tableau is optimal. The corresponding optimal solution is , and the optimal value is .
- Otherwise, if there is and the column , then the objective value is .
- Otherwise, choose and find that minimizes .
- Apply Gauss-Jordan exchange with as pivoting element, that is:
Perform row elementary operations to the above simplex tableau so that the -th column becomes .
Let me briefly explain the idea of this algorithm. Simplex method start with an initial basic feasible solution. If all is negative then moving towards any direction cannot reduce the objective function, so we can conclude that this solution is locally optimal, and also globally optimal. Step 2 means that we can move from the initial solution to that direction without stopping, so the objective function keep decreasing and the minimum is . As for step 3, we choose a , increase the value of the corresponding , which lower the value of objective function, and at the same time change the value of corresponding to the basis, so that the constraints still hold all the time. We also find in Step 3, because eventually we move to a solution such that one of the basis becomes and we cannot continue moving towards this direction, otherwise this will be negative, which violates the nonnegative constraints. In fact this solution is also a basic feasible solution. Now one basis becomes and one non basis becomes positive (from ). We can say we have changed the basis. Step 4 can be regarded as some kind of normalization.
Numerical example: Consider an LP
The corresponding standard LP is
The initial simplex tableau is
Solution | |||||
---|---|---|---|---|---|
We first find a positive in the first row. The column of has a positive so we select to enter the basis. Now we compare of rows with positive : since , we remove from the basis. We transform the simplex tableau into
Solution | |||||
---|---|---|---|---|---|
Similarly, now we remove from the basis and add to the basis.
Solution | |||||
---|---|---|---|---|---|
We can find the optimality condition has been satisfied. Thus the optimal solution is (with ) and the minimum of the original LP is .
Theorem 4: Assume that the feasible set is nonempty and every basic feasible solution is nondegenerate. Then, the simplex method terminates after a finite number of iterations.
Degeneracy
Theorem 4 is under the assumption that all basic feasible solutions are nondegenerate. Suppose now that the exact same algorithm is used in the presence of degeneracy. Then, the following new possibilities may be encountered in the course of the algorithm.
- If the current basic feasible solution is degenerate, the step size of moving can be equal to zero, that is to say we move to the new basic feasible solution which is the same as . Nevertheless, we still change the basis and continue the algorithm.
- We move to a basic feasible solution that is degenerate, from a nondegenerate one.
On the one hand, basis changes while staying at the same basic feasible solution are not in vain. On the other hand, a sequence of basis changes might lead back to the initial basis, in which case the algorithm may loop indefinitely. This undesirable phenomenon is called cycling. But we can avoid cycling by using specific pivoting rule.
Here we introduce two pivoting rules. Cycling can be avoided if we use the second one.
Dantzig’s pivoting rule (for standard LP): 1. Entering variable: choose the largest to enter the basis; 2. leaving variable: choose the one with the smallest index to leave the basis.
Bland’s pivoting rule (for standard LP): 1. Entering variable: choose the one with smallest index to enter the basis; 2. leaving variable: choose the one with the smallest index to leave the basis.
Bland's pivoting rule is also known as the smallest subscript pivoting rule.
Theorem 5: The simplex method with Bland’s pivoting rule always terminates in finite number of steps.
Useful techniques for simplex method
We first multiply some of the constraints by , change the problem so that .
Big technique
Suppose is a sufficiently large positive constant. First solve a modified LP, of which the new objective function is , while the equality constraints have equations inserted respectively on the left side and the new variables are .
From the construction we can see that if is a basic feasible solution of the modified LP, then is a basic feasible solution of the original LP. And, if the original LP has at least one feasible solution, then the optimal solution of the modified LP must satisfy . Otherwise the problem is infeasible.
Two-phase technique
Phase I is to solve another LP whose constraints and variable are the same with those of big technique above, but the objective function is .
If the optimal solution satisfies , then go to Phase II. Otherwise, the initial problem is clearly infeasible.
Phase II is to solve the original LP starting from obtained from Phase I.
The numbers of iterations in the big technique and two-phase technique are the same. It is said that the advantage of two-phase technique rests in eliminating manipulation of the constant .
Simplex method for maximization
Simplex method can be easily adapted to maximization problem. The main difference is the sign of in all steps.
- If , then the tableau is optimal.
- Otherwise, if there is and the column , then the optimal value is .
- Otherwise, choose and ...
Duality theory
Duality is a powerful theoretical tool that has numerous applications, provides new geometric insights, and leads to another algorithm for linear programming (the dual simplex method).
Motivation
Duality theory can be motivated as an outgrowth of the Lagrange multiplier method, often used in calculus to minimize a function subject to equality constraints. For example, to minimize subject to , we can use the method called Lagrange multiplier. Define . We know that there is a such that are stationary points to while are defined as extreme points to the original constrained problem. We obtain , and the original t gives us the additional relation , so the optimal solution to the original problem is .
The main idea in the above example is the following. Instead of enforcing the hard constraint , we allow it to be violated and associate a Lagrange multiplier, or price, with the amount by which it is violated. This leads to the unconstrained minimization. When the price is properly chosen (, in our example), the optimal solution to the constrained problem is also optimal for the unconstrained problem. In particular, under that specific value of , the presence or absence of the hard constraint does not affect the optimal cost.
The situation in linear programming is similar.
Consider the standard form problem
which we call the primal problem, and let be an optimal solution, assumed to exist. We introduce a relaxed problem in which the constraint is replaced by a penalty .
Let be the optimal cost for the relaxed problem, as a function of the price vector . The relaxed problem allows for more options than those present in the primal problem, and we expect to be no larger than the optimal primal cost .
Thus, each leads to a lower bound for the optimal cost . The problem: [maximize , subject to: no constraint], can be then interpreted as a search for the tightest possible lower bound of this type, and is known as the dual problem. The main result in duality theory asserts that the optimal cost in the dual problem is equal to the optimal cost in the primal. In other words, when the prices are chosen according to an optimal solution for the dual problem, the option of violating the constraints is of no value.
Now we calculate . . Therefore, it is not hard to see that if , otherwise it is . In maximizing , we only need to consider those values of for which is not equal to . The dual problem is the same as the LP
The dual problem
When the primal is not standard, the dual can be written through the following table:
Primal | minimize | maximize | Dual |
---|---|---|---|
constraints | variables | ||
variables | constraints |
Theorem 6: If we transform the dual into an equivalent minimization problem and then form its dual, we obtain a problem equivalent to the original problem.
A compact statement that is often used to describe Theorem 6 is that "the dual of the dual is the primal".
Theorem 7: Suppose that we have transformed a linear programming problem to another linear programming problem , by a sequence of transformations of the following types:
- Replace a free variable with the difference of two nonnegative variables.
- Replace an inequality constraint by an equality constraint involving a nonnegative slack variable.
- If some row of the matrix in a feasible standard form problem is a linear combination of the other rows, eliminate the corresponding equality constraint.
Then, the duals of and are equivalent, i.e., they are either both infeasible, or they have the same optimal cost.
The duality theorem
We generalize the result mentioned before into the following theorem.
Theorem 8 (Weak duality): If is a feasible solution to the primal problem and is a feasible solution to the dual problem, then
The proof can be simple if we prove by . But for the purpose of some further theorems, we prove in another way. We define for any and that and . By definition of the dual, they are all nonnegative, so
The weak duality theorem is not a deep result, yet it does provide some useful information about the relation between the primal and the dual. We have, for example, the following corollary.
Corollary: If the optimal cost in the primal is , then the dual problem must be infeasible; and if the optimal cost in the dual is , then the primal problem must be infeasible.
Corollary: Let and be feasible solutions to the primal and the dual, respectively, and suppose that . Then, and are optimal solutions to the primal and the dual, respectively.
The next theorem is the central result on linear programming duality.
Theorem 9 (Strong duality): If a linear programming problem has an optimal solution, so does its dual, and the respective optimal costs are equal.
Proof. Here we only consider the standard form problem. The rest is not hard to see, so it is omitted. Let us assume temporarily that the rows of are linearly independent and that there exists an optimal solution. Let us apply the simplex method to this problem. As long as cycling is avoided, e.g., by using the Bland's pivoting rule, the simplex method terminates with an optimal solution and an optimal basis matrix . Let be the corresponding vector of basic variables. When the simplex method terminates, we obtain . Let us define a vector by letting , then we know this is a feasible solution to the dual problem. In addition, . It follows that is an optimal solution to the dual (cf. the Corollary above), and the optimal dual cost is equal to the optimal primal cost.
Finite optimum | Unbounded | Infeasible | |
---|---|---|---|
Finite optimum | Possible | ||
Unbounded | Possible | ||
Infeasible | Possible | Possible |
Here is a table about different possibilities for the primal and the dual. (The blank entries should be impossible.)
There is another interesting relation between the primal and the dual which is known as Clark's theorem. It asserts that unless both problems are in feasible, at least one of them must have an unbounded feasible set.
Complementary slackness: An important relation between primal and dual optimal solutions is provided by the complementary slackness conditions, which we present next.
Theorem 10 (Complementary slackness): Let and be feasible solutions to the primal and the dual problem, respectively. The vectors and are optimal solutions for the two respective problems iff
Proof. In the proof of Theorem 8, we defined and which are both nonnegative. In addition, we showed that . By the strong duality theorem, if and are optimal, then , which implies that for all . Conversely, if all and are zero, then , by the corollary of Theorem 8, and are optimal.
An intuitive explanation of this theorem is that a constraint which is not active at an optimal solution can be removed from the problem without affecting the optimal cost, and there is no point in associating a nonzero price with such a constraint.
If the primal problem is in standard form (where the first condition is automatically satisfied) and a nondegenerate (otherwise, these conditions may be of very little help in determining an optimal solution to the dual problem) optimal basic feasible solution is known, the complementary slackness conditions determine a unique solution to the dual problem. Here is the reason. The second complementary slackness condition implies that for basis , and hence . Since is inversible, is uniquely determined.
We finally mention that if the primal constraints are of the form , , and the primal problem has an optimal solution, then there exist optimal solutions to the primal and the dual which satisfy strict complementary slackness; that is, a variable in one problem is nonzero iff the corresponding constraint in the other problem is active. This result has some interesting applications in discrete optimization, but these lie outside the scope of this appendix.
- Consider a standard LP and its dual. Assume that both problems have an optimal solution. Fix some . Suppose that every optimal solution to the primal satisfies . Then there exists an optimal solution to the dual such that . (Hint: Let be the optimal cost. Consider the problem of minimizing subject to , , and , and form its dual.)
There exist optimal solutions and to the primal and to the dual, respectively, such that for every we have either or . (Hint: Use part (1) for each , and then take the average of the vectors obtained.)
Consider now the following LP problem and its dual:
Assume that both problems have an optimal solution. Then there exist optimal solutions to the primal and to the dual, respectively, that satisfy strict complementary slackness, that is:
(i) For every we have either or .
(ii) For every , we have either or .
(Hint: Convert the primal to standard form and apply part (2).)
Dual simplex method
Let us consider a problem in standard form, under the usual assumption that the rows of the matrix are linearly independent. We also consider the corresponding tableau. We do not require the last column to be nonnegative, which means that we have a basic, but not necessarily feasible solution to the primal problem. However, we assume that those are nonpositive; equivalently, the vector satisfies , and we have a feasible solution to the dual problem. If the inequality happens to hold, we also have a primal feasible solution with the same cost, and optimal solutions to both problems have been found. If the inequality fails to hold, we perform a change of basis in a manner we describe next. After that those in the new tableau will also be nonpositive and dual feasibility has been maintained. At the same time, the dual cost increases. We hope holds after some iterations, then we obtain an optimal solution (for both dual and primal problem).
An iteration for the dual simplex method:
- A typical iteration starts with the tableau associated with a basis matrix and with all nonpositive.
- If all the components of the last column are nonnegative, we have an optimal basic feasible solution and the algorithm terminates; else, choose some such that .
- Consider the -th row, if all , then the optimal dual cost is and the algorithm terminates.
- For each such that , find that minimize .
- Apply Gauss-Jordan exchange with as pivoting element.
Let us now consider the possibility that some is zero. In this case the algorithm may cycle. This can be avoided by employing a suitable anticycling rule.
Some extra notes
Farkas' theorem
Theorem 11 (Farkas' theorem): Let and . Then exactly one of the following two alternatives holds:
- There exists some such that .
- There exists some such that and .
Proof. One direction is easy. If there exists some satisfying , and if , then , which shows that the second alternative cannot hold.
If there is no satisfying . Consider the pair of problems
Note that the first is the dual of the second. The maximization problem is infeasible, which implies that the minimization problem is either unbounded or infeasible. Since is a feasible solution to the minimization problem, it follows that the minimization problem is unbounded. Therefore, there exists some which is feasible () and .
Compact simplex tableau
The simplex tableaux that we have been using so far are usually called full simplex tableaux. However, we can observe that the same amount of information will be retained if we delete the basic columns from the full simplex tableau to obtain a more compact simplex tableau.
Computational efficiency of the simplex method
This is determined by two factors. The first is the computational effort at each iteration, and the second is the number of iterations. For each iteration, using simplex tableau needs arithmetic operations. For the second, there exists an LP and a pivoting rule under which the simplex method requires changes of basis before it terminates.
Other methods
There are some other algorithms not mentioned here, such as the ellipsoid method, and the interior point methods.
发表回复