Within computer science and operations research, many combinatorial optimization problems are computationally intractable to solve exactly (to optimality). Many such problems do admit fast (polynomial time) approximation algorithms—that is, algorithms that are guaranteed to return an approximately optimal solution given any input.
Contents
- Overview
- Comparison to other applications of the probabilistic method
- Set Cover example
- Step 3 The randomized rounding step
- proof
- Derandomization using the method of conditional probabilities
- Bounding the conditional probability of failure
- Keeping the conditional probability of failure below 1
- Randomized rounding algorithm for Set Cover
- Remarks
- References
Randomized rounding (Raghavan & Tompson 1987) is a widely used approach for designing and analyzing such approximation algorithms. The basic idea is to use the probabilistic method to convert an optimal solution of a relaxation of the problem into an approximately optimal solution to the original problem.
Overview
The basic approach has three steps:
- Formulate the problem to be solved as an integer linear program (ILP).
- Compute an optimal fractional solution
x to the linear programming relaxation (LP) of the ILP. - Round the fractional solution
x of the LP to an integer solutionx ′
(Although the approach is most commonly applied with linear programs, other kinds of relaxations are sometimes used. For example, see Goeman's and Williamson's semi-definite programming-based Max-Cut approximation algorithm.)
The challenge in the first step is to choose a suitable integer linear program. Familiarity with linear programming is required, in particular, familiarity with how to model problems using linear programs and integer linear programs. But, for many problems, there is a natural integer linear program that works well, such as in the Set Cover example below. (The integer linear program should have a small integrality gap; indeed randomized rounding is often used to prove bounds on integrality gaps.)
In the second step, the optimal fractional solution can typically be computed in polynomial time using any standard linear programming algorithm.
In the third step, the fractional solution must be converted into an integer solution (and thus a solution to the original problem). This is called rounding the fractional solution. The resulting integer solution should (provably) have cost not much larger than the cost of the fractional solution. This will ensure that the cost of the integer solution is not much larger than the cost of the optimal integer solution.
The main technique used to do the third step (rounding) is to use randomization, and then to use probabilistic arguments to bound the increase in cost due to the rounding (following the probabilistic method from combinatorics). There, probabilistic arguments are used to show the existence of discrete structures with desired properties. In this context, one uses such arguments to show the following:
Given any fractional solutionFinally, to make the third step computationally efficient, one either shows that
Comparison to other applications of the probabilistic method
The randomized rounding step differs from most applications of the probabilistic method in two respects:
- The computational complexity of the rounding step is important. It should be implementable by a fast (e.g. polynomial time) algorithm.
- The probability distribution underlying the random experiment is a function of the solution
x of a relaxation of the problem instance. This fact is crucial to proving the performance guarantee of the approximation algorithm --- that is, that for any problem instance, the algorithm returns a solution that approximates the optimal solution for that specific instance. In comparison, applications of the probabilistic method in combinatorics typically show the existence of structures whose features depend on other parameters of the input. For example, consider Turán's theorem, which can be stated as "any graph withn vertices of average degreed must have an independent set of size at leastn / ( d + 1 ) . (See this for a probabilistic proof of Turán's theorem.) While there are graphs for which this bound is tight, there are also graphs which have independent sets much larger thann / ( d + 1 ) . Thus, the size of the independent set shown to exist by Turán's theorem in a graph may, in general, be much smaller than the maximum independent set for that graph.
Set Cover example
The method is best illustrated by example. The following example illustrates how randomized rounding can be used to design an approximation algorithm for the Set Cover problem.
Fix any instance
For step 1, let IP be the standard integer linear program for set cover for this instance.
For step 2, let LP be the linear programming relaxation of IP, and compute an optimal solution
(The feasible solutions to LP are the vectors
The optimal solution
is as small as possible.)
Note that any set cover
In other words, the linear program LP is a relaxation of the given set-cover problem.
Since
Step 3: The randomized rounding step
Here is a description of the third step—the rounding step, which must convert the minimum-cost fractional set cover
The rounding step should produce an
As a starting point, consider the most natural rounding scheme:
With this rounding scheme, the expected cost of the chosen sets is at most
So only a constant fraction of the elements will be covered in expectation.
To make
Scaling the probabilities up by
(Note: with care the
proof
The output
- the cost
c ⋅ x ′ x ′ 2 λ c ⋅ x ∗ - for some element
e ,x ′ e .
The expectation of each
For the remaining bad events (one for each element
(This uses the inequality
Thus, for each of the
By the naive union bound, the probability that one of the
Derandomization using the method of conditional probabilities
The lemma above shows the existence of a set cover of cost
One approach would be to increase
That approach weakens the approximation ratio. We next describe a different approach that yields a deterministic algorithm that is guaranteed to match the approximation ratio of the existence proof above. The approach is called the method of conditional probabilities.
The deterministic algorithm emulates the randomized rounding scheme: it considers each set
Bounding the conditional probability of failure
We want to be able to set each variable
where
is the set of elements left uncovered at the end.
The random variable
To apply the method of conditional probabilities, we need to extend the argument to bound the conditional probability of failure as the rounding step proceeds. Usually, this can be done in a systematic way, although it can be technically tedious.
So, what about the conditional probability of failure as the rounding step iterates through the sets? Since
Next we calculate the conditional expectation of
Note that
Keeping the conditional probability of failure below 1
To keep the conditional probability of failure below 1, it suffices to keep the conditional expectation of
(where
In the
To see why, focus on the point in time when iteration
Since a weighted average of two quantities is always at least the minimum of those two quantities, it follows that
Thus, setting
In detail, what does this mean? Considered as a function of
Thus, the algorithm should set
Randomized-rounding algorithm for Set Cover
input: set system
output: set cover
- Compute a min-cost fractional set cover
x ∗ - Let
λ ← ln ( 2 | U | ) . Letp s ← min ( λ x s ∗ , 1 ) for eachs ∈ S . - For each
s ′ ∈ S do:- Let
S ← S − { s ′ } . (S contains the not-yet-decided sets.) - If
c s ′ 2 λ c ⋅ x ∗ > ∑ e ∈ s ′ ∩ U ∏ s ∈ S , s ∋ e ( 1 − p s ) then setx s ′ ← 0 ,else setx s ′ ← 1 andU ← U − s ′ U contains the not-yet-covered elements.)
- Let
- Return
x ′
The algorithm ensures that the conditional expectation of
Remarks
In the example above, the algorithm was guided by the conditional expectation of a random variable