Probabilistic method

From Wikipedia, the free encyclopedia

This article is not about probabilistic algorithms, which give the right answer with high probability but not with certainty, nor about Monte Carlo methods, which are simulations relying on pseudo-randomness.

The probabilistic method is a nonconstructive method, primarily used in combinatorics and pioneered by Paul Erdős, for proving the existence of a prescribed kind of mathematical object. This method has now been applied to other areas of mathematics such as number theory, linear algebra, and real analysis. It works by showing that if one randomly chooses objects from a specified class, the probability that the result is of the prescribed kind is more than zero. Although the proof uses probability, the final conclusion is determined for certain, without any possible error.

One way of doing this is by considering a randomly selected thing from a finite-sized universe. If the probability that the random thing satisfies certain properties is greater than zero, then this proves the existence of a thing that satisfies the properties. It doesn't matter if the probability is vanishingly small; any probability strictly greater than zero will do. Also, showing that the probability is (strictly) less than 1 can be used to prove the existence of an object that does not satisfy the prescribed properties.

Another way to use the probabilistic method is by calculating the expected value of some random variable. If it can be shown that the random variable can take on a value less than the expected value, this proves that the random variable can also take on some value greater than the expected value.

Common tools used in the probabilistic method include Markov's inequality, the Chernoff bound, and the Lovász local lemma.

Contents

[edit] Two examples due to Erdős

Although others before him proved theorems via the probabilistic method (for example, Szele's 1943 result that there exist tournaments containing a large number of Hamiltonian cycles), many of the most well known proofs using this method are due to Erdős. One such result is this 1947 proof giving a lower bound on the Ramsey number R(r, r; 2).

[edit] First example

Suppose we have a complete graph on n vertices. We wish to show (for small enough values of n) that it is possible to color the edges of the graph in two colors (say red and blue) so that there is no complete subgraph on r vertices which is monochromatic (every edge colored the same color).

To do so, we color the graph randomly. Color each edge independently with probability 1/2 of being red and 1/2 of being blue. We calculate the expected number of monochromatic subgraphs on r vertices as follows:

For any set S of r vertices from our graph, define the variable X(S) to be 1 if every edge amongst the r vertices is the same color, and 0 otherwise. Note that the number of monochromatic r-subgraphs is the sum of X(S) over all possible subsets. For any S, the expected value of X(S) is simply the probability that all of the {r \choose 2} edges in S are the same color,

2 \cdot 2^{-{r \choose 2}}

(the factor of 2 comes because there are two possible colors).

This holds true for any of the C(n,r) possible subsets we could have chosen, so we have that the sum of E[X(S)] over all S is

{n \choose r}2^{1-{r \choose 2}}

The sum of an expectation is the expectation of the sum (regardless of whether the variables are independent), so the expectation of the sum (the expected number of monochromatic r-subgraphs) is

{n \choose r}2^{1-{r \choose 2}}

Consider what happens if this value is less than 1. The number of monochromatic r-subgraphs in our random coloring will always be an integer, so must for at least one coloring be less than the expected value. But the only integer which satisfies this criterion is 0. Thus if

C(n,r) < 2^{{r \choose 2} - 1},

some coloring fits our desired criterion, so by definition R(r, r; 2) must be bigger than n. In particular, R(r, r; 2) must grow at least exponentially with r.

A peculiarity of this argument is that it is entirely nonconstructive. Even though it proves (for example) that almost every coloring of the complete graph on (1.1)r vertices contains no monochromatic r-subgraph, it gives no explicit example of such a coloring. The problem of finding such a coloring has been open for more than 50 years.

[edit] Second example

A 1959 paper of Erdős (see reference cited below) addressed the following problem in graph theory: given positive integers g and k, does there exist a graph G containing no cycles of length at most g, such that the chromatic number of G is at least k?

It can be shown that such a graph exists for any g and k, and the proof is reasonably simple. Let n be very large and consider a random graph G on n vertices, where every edge in G exists with probability n(1-g)/g. It can be shown that with positive probability, the following two properties hold:

  • G contains no stable set of size \lceil n/2k \rceil
  • G contains at most n/2 cycles of length at most g.

Here comes the trick: since G has these two properties, we can remove at most n/2 vertices from G to obtain a new graph G' on n' vertices that contains no cycles of length at most g. We can see that this new graph has no stable set of size \lceil n'/k \rceil, hence G' has chromatic number at least k.

This result gives a hint as to why the computation of the chromatic number of a graph is so difficult: even when there are no local reasons (such as small cycles) for a graph to require many colors the chromatic number can still be arbitrarily large.

[edit] References