Big O notation

From Wikipedia, the free encyclopedia

For other uses, see Big O.

Big O notation or Big Oh notation, and also Landau notation or asymptotic notation, is a mathematical notation used to describe the asymptotic behavior of functions. Its purpose is to characterize a function's behavior for very large (or very small) inputs in a simple but rigorous way that enables comparison to other functions. More precisely, the symbol O is used to describe an asymptotic upper bound for the magnitude of a function in terms of another, usually simpler, function. There are also other symbols o, Ω, ω, and Θ for various other upper, lower, and tight bounds. It has two main areas of application: in mathematics, it is usually used to characterize the residual term of a truncated infinite series, especially an asymptotic series, and in computer science, it is useful in the analysis of the complexity of algorithms.

Informally, the O notation is commonly employed to describe an asymptotic tight bound, but tight bounds are more formally and precisely denoted by the Θ (capital theta) symbol as described below.

It was first introduced by German number theorist Paul Bachmann in 1894, in the second volume of his book Analytische Zahlentheorie ("analytic number theory"), the first volume of which (not yet containing big O notation) came out in 1892. The notation was popularized in the work of another German number theorist Edmund Landau, hence it is sometimes called a Landau symbol. The big-O, standing for "order of", was originally a capital omicron; today the identical-looking Latin capital letter O is also used, but never the digit zero.

Contents

[edit] Uses

There are two formally close, but noticeably different usages of this notation: infinite asymptotics and infinitesimal asymptotics. This distinction is only in application and not in principle, however—the formal definition for the "big O" is the same for both cases, only with different limits for the function argument.

[edit] Infinite asymptotics

Big O notation is useful when analyzing algorithms for efficiency. For example, the time (or the number of steps) it takes to complete a problem of size n might be found to be T(n) = 4n² - 2n + 2.

As n grows large, the n² term will come to dominate, so that all other terms can be neglected—for instance when n = 500, the term 4n² is 1000 times as large as the 2n term, and so ignoring the latter would have negligible effect on the expression's value for most purposes.

Further, the coefficients become irrelevant as well if we compare to any other order of expression, such as an expression containing a term n³ or 2n. Even if T(n) = 1,000,000n², if U(n) = n³, the latter will always exceed the former once n grows larger than 1,000,000 (T(1,000,000) = 1,000,000³ = U(1,000,000)).

So the big O notation captures what remains: we write

T(n)\in O(n^2)

(pronounced "big O of n squared") and say that the algorithm has order of n²   time complexity.

[edit] Infinitesimal asymptotics

Big O can also be used to describe the error term in an approximation to a mathematical function. For example,

e^x=1+x+\frac{x^2}{2}+\hbox{O}(x^3)\qquad\hbox{as}\ x\to 0

expresses the fact that the error, the difference e^x - \left(1 + x +\frac{x^2}{2}\right), is smaller in absolute value than some constant times |x|3 when x is close enough to 0.

[edit] Formal definition

Suppose f(x) and g(x) are two functions defined on some subset of the real numbers. We say

f(x)\mbox{ is }O(g(x))\mbox{ as }x\to\infty

if and only if

\exists \;x_0,\exists \;M>0\mbox{ such that } |f(x)| \le \; M |g(x)|\mbox{ for }x>x_0.

The notation can also be used to describe the behavior of f near some real number a: we say

f(x)\mbox{ is }O(g(x))\mbox{ as }x\to a

if and only if

\exists \;\delta >0,\exists \; M>0\mbox{ such that }|f(x)| \le \; M |g(x)|\mbox{ for }|x - a| < \delta.

If g(x) is non-zero for values of x sufficiently close to a, both of these definitions can be unified using the limit superior:

f(x)\mbox{ is }O(g(x))\mbox{ as }x \to a

if and only if

\limsup_{x\to a} \left|\frac{f(x)}{g(x)}\right| < \infty.
Theory of O-Notation: f is in the order of g (f(x) = O(g(x))) if and only if there exists a positive real number M and a real number x0 such that for all X,|f(x)| <= M . |g(x)|, wherever x > x0
Enlarge
Theory of O-Notation: f is in the order of g (f(x) = O(g(x))) if and only if there exists a positive real number M and a real number x0 such that for all X,
|f(x)| <= M . |g(x)|, wherever x > x0

In mathematics, both asymptotic behaviours near ∞ and near a are considered. In computational complexity theory, only asymptotics near ∞ are used; furthermore, only positive functions are considered, so the absolute value bars may be left out.

[edit] Example

Take the polynomials:

f(x) = 6x^4 -2x^3 +5 \,
g(x) = x^4.  \,

We say f(x) has order O(g(x)) or O(x4). From the definition of order, |f(x)| ≤ C |g(x)| for all x>1, where C is a constant.

Proof:

|6x^4 - 2x^3 + 5| \le 6x^4 + 2x^3 + 5 \,         where x > 1
|6x^4 - 2x^3 + 5| \le 6x^4 + 2x^4 + 5x^4 \,     because x3 < x4, and so on.
|6x^4 - 2x^3 + 5| \le 13x^4 \,
|6x^4 - 2x^3 + 5| \le 13 \,|x^4 |. \,                       where C = 13 in this example

[edit] Matters of notation

The statement "f(x) is O(g(x))" as defined above is usually written as f(x) = O(g(x)). This is a slight abuse of notation; equality of two functions is not asserted. The property of being O(g(x)) is not symmetric:

O(x)=O(x^2)\mbox{ but }O(x^2)\ne O(x).

For this reason, some authors prefer set notation and write f \in O(g), thinking of O(g) as the set of all functions dominated by g.

In more complex usage, O( ) can appear in different places in an equation, even several times on each side. For example, the following are true for n\to\infty

(n + 1)2 = n2 + O(n)
(n+O(n^{1/2}))(n + O(\log\,n))^2 = n^3 + O(n^{5/2})
nO(1) = O(en)

The meaning of such statements is as follows: for any functions which satisfy each O( ) on the left side, there are some functions satisfying each O( ) on the right side, such that substituting all these functions into the equation makes the two sides equal. For example, the third equation above means: "For any function f(n)=O(1), there is some function g(n)=O(en) such that nf(n)=g(n)." In terms of the "set notation" above, the meaning is that the class of functions represented by the left side is a subset of the class of functions represented by the right side.

[edit] Common orders of functions

Here is a list of classes of functions that are commonly encountered when analyzing algorithms. All of these are as n increases to infinity. The slower-growing functions are listed first. c is an arbitrary constant.

Notation Name Example
O(1) constant Determining if a number is even or odd
O(log* n) iterated logarithmic The find algorithm of Hopcroft and Ullman on a disjoint set
O(log n) logarithmic Finding an item in a sorted list with the binary search algorithm
O((log n)c) polylogarithmic Deciding if n is prime with the AKS primality test
O(n) linear Finding an item in an unsorted list
O(n log n) linearithmic, loglinear, or quasilinear Sorting a list with heapsort
O(n2) quadratic Sorting a list with insertion sort
O(nc), c > 1 polynomial, sometimes called algebraic Finding the shortest path on a weighted digraph with the Floyd-Warshall algorithm
O(cn) exponential, sometimes called geometric Finding the (exact) solution to the traveling salesman problem
O(n!) factorial, sometimes called combinatorial Determining if two logical statements are equivalent [1]
O\left(2^{c^n}\right) double exponential Finding a complete set of AC-unifiers [2]

Not as common, but even larger growth is possible, such as the single-valued version of the Ackermann function, A(n,n). Conversely, extremely slowly-growing functions such as the inverse of this function, often denoted α(n), are possible. Although unbounded, these functions are often regarded as being constant factors for all practical purposes.

[edit] Properties

If a function f(n) can be written as a finite sum of other functions, then the fastest growing one determines the order of f(n). For example

f(n) = 10 \log n + 5 (\log n)^3 + 7n + 3n^2 + 6n^3 \in \hbox{O}(n^3)\,\!.

In particular, if a function may be bounded by a polynomial in n, then as n tends to infinity, one may disregard lower-order terms of the polynomial.

O(nc) and O(cn) are very different. The latter grows much, much faster, no matter how big the constant c is (as long as it is greater than one). A function that grows faster than any power of n is called superpolynomial. One that grows slower than any exponential function of the form cn is called subexponential. An algorithm can require time that is both superpolynomial and subexponential; examples of this include the fastest known algorithms for integer factorization.

O(logn) is exactly the same as O(log(nc)). The logarithms differ only by a constant factor, (since log(nc) = clogn) and thus the big O notation ignores that. Similarly, logs with different constant bases are equivalent. Exponentials with different bases, on the other hand, are not of the same order; assuming that they are is a common mistake. For example, 2n and 3n are not of the same order.

Changing units may or may not affect the order of the resulting algorithm. Changing units is equivalent to multiplying the appropriate variable by a constant wherever it appears. For example, if an algorithm runs in the order of n2, replacing n by cn means the algorithm runs in the order of c2n2, and the big O notation ignores the constant c2. This can be written as c^2n^2 \in O(n^2). If, however, an algorithm runs in the order of 2n, replacing n with cn gives 2cn = (2c)n. This is not equivalent to 2n (unless, of course, c=1).

[edit] Product

f_1(n)  f_2(n)=O(g_1(n)  g_2(n))\,

[edit] Sum

O(f(n)) + O(g(n)) = O(\max \lbrace |f(n)|,|g(n)| \rbrace) \,

[edit] Multiplication by a constant

O(k g(n)) = O(g(n)),\quad k \ne 0

Other useful relations are given in section Big O and little o below.

[edit] Related asymptotic notations: O, o, Ω, ω, Θ, Õ

Big O is the most commonly used asymptotic notation for comparing functions, although it is often actually an informal substitute for Θ (Theta, see below). Here, we define some related notations in terms of "big O":

Notation Definition Mathematical definition Alternative definition
f(n) \in O(g(n)) asymptotic upper bound \limsup_{n \to \infty} \left|\frac{f(n)}{g(n)}\right| < \infty \exists \;x_0,\exists \;M>0\mbox{ such that } |f(x)| \le \; M |g(x)|\mbox{ for all }x>x_0.
f(n) \in \Omega(g(n)) asymptotic lower bound \liminf_{n \to \infty} \left|\frac{f(n)}{g(n)}\right| > 0 \exists \;x_0,\exists \;M>0\mbox{ such that } |f(x)| \ge \; M |g(x)|\mbox{ for all }x>x_0.
f(n) \in \Theta(g(n)) asymptotically tight bound 0 < \liminf_{n \to \infty} \left|\frac{f(n)}{g(n)}\right| \leq \limsup_{n \to \infty} \left|\frac{f(n)}{g(n)}\right|< \infty f(n) \in O(g(n)) \mbox{ and } f(n) \in \Omega(g(n))
f(n) \in o(g(n)) asymptotically negligible \lim_{n \to \infty} \left|\frac{f(n)}{g(n)}\right| = 0 \forall M>0 \; \exists \;x_0 \mbox{ such that } |f(x)| < \; M |g(x)| \mbox{ for all }x>x_0.
f(n) \in \omega(g(n)) asymptotically dominant \lim_{n \to \infty} \left|\frac{f(n)}{g(n)}\right| = \infty \forall M \; \exists \;x_0 \mbox{ such that } |f(x)| > \; M |g(x)| \mbox{ for all }x>x_0.

(A mnemonic for these Greek letters is that "omicron" can be read "o-micron", i.e., "o-small", whereas "omega" can be read "o-mega" or "o-big".)

The relation f(n) = o(g(n)) is read as "f(n) is little-oh of g(n)". Intuitively, it means that g(n) grows much faster than f(n). Formally, it states that the limit of f(n) / g(n) is zero.

Aside from big-O, the notations Θ and Ω are the two most often used in computer science; the lower-case o is common in mathematics but rarer in computer science. The lower-case ω is rarely used.

Another notation sometimes used in computer science is Õ (read soft-O). f(n) = \tilde{O} (g(n)) is shorthand for f(n) = O(g(n)logkg(n)) for some k. Essentially, it is Big-O, ignoring logarithmic factors. This notation is often used to describe a class of "nitpicking" estimates (since logkn is always o(nε) for any constant k and any ε > 0).

The following properties can be useful:

  • o(f) + o(f) \subseteq o(f)
  • o(f)o(g) \subseteq o(fg)
  • o(o(f)) \subseteq o(f)
  • o(f) \subset O(f) (and thus the above properties apply with most combinations of o and O).

[edit] Multiple variables

Big O (and little o, and Ω...) can also be used with multiple variables. For example, the statement

f(n,m) = n^2 + m^3 + \hbox{O}(n+m) \mbox{ as } n,m\to\infty

asserts that there exist constants C and N such that

\forall n, m>N: |g(n,m)| \le C(n+m).

where g(n,m) is defined by

f(n,m) = n2 + m3 + g(n,m).

To avoid ambiguity, the running variable should always be specified: the statement

f(n,m) = \hbox{O}(n^m) \mbox{ as } n,m\to\infty

is quite different from

\forall m: f(n,m) = \hbox{O}(n^m) \mbox{ as } n\to\infty.

[edit] Graph theory

It is often useful to bound the running time of graph algorithms. Unlike most other computational problems, in graphs, there are two relevant parameters describing the size of the input, |V| and |E|; |V| is the number of vertices in the graph, while |E| is the number of edges in the graph. Inside asymptotic notation (and only there), it is common to use the symbols V and E, when someone really means |V| and |E|. We adopt this convention here to simplify asymptotic functions and make them easily readable. Keep in mind that the symbols V and E are never used inside asymptotic notation with their literal meaning, so there is no risk of ambiguity. For example O(E + VlogV) means O((E,V) \mapsto |E| + |V|\cdot\log|V|) for a suitable metric of graphs.

[edit] Generalizations and related usages

The generalization to functions taking values in any normed vector space is straightforward (replacing absolute values by norms), where f and g need not take their values in the same space. A generalization to functions g taking values in any topological group is also possible.

The "limiting process" x→xo can also be generalized by introducing an arbitrary filter base, i.e. to directed nets f and g.

The o notation can be used to define derivatives and differentiability in quite general spaces, and also (asymptotical) equivalence of functions,

f\sim g \iff (f-g) = o(g)

which is an equivalence relation and a more restrictive notion than the relationship "f is Θ(g)" from above. (It reduces to \lim f/g = 1 if f and g are positive real valued functions.) For example, 2x is Θ(x), but 2x − x is not o(x).

[edit] See also

[edit] References