Jump to content

Big O notation

From Wikipedia, the free encyclopedia

This is an old revision of this page, as edited by 128.187.0.164 (talk) at 03:39, 7 January 2006 (→‎Related asymptotic notations: ''O'', ''o'', Ω, ω, Θ, ''Õ''). The present address (URL) is a permanent link to this revision, which may differ significantly from the current revision.

Big O notation is a mathematical notation used to describe the asymptotic behavior of functions. More precisely, it is used to describe an asymptotic upper bound for the magnitude of a function in terms of another, usually simpler, function. It has two main areas of application: in mathematics, it is usually used to characterize the residual term of a truncated infinite series, especially an asymptotic series, and in computer science, it is useful in the analysis of the complexity of algorithms.

It was first introduced by German number theorist Paul Bachmann in his 1892 book Analytische Zahlentheorie. The notation was popularized in the work of another German number theorist Edmund Landau, hence it is sometimes called a Landau symbol. The big-O, standing for "order of", was originally a capital omicron; today the capital letter O is used, but never the digit zero.

Uses

There are two formally close, but noticeably different usages of this notation: infinite asymptotics and infinitesimal asymptotics. This distinction is only in application and not in principle, however—the formal definition for the "big O" is the same for both cases, only with different limits for the function argument.

Infinite asymptotics

Big O notation is useful when analyzing algorithms for efficiency. For example, the time (or the number of steps) it takes to complete a problem of size n might be found to be T(n) = 4n² - 2n + 2.

As n grows large, the n² term will come to dominate, so that all other terms can be neglected—for instance when n = 500, the term 4n² is 1000 times as large as the 2n term, and so ignoring the latter would have negligible effect on the expression's value for most purposes.

Further, the coefficients become irrelevant as well if we compare to any other order of expression, such as an expression containing a term n³ or 2n. Even if T(n) = 1,000,000n², if U(n) = n³, the latter will always exceed the former once n grows larger than 1,000,000 (T(1,000,000) = 1,000,000³ = U(1,000,000)).

So Big O notation captures what remains: we write

and say that the algorithm has order of n2 time complexity.

Infinitesimal asymptotics

Big O can also be used to describe the error term in an approximation to a mathematical function. For example,

expresses the fact that the error (the difference ) is smaller in absolute value than some constant times x3 if x is close enough to 0.

Formal definition

Suppose f(x) and g(x) are two functions defined on some subset of the real numbers. We say

f(x) is O(g(x)) as x

if and only if

there exist numbers x0 and M such that |f(x)| ≤ M |g(x)| for x > x0.

The notation can also be used to describe the behavior of f near some real number a: we say

f(x) is O(g(x)) as x a

if and only if

there exists numbers δ>0 and M such that |f(x)| ≤ M |g(x)| for |x - a| < δ.

If g(x) is non-zero for values of x sufficiently close to a, both of these definitions can be unified using the limit superior:

f(x) is O(g(x)) as x a

if and only if

In mathematics, both asymptotic behaviors near ∞ and near a are considered. In computational complexity theory, only asymptotics near ∞ are used; furthermore, only positive functions are considered, so the absolute value bars may be left out.

Example

Take the polynomials:

We say f(x) has order O(g(x)) or O(x4). From the definition of order, |f(x)| ≤ C |g(x)| for all x>1, where C is a constant.

Proof:

        where x > 1
    because x3 < x4, and so on.

Matters of notation

The statement "f(x) is O(g(x))" as defined above is often written as f(x) = O(g(x)). This is a slight abuse of notation: we are not really asserting the equality of two functions. The property of being O(g(x)) is not symmetric:

but .

For this reason, some authors prefer a set notation and write f O(g), thinking of O(g) as the set of all functions dominated by g.

Furthermore, an "equation" of the form

should be understood as "the difference of f(x) and h(x) is O(g(x))".

Common orders of functions

Here is a list of classes of functions that are commonly encountered when analyzing algorithms. All of these are as n increases to infinity. The slower-growing functions are listed first. c is an arbitrary constant.

notation name
constant
iterative logarithmic
logarithmic
polylogarithmic
sublinear
linear
linearithmic, loglinear, quasilinear or supralinear
quadratic
, polynomial, sometimes called "algebraic"
exponential, sometimes called "geometric"
factorial, sometimes called "combinatorial"

Not as common, but even larger growth is possible, such as the single-valued version of the Ackermann function, A(n,n). Conversely, extremely slowly-growing functions such as the inverse of this function, often denoted α(n), are possible. Although unbounded, these functions are often regarded as being constant factors for all practical purposes.

Properties

If a function f(n) can be written as a finite sum of other functions, then the fastest growing one determines the order of f(n). For example

.

In particular, if a function may be bounded by a polynomial in n, then as n tends to infinity, one may disregard lower-order terms of the polynomial.

O(nc) and O(cn) are very different. The latter grows much, much faster, no matter how big the constant c is (so long as it is greater than one). A function that grows faster than any power of n is called superpolynomial. One that grows slower than any exponential function of the form cn is called subexponential. An algorithm can require time that is both superpolynomial and subexponential; examples of this include the fastest known algorithms for integer factorization.

O(log n) is exactly the same as O(log(nc)). The logarithms differ only by a constant factor, (since log(nc)=c log n) and thus the big O notation ignores that. Similarly, logs with different constant bases are equivalent.

Product

Sum

Multiplication by a constant

, k≠0

Addition of a constant

unless g(n) ∈ o(1), in which case it is O(1).

Other useful relations are given in section Big O and little o below.

Related asymptotic notations: O, o, Ω, ω, Θ, Õ

Big O is the most commonly used asymptotic notation for comparing functions, although it is often actually an informal substitute for Θ (Theta, see below). Here, we define some related notations in terms of "big O":

Notation Definition Mathematical definition
asymptotic upper bound
asymptotically negligible
asymptotic lower bound
asymptotically dominant
asymptotically tight bound and

Here "lim sup" and "lim inf" denote limit superior and limit inferior. When the limit exists, it is the same as both the lim sup and lim inf.

(A mnemonic for these Greek letters is that "omicron" can be read "o-micron", i.e., "o-small", whereas "omega" can be read "o-mega" or "o-big".)

The relation f(n) = o(g(n)) is read as "f(n) is little-oh of g(n)". Intuitively, it means that g(n) grows much faster than f(n). Formally, it states that the limit of f(n)/g(n) is zero.

Aside from big-O, the notations Θ and Ω are the two most often used in computer science; the lower-case o is common in mathematics but rarer in computer science. The lower-case ω is rarely used.

In casual use, O is commonly used where Θ is meant, i.e., when a tight estimate is implied. For example, one might say "heapsort is O(n log n) in the average case" when the intended meaning was "heapsort is Θ(n log n) in the average case". Both statements are true, but the latter is a stronger claim.

Another notation sometimes used in computer science is Õ (read Soft-O). f(n) = Õ(g(n)) is shorthand for f(n) = O(g(n) logkg(n)) for some k. Essentially, it is Big-O, ignoring logarithmic factors. This notation is often used to describe a class of "nitpicking" estimates (since logkn is always o(n) for any constant k).

Big O and little o

The following properties can be useful:

  • o(f ) + o(f ) ∈ o(f )
  • o(f ) o(g ) ∈ o(fg )
  • o(o(f )) ∈ o(f )
  • o(f ) ∈ O(f ) (and thus the above properties apply with most combinations of o and O).

Multiple variables

Big O (and little o, and Ω...) can also be used with multiple variables. For example, the statement

asserts that there exist constants C and N such that

To avoid ambiguity, the running variable should always be specified: the statement

is quite different from

External links

References

  • Marian Slodička. Mathematical Analysis I. University of Ghent, 2003.
  • Donald Knuth. The Art of Computer Programming, Volume 1: Fundamental Algorithms, Third Edition. Addison-Wesley, 1997. ISBN 0-201-89683-4. Section 1.2.11: Asymptotic Representations, pp.107–123.
  • Thomas H. Cormen, Charles E. Leiserson, Ronald L. Rivest, and Clifford Stein. Introduction to Algorithms, Second Edition. MIT Press and McGraw-Hill, 2001. ISBN 0262032937. Section 3.1: Asymptotic notation, pp.41–50.
  • . ISBN 0-534-94728-X. {{cite book}}: Missing or empty |title= (help); Unknown parameter |Author= ignored (|author= suggested) (help); Unknown parameter |Publisher= ignored (|publisher= suggested) (help); Unknown parameter |Title= ignored (|title= suggested) (help); Unknown parameter |Year= ignored (|year= suggested) (help) Pages 226–228 of section 7.1: Measuring complexity.