Definition:Big-O Notation/Sequence/Definition 1

From ProofWiki
Jump to navigation Jump to search

Definition

Let $g: \N \to \R$ be a real sequence, expressed here as a real-valued function on the set of natural numbers $\N$.

Then $\map \OO g$ is defined as:

$\map \OO g = \set {f: \N \to \R: \exists c \in \R_{>0}: \exists n_0 \in \N: \forall n > n_0: 0 \le \size {\map f n} \le c \cdot \size {\map g n} }$


Notation

The expression $\map f n \in \map \OO {\map g n}$ is read as:

$\map f n$ is big-O of $\map g n$


While it is correct and accurate to write:

$\map f n \in \map \OO {\map g n}$

it is a common abuse of notation to write:

$\map f n = \map \OO {\map g n}$

This notation offers some advantages.


Also defined as

Some authors require that $b_n$ be nonzero for $n$ sufficiently large.

Some authors require that the functions appearing in the $\OO$-estimate be positive or strictly positive.


Also denoted as

The big-$\OO$ notation, along with little-$\oo$ notation, are also referred to as Landau's symbols or the Landau symbols, for Edmund Georg Hermann Landau.


In analytic number theory, sometimes Vinogradov's notations $f \ll g$ or $g \gg f$ are used to mean $f = \map \OO g$.

This can often be clearer for estimates leading to typographically complex error terms.


Some sources use an ordinary $O$:

$f = \map O g$


Also see


Sources