Definition:
– A theoretical measure of the execution of
an algorithm, usually the time or memory
needed, given the problem size n, which
is usually the number of items.
– Informally, saying some equation f(n) =
O(g(n)) means it is less than some
constant multiple of g(n). The notation is
read, "f of n is big oh of g of n".