Home > Term: little-o notation
little-o notation
A theoretical measure of the execution of an algorithm, usually the time or memory needed, given the problem size n, which is usually the number of items. Informally, saying some equation f(n) = o(g(n)) means f(n) becomes insignificant relative to g(n) as n approaches infinity. The notation is read, "f of n is little oh of g of n". Formal Definition: f(n) = o(g(n)) means for all c > 0 there exists some k > 0 such that 0 ≤ f(n) < cg(n) for all n ≥ k. The value of k must not depend on n, but may depend on c.
- Μέρος του λόγου: noun
- Κλάδος/Τομέας: Υπολογιστές
- Category: Αλγόριθμοι & δομές
- Government Agency: NIST
0
Δημιουργός
- GeorgeV
- 100% positive feedback