# Calculating the Fibonacci numbers

The Fibonacci sequence is defined as the sequence that satisfies the following properties:

1. It starts with 0, 1
2. Every subsequent term is the sum of the two preceding terms

So, it starts as 0, 1, 1, 2, 3, 5, 8, 13, …

Mathematically, it is defined as

Now, let’s write a fast Python function to calculate the th Fibonacci number. (Python is not normally a language I’d use to write something that needs to be fast, but this is just for fun. In this case I choose it because its built-in big number support is so convenient. It could also be written in, say, C++, although you’d need something like boost to provide the support for big numbers.)

## A naive implementation

Naively, we can code the mathematical definition into a recursive function:

def fibonacci_naive(n):
if n <= 1:
return n
return fibonacci_naive(n - 1) + fibonacci_naive(n - 2)


However, it turns out that this is very slow, even just calculating already takes more than a second on my machine. Even worse, the runtime of fibonacci_naive(n) seems to increase very rapidly when we increase n: calculating already takes more than five seconds.

If we think about this function a bit, we can see that the only way the recursion can “end”, is when it returns 0 or 1. So to calculate , at least function calls need to be made. Now, turns out to be exponential as a function of . The runtime of fibonacci_naive(n) is at least exponential in n.

## A simple and fast implementation

Noting that

we can come up with the following simple algorithm:

def fibonacci_linear(n):
f_n, f_nm1 = 0, 1
for _ in range(0, n):
f_n, f_nm1 = f_n + f_nm1, f_n
return f_n


This is blazing fast compared to the naive algorithm. Calculating , a number with 2090 digits, takes about a millisecond on my machine.

We note that the number of operations that fibonacci_linear(n) does is linear in n. However, since the length of the numbers involved is also linear in n, the runtime is quadratic. Usually you want to steer clear of quadratic algorithms, but this particular algorithm is probably good enough for most applications.

## Using matrices

If we really want to be a tryhard we can note that we can turn the recursive definition into a matrix equation:

So now we want to calculate the exponential of a matrix. This can be done in operations by using the Russian peasant algorithm (which is typically presented as an algorithm for multiplication, but it works for exponentiation as well if we replace the doubling by squaring and the addition by multiplication).

The idea is that to compute we loop over the bit positions b = 0, 1, 2, ... of n and keep in a variable temp. To find , we can simply square . Then we have an accumulator acc that is initialized to 1 (or the identity matrix, or whatever), and we multiply it by whenever the bit at position b in n is set. When the function terminates, we have

in the accumulator variable acc. I used to denote the bit positions that are 1 in the binary representation of , so that .

Putting this together we get

class Matrix:
def __init__(self, a, b, c, d):
self.a, self.b, self.c, self.d = a, b, c, d

def __mul__(self, other):
return Matrix(self.a * other.a + self.b * other.c,
self.a * other.b + self.b * other.d,
self.c * other.a + self.d * other.c,
self.c * other.b + self.d * other.d)

def russian_peasant_exponentiation(x, n, id):
acc = id
temp = x
while n > 0:
if n & 1:
acc *= temp
temp *= temp
n >>= 1
return acc

def fibonacci_matrix(n):
fibonacci_matrix = Matrix(1, 1, 1, 0)
identity_matrix = Matrix(1, 0, 0, 1)
return russian_peasant_exponentiation(fibonacci_matrix, n, identity_matrix).b


Granted, this could be optimized further by inlining and simplifying things, but doing so would destory the whole idea of using a matrix, and we’re going to simplify everything by using some mathematical identities anyway in the next section.

## Micro-optimizing using math

Now we note that

and

We see here that the entry in the right top of the right matrix is calculated as even though is already in the matrix on the left. From this we already see that the previous algorithm, which uses this exact algorithm, is not optimal.

From these two identities, it follows by induction that

Using we see

Looking at the element in the left top and the element in the right bottom of the left and right side, we obtain the identities

If we set we get

Now, if we work with pairs of the form , we know how to “double” the index (e.g. obtain from ), and how to “add” (e.g., obtain from and ). So we can use the Russian peasant multiplication algorithm again. Factoring out some common expressions, we get:

def fibonacci_fast(n):
acc = (0, 1)
temp = (1, 0)
while n > 0:
if n & 1:
square = acc * temp
acc = (square + acc * temp + acc * temp,
square + acc * temp)
square = temp * temp
temp = (square + 2 * temp * temp,
square + temp * temp)
n >>= 1
return acc


And this is the fastest code I could write to calculate the Fibonacci numbers (in Python, that is). It’s runtime should be comparable to the runtime fibonacci_matrix since it’s still using the Russian peasant exponentiation scheme. However, it should be a bit faster since we got rid of a lot of overhead (function calls, lists, duplicate computations).

## Benchmarks

Note that both axes are logarithmic. In this type of graphs, graphs of the form map to a straight line. Two graphs and have the same slope if the exponents and are the same. They have a vertical offset if the constants and are different. Some observations:

• fibonacci_naive is so slow that it doesn’t even make sense to compare it to the other methods
• fibonacci_linear is much better, but it is still asymptotically slower than fibonacci_matrix and fibonacci_fast (although for small n – say, smaller than 250, it might actually be faster)
• fibonacci_fast is roughly 2.5 times as fast as fibonacci_matrix