Python provides a powerful set of tools to create and manipulate lists of data. In this part, we take a deep dive into the Python list type. We use Python lists to implement and optimize the Sieve of Eratosthenes, which will produce a list of all prime numbers up to a big number (like 10 million) in a snap. Along the way, we introduce some Python techniques for mathematical functions and data analysis. This programming lesson is meant to complement Chapter 2 of An Illustrated Theory of Numbers, and mathematical background can be found there.

Before diving into lists, we recall the brute force primality test that we created in the last lesson. To test whether a number n is prime, we can simply check for factors. This yields the following primality test.

In [ ]:

defis_prime(n):''' Checks whether the argument n is a prime number. Uses a brute force search for factors between 1 and n. '''forjinrange(2,n):# the list of numbers 2,3,...,n-1.ifn%j == 0: # is n divisible by j?
print"%d is a factor of %d."%(j,n)returnFalsereturnTrue

We can also implement this test with a while loop instead of a for loop. This will save some memory (space), since it avoids the storage of the list range(2,n) of numbers.

In [ ]:

defis_prime(n):''' Checks whether the argument n is a prime number. Uses a brute force search for factors between 1 and n. '''j=2whilej<n:# j will proceed through the list of numbers 2,3,...,n-1.ifn%j == 0: # is n divisible by j?
print"%d is a factor of %d."%(j,n)returnFalsej=j+1# There's a Python abbreviation for this: j += 1.returnTrue

In [ ]:

is_prime(10001)

In [ ]:

is_prime(101)

If $n$ is a prime number, then the is_prime(n) function will iterate through all the numbers between $2$ and $n-1$. But this is overkill! Indeed, if $n$ is not prime, it will have a factor between $2$ and the square root of $n$. This is because factors come in pairs: if $ab = n$, then one of the factors, $a$ or $b$, must be less than or equal to the square root of $n$. So it suffices to search for factors up to (and including) the square root of $n$.

We haven't worked with square roots in Python yet. But Python comes with a standard math package which enables square roots, trig functions, logs, and more. Click the previous link for documentation. This package doesn't load automatically when you start Python, so you have to load it with a little Python code.

In [ ]:

frommathimportsqrt

This command imports the square root command (sqrt) from the package called math. Now you can find square roots.

In [ ]:

sqrt(1000)

There are a few different ways to import commands from packages. The above syntax is a good starting point, but sometimes problems can arise if different packages have functions with the same name. Here are a few methods of importing the sqrt function and how they differ.

from math import sqrt: After this command, the sqrt command will refer to the function from the math package (overriding any previous definition).

import math: After this command, all the functions from the math package will be imported. But to call sqrt, you would type a command like math.sqrt(1000). This is convenient if there are potential conflicts with other packages.

from math import *: After this command, all the functions from the math package will be imported. To call them, you can access them directly with a command like sqrt(1000). This can easily cause conflicts with other packages, since packages can have hundreds of functions in them!

import math as mth: Some people like abbreviations. This imports all the functions from the math package. To call one, you type a command like mth.sqrt(1000).

In [ ]:

importmath

In [ ]:

math.sqrt(1000)

In [ ]:

factorial(10)# This will cause an error!

In [ ]:

math.factorial(10)# This is ok, since the math package comes with a function called factorial.

Now let's improve our is_prime(n) function by searching for factors only up to the square root of the number n. We consider two options.

In [ ]:

defis_prime_slow(n):''' Checks whether the argument n is a prime number. Uses a brute force search for factors between 1 and n. '''j=2whilej<=sqrt(n):# j will proceed through the list of numbers 2,3,... up to sqrt(n).ifn%j == 0: # is n divisible by j?
print"%d is a factor of %d."%(j,n)returnFalsej=j+1# There's a Python abbreviation for this: j += 1.returnTrue

In [ ]:

defis_prime_fast(n):''' Checks whether the argument n is a prime number. Uses a brute force search for factors between 1 and n. '''j=2root_n=sqrt(n)whilej<=root_n:# j will proceed through the list of numbers 2,3,... up to sqrt(n).ifn%j == 0: # is n divisible by j?
print"%d is a factor of %d."%(j,n)returnFalsej=j+1# There's a Python abbreviation for this: j += 1.returnTrue

In [ ]:

is_prime_fast(1000003)

In [ ]:

is_prime_slow(1000003)

I've chosen function names with "fast" and "slow" in them. But what makes them faster or slower? Are they faster than the original? And how can we tell?

Python comes with a great set of tools for these questions. The simplest (for the user) are the time utilities. By placing the magic%timeit before a command, Python does something like the following:

Python makes a little container in your computer devoted to the computations, to avoid interference from other running programs if possible.

Python executes the command lots and lots of times.

Python averages the amoung of time taken for each execution.

Give it a try below, to compare the speed of the functions is_prime (the original) with the new is_prime_fast and is_prime_slow. Note that the %timeit commands might take a little while.

In [ ]:

%timeit is_prime_fast(1000003)

In [ ]:

%timeit is_prime_slow(1000003)

In [ ]:

%timeit is_prime(1000003)

Time is measured in seconds, milliseconds (1 ms = 1/1000 second), microseconds (1 µs = 1/1,000,000 second), and nanoseconds (1 ns = 1/1,000,000,000 second). So it might appear at first that is_prime is the fastest, or about the same speed. But check the units! the other two approaches are about a thousand times faster! How much faster were they on your computer?

In [ ]:

is_prime_fast(10000000000037)# Don't try this with `is_prime` unless you want to wait for a long time!

Indeed, the is_prime_fast(n) function will go through a loop of length about sqrt(n) when n is prime. But is_prime(n) will go through a loop of length about n. Since sqrt(n) is much less than n, especially when n is large, the is_prime_fast(n) function is much faster.

Between is_prime_fast and is_prime_slow, the difference is that the fast version precomputes the square root sqrt(n) before going through the loop, where the slow version repeats the sqrt(n) every time the loop is repeated. Indeed, writing while j <= sqrt(n): suggests that Python might execute sqrt(n) every time to check. This might lead to Python computing the same square root a million times... unnecessarily!

A basic principle of programming is to avoid repetition. If you have the memory space, just compute once and store the result. It will probably be faster to pull the result out of memory than to compute it again.

Python does tend to be pretty smart, however. It's possible that Python is precomputingsqrt(n) even in the slow loop, just because it's clever enough to tell in advance that the same thing is being computed over and over again. This depends on your Python version and takes place behind the scenes. If you want to figure it out, there's a whole set of tools (for advanced programmers) like the disassembler to figure out what Python is doing.

In [ ]:

is_prime_fast(10**14+37)# This might get a bit of delay.

Now we have a function is_prime_fast(n) that is speedy for numbers n in the trillions! You'll probably start to hit a delay around $10^{15}$ or so, and the delays will become intolerable if you add too many more digits. In a future lesson, we will see a different primality test that will be essentially instant even for numbers around $10^{1000}$!

To check whether a number n is prime, you can first check whether n is even, and then check whether n has any odd factors. Change the is_prime_fast function by implementing this improvement. How much of a speedup did you get?

Use the %timeit tool to study the speed of is_prime_fast for various sizes of n. Using 10-20 data points, make a graph relating the size of n to the time taken by the is_prime_fast function.

Write a function is_square(n) to test whether a given integer n is a perfect square (like 0, 1, 4, 9, 16, etc.). How fast can you make it run? Describe the different approaches you try and which are fastest.

We have already (briefly) encountered the list type in Python. Recall that the range command produced a list. For example, range(10) produces the list [0,1,2,3,4,5,6,7,8,9]. You can also create your own list by a writing out its terms, e.g. L = [4,7,10].

Here we work with lists, and a very Pythonic approach to list manipulation. With practice, this can be a powerful tool to write fast algorithms, exploiting the hard-wired capability of your computer to shift and slice large chunks of data. Our application will be to implement the Sieve of Eratosthenes, producing a long list of prime numbers (without using any is_prime test along the way).

Notice that the entries in a list can be of any type. The above list L has some integer entries and some string entries. The list B has boolean entries. Lists are ordered in Python, starting at zero. One can access the $n^{th}$ entry in a list with a command like L[n].

In [ ]:

printL[3]

In [ ]:

printL[4]

In [ ]:

printL[0]

The location of an entry is called its index. So at the index 3, the list L stores the entry three. Note that the same entry can occur in many places in a list. E.g. [7,7,7] is a list with 7 at the zeroth, first, and second index.

In [ ]:

printL[-1]printL[-2]

The last bit of code demonstrates a cool Python trick. The "-1st" entry in a list refers to the last entry. The "-2nd entry" refers to the second-to-last entry, and so on. It gives a convenient way to access both sides of the list, even if you don't know how long it is.

Slicing lists allows us to create new lists from old lists, by chopping off one end or the other, or even slicing out entries at a fixed interval. The simplest syntax has the form L[a:b] where a denotes the index of the starting entry and index of the final entry is one less than b. It is best to try a few examples to get a feel for it.

Slicing a list with a command like L[a:b] doesn't actually change the original list L. It just extracts some terms from the list and outputs those terms. Soon enough, we will change the list L using a list assignment.

In [ ]:

L[0:5]

In [ ]:

L[5:11]

In [ ]:

L[3:7]

This continues the strange (for beginners) Python convention of starting at the first number and ending just before the last number. Compare to range(3,7), for example.

The command L[0:5] can be replaced by L[:5] to abbreviate. The empty opening index tells Python to start at the beginning. Similarly, the command L[5:11] can be replaced by L[5:]. The empty closing index tells Python to end the slice and the end. This is helpful if one doesn't know where the list ends.

In [ ]:

L[:5]

In [ ]:

L[3:]

Just like the range command, list slicing can take an optional third argument to give a step size. To understand this, try the command below.

In [ ]:

L[2:10]

In [ ]:

L[2:10:3]

If, in this three-argument syntax, the first or second argument is absent, then the slice starts at the beginning of the list or ends at the end of the list accordingly.

In [ ]:

L# Just a reminder. We haven't modified the original list!

In [ ]:

L[:9:3]# Start at zero, go up to (but not including) 9, by steps of 3.

In [ ]:

L[2::3]# Start at two, go up through the end of the list, by steps of 3.

In [ ]:

L[::3]# Start at zero, go up through the end of the list, by steps of 3.

The sieve of Eratosthenes (hereafter called "the Sieve") is a very fast way of producing long lists of primes, without doing repeated primality checking. It is described in more detail in Chapter 2 of An Illustrated Theory of Numbers. The basic idea is to start with all of the natural numbers, and successively filter out, or sieve, the multiples of 2, then the multiples of 3, then the multiples of 5, etc., until only primes are left.

Using list slicing, we can carry out this sieving process efficiently. And with a few more tricks we encounter here, we can carry out the Sieve very efficiently.

The first approach we introduce is a bit naive, but is a good starting place. We will begin with a list of numbers up to 100, and sieve out the appropriate multiples of 2,3,5,7.

In [ ]:

primes=range(100)# Let's start with the numbers 0...99.

Now, to "filter", i.e., to say that a number is not prime, let's just change the number to the value None.

In [ ]:

primes[0]=None# Zero is not prime.primes[1]=None# One is not prime.printprimes# What have we done?

Now let's filter out the multiples of 2, starting at 4. This is the slice primes[4::2]

In [ ]:

primes[4::2]=[None]*len(primes[4::2])# The right side is a list of Nones, of the necessary length.printprimes# What have we done?

Now we filter out the multiples of 3, starting at 9.

In [ ]:

primes[9::3]=[None]*len(primes[9::3])# The right side is a list of Nones, of the necessary length.printprimes# What have we done?

Next the multiples of 5, starting at 25 (the first multiple of 5 greater than 5 that's left!)

In [ ]:

primes[25::5]=[None]*len(primes[25::5])# The right side is a list of Nones, of the necessary length.printprimes# What have we done?

Finally, the multiples of 7, starting at 49 (the first multiple of 7 greater than 7 that's left!)

In [ ]:

primes[49::7]=[None]*len(primes[49::7])# The right side is a list of Nones, of the necessary length.printprimes# What have we done?

What's left? A lot of Nones and the prime numbers up to 100. We have successfully sieved out all the nonprime numbers in the list, using just four sieving steps (and setting 0 and 1 to None manually).

But there's a lot of room for improvement, from beginning to end!

The format of the end result is not so nice.

We had to sieve each step manually. It would be much better to have a function prime_list(n) which would output a list of primes up to n without so much supervision.

The memory usage will be large, if we need to store all the numbers up to a large n at the beginning.

We solve these problems in the following way.

We will use a list of booleans rather than a list of numbers. The ending list will have a True value at prime indices and a False value at composite indices. This reduces the memory usage and increases the speed.

A which function (explained soon) will make the desired list of primes after everything else is done.

We will proceed through the sieving steps algorithmically rather than entering each step manually.

Here is a somewhat efficient implementation of the Sieve in Python.

In [ ]:

defisprime_list(n):''' Return a list of length n+1 with Trues at prime indices and Falses at composite indices. '''flags=[True]*(n+1)# A list [True, True, True,...] to start.flags[0]=False# Zero is not prime. So its flag is set to False.flags[1]=False# One is not prime. So its flag is set to False.p=2# The first prime is 2. And we start sieving by multiples of 2.whilep<=sqrt(n):# We only need to sieve by p is p <= sqrt(n).ifflags[p]:# We sieve the multiples of p if flags[p]=True.flags[p*p::p]=[False]*len(flags[p*p::p])# Sieves out multiples of p, starting at p*p.p=p+1# Try the next value of p.returnflags

In [ ]:

printisprime_list(100)

If you look carefully at the list of booleans, you will notice a True value at the 2nd index, the 3rd index, the 5th index, the 7th index, etc.. The indices where the values are True are precisely the prime indices. Since booleans take the smallest amount of memory of any data type (one bit of memory per boolean), your computer can carry out the isprime_list(n) function even when n is very large.

To be more precise, there are 8 bits in a byte. There are 1024 bytes (about 1000) in a kilobyte. There are 1024 kilobytes in a megabyte. There are 1024 megabytes in a gigabyte. Therefore, a gigabyte of memory is enough to store about 8 billion bits. That's enough to store the result of isprime_list(n) when n is about 8 billion. Not bad! And your computer probably has 4 or 8 or 12 or 16 gigabytes of memory to use.

To transform the list of booleans into a list of prime numbers, we create a function called where. This function uses another Python technique called list comprehension. We discuss this technique later in this lesson, so just use the where function as a tool for now, or read about list comprehension if you're curious.

In [ ]:

defwhere(L):''' Take a list of booleans as input and outputs the list of indices where True occurs. '''return[nforninrange(len(L))ifL[n]]

Combined with the isprime_list function, we can produce long lists of primes.

In [ ]:

printwhere(isprime_list(100))

Let's push it a bit further. How many primes are there between 1 and 1 million? We can figure this out in three steps:

Create the isprime_list.

Use where to get the list of primes.

Find the length of the list of primes.

But it's better to do it in two steps.

Create the isprime_list.

Sum the list! (Note that True is 1, for the purpose of summation!)

In [ ]:

sum(isprime_list(1000000))# The number of primes up to a million!

In [ ]:

%timeit isprime_list(10**6)

In [ ]:

%timeit sum(isprime_list(10**6))

This isn't too bad! It takes a fraction of a second to identify the primes up to a million, and a smaller fraction of a second to count them! But we can do a little better.

The first improvement is to take care of the even numbers first. If we count carefully, then the sequence 4,6,8,...,n (ending at n-1 if n is odd) has the floor of (n-2)/2 terms. Thus the line flags[4::2] = [False] * ((n-2)/2) will set all the flags to False in the sequence 4,6,8,10,... From there, we can begin sieving by odd primes starting with 3.

The next improvement is that, since we've already sieved out all the even numbers (except 2), we don't have to sieve out again by even multiples. So when sieving by multiples of 3, we don't have to sieve out 9,12,15,18,21,etc.. We can just sieve out 9,15,21,etc.. When p is an odd prime, this can be taken care of with the code flags[p*p::2*p] = [False] * len(flags[p*p::2*p]).

In [ ]:

defisprime_list(n):''' Return a list of length n+1 with Trues at prime indices and Falses at composite indices. '''flags=[True]*(n+1)# A list [True, True, True,...] to start.flags[0]=False# Zero is not prime. So its flag is set to False.flags[1]=False# One is not prime. So its flag is set to False.flags[4::2]=[False]*((n-2)/2)p=3whilep<=sqrt(n):# We only need to sieve by p is p <= sqrt(n).ifflags[p]:# We sieve the multiples of p if flags[p]=True.flags[p*p::2*p]=[False]*len(flags[p*p::2*p])# Sieves out multiples of p, starting at p*p.p=p+2# Try the next value of p. Note that we can proceed only through odd p!returnflags

In [ ]:

%timeit sum(isprime_list(10**6)) # How much did this speed it up?

Another modest improvement is the following. In the code above, the program counts the terms in sequences like 9,15,21,27,..., in order to set them to False. This is accomplished with the length command len(flags[p*p::2*p]). But that length computation is a bit too intensive. A bit of algebraic work shows that the length is given formulaically in terms of p and n by the formula:

$$len = \lfloor \frac{n - p^2 - 1}{2p} \rfloor + 1$$

(Here $\lfloor x \rfloor$ denotes the floor function, i.e., the result of rounding down.) Putting this into the code yields the following.

In [ ]:

defisprime_list(n):''' Return a list of length n+1 with Trues at prime indices and Falses at composite indices. '''flags=[True]*(n+1)# A list [True, True, True,...] to start.flags[0]=False# Zero is not prime. So its flag is set to False.flags[1]=False# One is not prime. So its flag is set to False.flags[4::2]=[False]*((n-2)/2)p=3whilep<=sqrt(n):# We only need to sieve by p is p <= sqrt(n).ifflags[p]:# We sieve the multiples of p if flags[p]=True.flags[p*p::2*p]=[False]*((n-p*p-1)/(2*p)+1)# Sieves out multiples of p, starting at p*p.p=p+2# Try the next value of p.returnflags

In [ ]:

%timeit sum(isprime_list(10**6)) # How much did this speed it up?

That should be pretty fast! It should be under 100 ms (one tenth of one second!) to determine the primes up to a million, and on a newer computer it should be under 50ms. We have gotten pretty close to the fastest algorithms that you can find in Python, without using external packages (like SAGE or sympy). See the related discussion on StackOverflow... the code in this lesson was influenced by the code presented there.

A natural number $n$ is called squarefree if it has no perfect square divides $n$ except for 1. Write a function squarefree_list(n) which outputs a list of booleans: True if the index is squarefree and False if the index is not squarefree. For example, if you execute squarefree_list(12), the output should be [False, True, True, True, False, True, True, True, False, False, True, True, False]. Note that the False entries are located the indices 0, 4, 8, 9, 12. These natural numbers have perfect square divisors besides 1.

Your DNA contains about 3 billion base pairs. Each "base pair" can be thought of as a letter, A, T, G, or C. How many bits would be required to store a single base pair? In other words, how might you convert a sequence of booleans into a letter A,T,G, or C? Given this, how many megabytes or gigabytes are required to store your DNA? How many people's DNA would fit on a thumb-drive?

Now that we can produce a list of prime numbers quickly, we can do some data analysis: some experimental number theory to look for trends or patterns in the sequence of prime numbers. Since Euclid (about 300 BCE), we have known that there are infinitely many prime numbers. But how are they distributed? What proportion of numbers are prime, and how does this proportion change over different ranges? As theoretical questions, these belong the the field of analytic number theory. But it is hard to know what to prove without doing a bit of experimentation. And so, at least since Gauss (read Tschinkel's article about Gauss's tables) started examining his extensive tables of prime numbers, mathematicians have been carrying out experimental number theory.

To carry out serious analysis, we will use the method of list comprehension to place our population into "bins" for statistical analysis. Our first type of list comprehension has the form [x for x in LIST if CONDITION]. This produces the list of all elements of LIST satisfying CONDITION. It is similar to list slicing, except we pull out terms from the list according to whether a condition is true or false.

For example, let's divide the (odd) primes into two classes. Red primes will be those of the form 4n+1. Blue primes will be those of the form 4n+3. In other words, a prime p is red if p%4 == 1 and blue if p%4 == 3. And the prime 2 is neither red nor blue.

print"There are %d red primes and %d blue primes, up to 1 million."%(len(redprimes),len(blueprimes))

This is pretty close! It seems like prime numbers are about evenly distributed between red and blue. Their remainder after division by 4 is about as likely to be 1 as it is to be 3. In fact, it is proven that asymptotically the ratio between the number of red primes and the number of blue primes approaches 1. However, Chebyshev noticed a persistent slight bias towards blue primes along the way.

Some of the deepest conjectures in mathematics relate to the prime counting function $\pi(x)$. Here $\pi(x)$ is the number of primes between 1 and $x$ (inclusive). So $\pi(2) = 1$ and $\pi(3) = 2$ and $\pi(4) = 2$ and $\pi(5) = 3$. One can compute a value of $\pi(x)$ pretty easily using a list comprehension.

In [ ]:

defprimes_upto(x):returnlen([pforpinprimesifp<=x])# List comprehension recovers the primes up to x.

In [ ]:

primes_upto(1000)# There are 168 primes between 1 and 1000.

Now we graph the prime counting function. To do this, we use a list comprehension, and the visualization library called matplotlib. For graphing a function, the basic idea is to create a list of x-values, a list of corresponding y-values (so the lists have to be the same length!), and then we feed the two lists into matplotlib to make the graph.

We begin by loading the necessary packages.

In [ ]:

importmatplotlib# A powerful graphics package.importnumpy# A math packageimportmatplotlib.pyplotasplt# A plotting subpackage in matplotlib.

Now let's graph the function $y = x^2$ over the domain $-2 \leq x \leq 2$ for practice. As a first step, we use numpy's linspace function to create an evenly spaced set of 11 x-values between -2 and 2.

In [ ]:

x_values=numpy.linspace(-2,2,11)# The argument 11 is the *number* of terms, not the step size!printx_valuestype(x_values)

You might notice that the format looks a bit different from a list. Indeed, if you check type(x_values), it's not a list but something else called a numpy array. Numpy is a package that excels with computations on large arrays of data. On the surface, it's not so different from a list. The numpy.linspace command is a convenient way of producing an evenly spaced list of inputs.

The big difference is that operations on numpy arrays are interpreted differently than operations on ordinary Python lists. Try the two commands for comparison.

In [ ]:

[1,2,3]+[1,2,3]

In [ ]:

x_values+x_values

In [ ]:

y_values=x_values*x_values# How is multiplication interpreted on numpy arrays?printy_values

The first line contains the magic%matplotlib inline. We have seen a magic word before, in %timeit. Magic words can call another program to assist. So here, the magic %matplotlib inline calls matplotlib for help, and places the resulting figure within the notebook.

The next line plt.plot(x_values, y_values) creates a plot object based on the data of the x-values and y-values. It is an abstract sort of object, behind the scenes, in a format that matplotlib understands. The following lines set the title of the plot, the axis labels, and turns a grid on. The last line plt.show renders the plot as an image in your notebook. There's an infinite variety of graphs that matplotlib can produce -- see the gallery for more! Other graphics packages include bokeh and seaborn, which extends matplotlib.

Now, to analyze the prime counting function, let's graph it. To make a graph, we will first need a list of many values of x and many corresponding values of $\pi(x)$. We do this with two commands. The first might take a minute to compute.

We created an array of x-values as before. But the creation of an array of y-values (here, called pix_values to stand for $\pi(x)$) probably looks strange. We have done two new things!

We have used a list comprehension [primes_upto(x) for x in x_values] to create a list of y-values.

We have used numpy.array(LIST) syntax to convert a Python list into a numpy array.

First, we explain the list comprehension. Instead of pulling out values of a list according to a condition, with [x for x in LIST if CONDITION], we have created a new list based on performing a function each element of a list. The syntax, used above, is [FUNCTION(x) for x in LIST]. These two methods of list comprehension can be combined, in fact. The most general syntax for list comprehension is [FUNCTION(x) for x in LIST if CONDITION].

Second, a list comprehension can be carried out on a numpy array, but the result is a plain Python list. It will be better to have a numpy array instead for what follows, so we use the numpy.array() function to convert the list into a numpy array.

In [ ]:

type(numpy.array([1,2,3]))# For example.

Now we have two numpy arrays: the array of x-values and the array of y-values. We can make a plot with matplotlib.

In [ ]:

len(x_values)==len(pix_values)# These better be the same, or else matplotlib will be unhappy.

In this range, the prime counting function might look nearly linear. But if you look closely, there's a subtle downward bend. This is more pronounced in smaller ranges. For example, let's look at the first 10 x-values and y-values only.

It still looks almost linear, but there's a visible downward bend here. How can we see this bend more clearly? If the graph were linear, its equation would have the form $\pi(x) = mx$ for some fixed slope $m$ (since the graph does pass through the origin). Therefore, the quantity $\pi(x)/x$ would be constant if the graph were linear.

Hence, if we graph $\pi(x) / x$ on the y-axis and $x$ on the x-axis, and the result is nonconstant, then the function $\pi(x)$ is nonlinear.

In [ ]:

m_values=pix_values[1:]/x_values[1:]# We start at 1, to avoid a division by zero error.

That is certainly not constant! The decay of $\pi(x) / x$ is not so different from $1 / \log(x)$, in fact. To see this, let's overlay the graphs. We use the numpy.log function, which computes the natural logarithm of its input (and allows an entire array as input).

In [ ]:

%matplotlib inline
plt.plot(x_values[1:],m_values,label='$\pi(x)/x$')# The same as the plot above.plt.plot(x_values[1:],1/numpy.log(x_values[1:]),label='$1 / \log(x)$')# Overlay the graph of 1 / log(x)plt.title('The ratio of $\pi(x) / x$ as $x$ varies.')plt.xlabel('x')plt.ylabel('$\pi(x) / x$')plt.grid(True)plt.legend()# Turn on the legend.plt.show()

The shape of the decay of $\pi(x) / x$ is very close to $1 / \log(x)$, but it looks like there is an offset. In fact, there is, and it is pretty close to $1 / \log(x)^2$. And that is close, but again there's another little offset, this time proportional to $2 / \log(x)^3$. This goes on forever, if one wishes to approximate $\pi(x) / x$ by an "asymptotic expansion" (not a good idea, it turns out).

Comparing the graph to the theoretical result, we find that the ratio $\pi(x) / (x / \log(x))$ approaches $1$ (the theoretical result) but very slowly (see the graph above!).

A much stronger result relates $\pi(x)$ to the "logarithmic integral" $li(x)$. The Riemann hypothesis is equivalent to the statement
$$\left\vert \pi(x) - li(x) \right\vert = O(\sqrt{x} \log(x)).$$
In other words, the error if one approximates $\pi(x)$ by $li(x)$ is bounded by a constant times $\sqrt{x} \log(x)$. The logarithmic integral function isn't part of Python or numpy, but it is in the mpmath package. If you have this package installed, then you can try the following.

In [ ]:

frommpmathimportli

In [ ]:

printprimes_upto(1000000)# The number of primes up to 1 million.printli(1000000)# The logarithmic integral of 1 million.

As a last bit of data analysis, we consider the prime gaps. These are the numbers that occur as differences between consecutive primes. Since all primes except 2 are odd, all prime gaps are even except for the 1-unit gap between 2 and 3. There are many unsolved problems about prime gaps; the most famous might be that a gap of 2 occurs infinitely often (as in the gaps between 3,5 and between 11,13 and between 41,43, etc.).

Once we have our data set of prime numbers, it is not hard to create a data set of prime gaps. Recall that primes is our list of prime numbers up to 1 million.

In [ ]:

len(primes)# The number of primes up to 1 million.

In [ ]:

primes_allbutlast=primes[:-1]# This excludes the last prime in the list.primes_allbutfirst=primes[1:]# This excludes the first (i.e., with index 0) prime in the list.

What have we done? It is useful to try out this method on a short list.

In [ ]:

L=[1,3,7,20]# A nice short list.

In [ ]:

printL[:-1]printL[1:]

Now we have two lists of the same length. The gaps in the original list L are the differences between terms of the same index in the two new lists. One might be tempted to just subtract, e.g., with the command L[1:] - L[:-1], but subtraction is not defined for lists.

Fortunately, by converting the lists to numpy arrays, we can use numpy's term-by-term subtraction operation.

In [ ]:

L[1:]-L[:-1]# This will give a TypeError. You can't subtract lists!

In [ ]:

numpy.array(L[1:])-numpy.array(L[:-1])# That's better. See the gaps in the list [1,3,7,20] in the output.

Now let's return to our primegaps data set. It contains all the gap-sizes for primes up to 1 million.

In [ ]:

printlen(primes)printlen(primegaps)# This should be one less than the number of primes.

As a last example of data visualization, we use matplotlib to produce a histogram of the prime gaps.

In [ ]:

max(primegaps)# The largest prime gap that appears!

In [ ]:

%matplotlib inline
plt.figure(figsize=(12,5))# Makes the resulting figure 12in by 5in.plt.hist(primegaps,bins=range(1,115))# Makes a histogram with one bin for each possible gap from 1 to 114.plt.ylabel('Frequency')plt.xlabel('Gap size')plt.grid(True)plt.title('The frequency of prime gaps, for primes up to 1 million')plt.show()

Observe that gaps of 2 (twin primes) are pretty frequent. There are over 8000 of them, and about the same number of 4-unit gaps! But gaps of 6 are most frequent in the population, and there are some interesting peaks at 6, 12, 18, 24, 30. What else do you observe?

Create functions redprimes_upto(x) and blueprimes_upto(x) which count the number of red/blue primes up to a given number x. Recall that we defined red/blue primes to be those of the form 4n+1 or 4n+3, respectively. Graph the relative proportion of red/blue primes as x varies from 1 to 1 million. E.g., are the proportions 50%/50% or 70%/30%, and how do these proportions change? Note: this is also visualized in An Illustrated Theory of Numbers and you can read an article by Rubinstein and Sarnak for more.

Does there seem to be a bias in the last digits of primes? Note that, except for 2 and 5, every prime ends in 1,3,7, or 9. Note: the last digit of a number n is obtained from n % 10.

Read about the "Prime Conspiracy", recently discovered by Lemke Oliver and Soundararajan. Can you detect their conspiracy in our data set of primes?

This website does not host notebooks, it only renders notebooks
available on other websites.