Theorem 33: the size of Gauss sums

Sorry for the delay in posting this: I’ve been working on it for a few days, but it took a little while to get together.  I hope you find it interesting!

I’d like to send you on some mystery tours.  In each case, I’d like you to think about where you end up (relative to where you started).  You’ll need a very accurate compass and plenty of space, or just a good imagination!  Or perhaps you could send a turtle for a walk, if you have some form of Logo on your computer (alas, I do not).  I’ve drawn some pictures for this post; I used GeoGebra (which I used last time too).  It was easy to use for this, but I couldn’t find a turtle.


In each case, I’m going to give you some bearings, and I’d like you take one step on the first bearing, one step on the second, and so on.  The bearings are all in degrees.

  1. Bearings: 0, 0, 0, 0, 0, 0, 0.
  2. Bearings: 360/7, 360/7, 360/7, 360/7, 360/7, 360/7, 360/7.
  3. Bearings: 0, 360\times (1/7), 360\times (2/7), 360\times (3/7), 360\times (4/7), 360\times (5/7), 360\times (6/7).
  4. Bearings: 0, 360\times (1^2/7), 360\times (2^2/7), 360\times (3^2/7), 360\times (4^2/7), 360\times (5^2/7), 360\times (6^2/7).

Done that?  I encourage you to try the same sort of thing with some other routes.  For example, you could try replacing 7 by other numbers.

What can we say about where we end up, following each of these four routes?  Can we predict this in advance?  Can we predict what would happen for other numbers?

Here are some pictures.  Hopefully they look like yours!  (The angles are measured anticlockwise from the horizontal positive axis.  It now occurs to me that this is not the standard direction if you’re using a compass, but it matches standard mathematical practice better.  So I hope you won’t be too confused and will forgive me.)

  1. Bearings: 0, 0, 0, 0, 0, 0, 0.
  2. Straight line

  3. Bearings: 360/7, 360/7, 360/7, 360/7, 360/7, 360/7, 360/7.
  4. Straight line at angle 360/7

  5. Bearings: 0, 360\times (1/7), 360\times (2/7), 360\times (3/7), 360\times (4/7), 360\times (5/7), 360\times (6/7).
  6. Polygon using angles 0, 360/7, 2*360/7, etc.

  7. Bearings: 0, 360\times (1^2/7), 360\times (2^2/7), 360\times (3^2/7), 360\times (4^2/7), 360\times (5^2/7), 360\times (6^2/7).

Path using bearings 0, 360/7, 4*360/7, etc.


I think that we can hopefully see reasonably easily that in examples 1 and 2, with a constant bearing, we’re going to get straight lines.  So let’s not dwell on that.  What happens with the other two?  Here are the corresponding things with 7 replaced by 13.

Bearings: 0, 360\times (1/13), 360\times (2/13), 360\times (3/13), 360\times (4/13), 360\times (5/13), 360\times (6/13), 360\times (7/13), 360\times (8/13), 360\times (9/13), 360\times (10/13), 360\times (11/13), 360\times (12/13).

Polygon with bearings 0, 360/13, 2*360/13, etc.

Bearings: 0, 360\times (1^2/13), 360\times (2^2/13), 360\times (3^2/13), 360\times (4^2/13), 360\times (5^2/13), 360\times (6^2/13) 360\times (7^2/13), 360\times (8^2/13), 360\times (9^2/13), 360\times (10^2/13), 360\times (11^2/13), 360\times (12^2/13).

Path with bearings 0, 360/13, 4*360/13, etc.

Notice anything interesting?

I’d like to draw your attention to a few points.

  • In example 3, we have a closed regular polygon, starting and finishing at the origin (so we end up exactly where we started).
  • In example 4, we take quite an erratic-looking path, but finish on an axis, and at distance \sqrt{7} (or \sqrt{13}) from the origin.  (I agree it’s not obvious from my pictures above that these are the distances, so try it for yourself if you don’t believe me!)

(There are other interesting things to notice and explain too, such as some symmetry, but I’m not going to talk about them here.)

In order to explore this further, I want to introduce some notation.  I’ll try to keep it as non-technical as possible, but I’ll use the standard notation for the benefit of people who’ve come across it before.

I’m going to write e(\theta) for a single step with `bearing’ (in the above sense) 360 \times \theta.  (As is fairly standard, at least in certain circles, I’m writing e(\theta) for \exp(2\pi i\theta)=\cos(2\pi\theta)+i\sin(2\pi\theta).)

We take a sequence of steps by adding such things.  So example 3 corresponds to e(0) + e(1/7) + e(2/7) + e(3/7) + e(4/7) + e(5/7) + e(6/7).  This is not enormously convenient notation (I don’t really want to type the version with 13 instead of 7!), so let me introduce the sigma sign (if you haven’t met it before).  I’d record the above sum as \sum_{n=0}^6 e(n/7).  The symbol \sum_{n=0}^6 means that we take the sum of the following term as n ranges from 0 to 6, and this is precisely e(0) + e(1/7) + e(2/7) + e(3/7) + e(4/7) + e(5/7) + e(6/7).  In this notation, we’d write example 4 as \sum_{n=0}^6 e(n^2/7).  OK?

It turns out not to be too difficult to see that for any N, we have \sum_{n=0}^{N-1} e(n/N) = 0.  That is, we always end up back where we started in this case (as we predicted based on our experience with N=7 and N=13).  I think I’ll leave this as an exercise for you.

What about the squares?  I want to concentrate on primes here, so I want to look at things like \sum_{n=0}^{p-1} e(n^2/p), where p is an odd prime.  These are called Gauss sums, and that brings me to this week’s theorem.

Theorem Let p be an odd prime.  Then \sum_{n=0}^{p-1} e(n^2/p) = \epsilon_p \sqrt{p}, where \epsilon_p is 1 if p \equiv 1 \mod{4}, and i if p \equiv 3\mod{4}.

What does this mean?  It means that the size of the sum is always \sqrt{p}.  That is, the distance of the finishing point from the starting point is \sqrt{p}, as we suspected.  And it tells us that if p leaves remainder 1 when divided by 4, then the finishing point will be on the positive horizontal axis (as I’ve drawn it), whereas if p leaves remainder 3 when divided by 4, then the finishing point will be on the positive vertical axis.  You might like to use the pictures above to check this for p=7 and p=13.

I think I’m going to skip the proof here, because this post is probably long enough already.  Showing that the size of the sum is \sqrt{p} isn’t too difficult, when one knows a bit about manipulating these exponentials e(\theta).  It takes a bit more work to establish the direction.  There are probably proofs of this all over the place.  Off the top of my head, I happen to know that Davenport’s Multiplicative Number Theory and Iwaniec and Kowalski’s Analytic Number Theory both have sections on Gauss sums.

What’s the point of this?  It tells us something about the distribution of the quadratic residues modulo a prime.  We’ve seen three quite different types of behaviour with these mystery tours.  In one, we ended up as far away from the starting point as possible, when all the bearings were the same.  They were concentrated in one direction, rather than being evenly spread round the circle (compass).  In another, we had the opposite extreme: the bearings were so uniformly distributed round the circle that we ended up where we started.  Those are both quite artificial situations: if I picked seven bearings at random, I wouldn’t expect to get either of those types of behaviour.  I’d expect something in between, and this is the sort of behaviour that we see with the squares.  So this is telling us (being rather vague) that the quadratic residues are somehow quite randomly distributed.  They don’t all bunch up and conspire to line up in some direction.  It’s notoriously difficult to say precise things about the distribution of the quadratic residues (Tao has a nice post about some ideas about this), but this evaluation of the Gauss sum gives some information that turns out to be useful.  I think it’s pretty amazing that we can exactly compute the value of this sum: it takes a wiggly route along the way, but we can say exactly where it will end up.

If only I had a turtle…


5 Responses to “Theorem 33: the size of Gauss sums”

  1. Alan Crowe Says:

    I don’t have Logo, but I do have a Ghostscript, the GNU PostScript interpreter. I reproduced the drunken stagger you show for n=13. Given the square root and your comments on randomness, I was fairly sure what I would see when I cranked n up to 557: a drunkards random walk.

    So my little bit of PostScript

    2 setmiterlimit
    /unit 20 def % about half an inch
    72 400 moveto
    /n 557 def
    1 1 1 n mul
    { dup mul %square it
    dup %use it twice, for both x and y
    360 n div mul cos unit mul % that’s x
    exch % start again for y
    360 n div mul sin unit mul % that’s y
    } for

    gave me quite a surpris, see

    The path spirals in, jiggles in place, and spirals back out + repeat due to symmetry. I’m verifying that the sum comes to the square root of n with a very easy bit of Common Lisp

    (defun gauss-sum (n)
    (loop for r below n
    sum (cis (/ (* 2 pi r r) n))))

    but I’m not about to try proving anything. I thought that summing e to the index squared was the warning sign that theta functions are lurking nearby. Scary!

  2. theoremoftheweek Says:

    Nice picture, thanks Alan! There are some similar nice pictures in a book by Hugh Montgomery ( My picture for p=13 shows signs of that behaviour, but only when you know what you’re looking for. (On the other hand, I put in each line segment by hand, which would have taken a long time for p=557! So I’m grateful to you.)

    The crucial thing about these sums is that it’s n^2/p, and we’re summing over all values of n (mod p). So they’re finite sums (not scary theta functions!).

  3. Olof Says:

    Just a quick note on the argument/direction of the Gauss sums: it is not too hard to see that the sums must land on an axis. Indeed, writing G = \sum_{n=0}^{p-1} e(n^2/p), consider G + \overline{G} = 2 \mathrm{Re}(G). If -1 is a square mod p, then one has that G = \overline{G}, whereas if -1 is not a square mod p then one can argue slightly differently, making use of the linear exponential sums you’ve discussed above.

    Unfortunately, the only demonstrations of which side of the origin the Gauss sums land on that I know of are much harder than this!

  4. Carnival of Mathematics 69 « JD2718 Says:

    […] Neale writes about theorems. Here she discusses Theorem 33: the size of Gauss sums posted at Theorem of the week. Robin Whitty (who does the same thing, but 7 times as often) […]

  5. Analysis I: Lecture 5 « Theorem of the week Says:

    […] directly related to the Analysis I course, but hopefully might be interesting: a blog post about the size of Gauss sums, which links closely with this business of sums with and without cancellation (and which has some […]

Leave a Reply

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out / Change )

Twitter picture

You are commenting using your Twitter account. Log Out / Change )

Facebook photo

You are commenting using your Facebook account. Log Out / Change )

Google+ photo

You are commenting using your Google+ account. Log Out / Change )

Connecting to %s

%d bloggers like this: