vEnhance's avatar

May 29, 2015

🖉 Zeros and Primes

Prerequisites for this post: previous post, and complex analysis. For this entire post, ss is a complex variable with s=σ+its = \sigma + it.

1. The Γ\Gamma function

So there’s this thing called the Gamma function. Denoted Γ(s)\Gamma(s), it is defined by Γ(s)=0xs1exdx\Gamma(s) = \int_0^{\infty} x^{s-1} e^{-x} dx as long as σ>0\sigma > 0. Here are its values at the first few integers:

Γ(1)=1Γ(2)=1Γ(3)=2Γ(4)=6Γ(5)=24. \begin{aligned} \Gamma(1) &= 1 \\ \Gamma(2) &= 1 \\ \Gamma(3) &= 2 \\ \Gamma(4) &= 6 \\ \Gamma(5) &= 24. \end{aligned}

Yep: the Γ\Gamma function is contrived so that the identity Γ(s+1)=sΓ(s)\Gamma(s+1) = s\Gamma(s) always holds. (Proof: integration by parts.) Thus Γ\Gamma interpolates the factorial function a nice way. Moreover, this identity lets us extend Γ\Gamma to a meromorphic function on the entire complex plane.

We like the Γ\Gamma function because, unlike ζ\zeta, we know what it’s doing. For example we actually know how to compute Γ(n)\Gamma(n) for any positive integer nn. (Contrast this with, say, ζ(3)\zeta(3); we don’t even know if it’s an algebraic number, and it took until 1978 to prove that it was irrational). More to the point, we also know where all its zeros and poles are:

Proposition 1. Let Γ\Gamma be defined as above, extended to a meromorphic function on all of C\mathbb C.

  • The function Γ\Gamma has no zeros.
  • It has a simple pole at s=0,1,2,s= 0, -1, -2, \dots, and these are the only poles.

The pole at s=0s=0 should not be surprising: plug in s=0s=0 to Γ(s+1)=sΓ(s)\Gamma(s+1)=s\Gamma(s).

In any case, moral of story: Γ\Gamma is very friendly!

2. Functional Equation for Zeta

We will now do something really magical, due to Riemann. Pick an integer nn; we set t=πn2xt = \pi n^2x to derive the artificial equation Γ(s)=(n2π)s0xs1en2πxdx.\Gamma(s) = (n^2\pi)^{s} \int_0^{\infty} x^{s-1} e^{-n^2\pi x} dx. Replacing ss with 12s\frac{1}{2} s and rearranging gives πs/2Γ(s/2)ns=0xs/21en2πxdxσ>0.\pi^{-s/2} \cdot \Gamma(s/2) n^{-s} = \int_0^\infty x^{s/2 - 1} e^{-n^2\pi x} dx \qquad \sigma > 0. Then, due to absolute convergence we can sum over n1n \ge 1; this brings the Riemann zeta function into the right-hand side, to get

πs/2Γ(s/2)ζ(s)=0xs/21θ(x)12dxσ>1 \pi^{-s/2} \cdot \Gamma(s/2) \zeta(s) = \int_0^\infty x^{s/2 - 1} \frac{\theta(x)-1}{2} dx \qquad \sigma > 1

where θ(x)=n=en2πx;\theta(x) = \sum_{n=-\infty}^{\infty} e^{-n^2\pi x}; so that θ(x)12\frac{\theta(x)-1}{2} gives just the sum for n1n \ge 1. It turns out that this θ\theta is special: a Jacobi theta function, which happens to satisfy θ(x1)=xθ(x)x>0.\theta(x^{-1}) = \sqrt{x} \cdot \theta(x) \quad \forall x > 0. Also, ω(x)=O(eπx)\omega(x) = O(e^{-\pi x}) for x>0x > 0.

Using this and with some purely algebraic manipulations (namely, splitting the integral into two parts, 0x10 \le x \le 1 and x1x \ge 1, and then using the property of the theta function), one can derive that

πs/2Γ(s/2)ζ(s)=1s11s+1θ(x)12(xs/21+x(1s)/21)dx. \pi^{-s/2}\Gamma(s/2)\zeta(s) = \frac1{s-1} - \frac 1s + \int_1^\infty \frac{\theta(x)-1}{2} \cdot \left( x^{s/2-1} + x^{(1-s)/2-1} \right) dx.

The right-hand side has two very useful properties:

  • It is even around 12\frac{1}{2}, meaning it remains the same when ss is replaced by 1s1-s.
  • The integral is actually an entire function on all of C\mathbb C (the integral converges for all ss because θ(x)=O(eπx)\theta(x) = O(e^{-\pi x})).

So if we multiply both sides by 12s(s1)\frac{1}{2} s(s-1), we get a symmetric entire function, called the ξ\xi function:

Theorem 2. The function ξ(s)=12s(s1)πs/2Γ(s/2)ζ(s)\xi(s) = \frac{1}{2} s(s-1) \pi^{-s/2} \Gamma(s/2) \zeta(s) satisfies ξ(s)=ξ(1s)\xi(s) = \xi(1-s) and is an entire function on all of C\mathbb C.

In particular we can use this to extend ζ\zeta to a meromorphic function on the entire complex plane. We have

π12sΓ(s2)ζ(s)=π12(1s)Γ(1s2)ζ(1s). \pi^{-\frac{1}{2} s} \Gamma\left( \frac s2 \right) \zeta(s) = \pi^{-\frac{1}{2} (1-s)} \Gamma\left( \frac{1-s}{2} \right) \zeta(1-s).

We can count zeros and poles from this. The π\pi’s don’t do anything. The Γ\Gamma’s have no zeros and just poles at non-positive integers.

Since ζ\zeta has no zeros and no poles other than s=1s = 1 for σ>0\sigma > 0, the only zeros it will have are for σ0\sigma \le 0 are at s=2,4,6,s = -2, -4, -6, \dots in order to cancel out the poles of Γ(s/2)\Gamma(s/2). And so these are the so-called trivial zeros.

On the other hand in the strip 0<σ<10 < \sigma < 1 we get very confused. More on that later.

3. Explicit Formula for Chebyshev Function

Recall that last post we had ψ(x)=12πicic+iζ(s)ζ(s)xssds\psi(x) = \frac{1}{2\pi i} \int_{c-i\infty}^{c+i\infty} -\frac{\zeta'(s)}{\zeta(s)} \cdot \frac{x^s}{s} ds according to Perron’s Formula. Write this as

ψ(x)=12πilimT(ciTc+iTζ(s)ζ(s)xssds) \psi(x) = \frac{1}{2\pi i} \lim_{T \rightarrow \infty} \left( \int_{c-iT}^{c+iT} \frac{\zeta'(s)}{\zeta(s)} \cdot -\frac{x^s}{s} ds \right)

Recall that if ff is a meromorphic function then so is 12πiff\frac{1}{2\pi i} \cdot \frac{f'}{f} which has a simple pole for each zero/pole of ff; the residue is +1+1 for each simple zero (more generally +m+m for a zero of multiplicity mm) and 1-1 for each simple pole (more generally m-m for a pole of order mm). (See the argument principle).

Now we are going to do use the bestest theorem from complex analysis: Cauchy’s Residue Theorem. We replace the path ciTc+iTc-iT \rightarrow c+iT with ciTUiTU+iTc+iTc-iT \rightarrow -U-iT \rightarrow -U + iT \rightarrow c+iT. Doing so picks up the following residues:

  • Since ζ\zeta has a pole at s=1s=1, ζ/ζ\zeta'/\zeta has a simple pole there of residue 1-1, so we pick up a residue of +x+x corresponding to

    (1)x1/1=+x.(-1) \cdot -x^1/1 = +x. This is the “main term”.

  • ζ\zeta has simple zeros at s=2,4,6,s=-2,-4,-6,\dots. So we pick up residues of x22-\frac{x^2}{2}, x44-\frac{x^4}{4}, and so on, or

    n1x2n2n=12log(1x2).- \sum_{n \ge 1} \frac{x^{2n}}{2n} = -\frac{1}{2} \log(1-x^{-2}).

  • xs/sx^s/s itself has a pole at s=0s=0, which gives us an additional term

    ζ(0)ζ(0).-\frac{\zeta'(0)}{\zeta(0)}. It turns out that this equals log(2π)\log(2\pi), because why not.

  • Finally, the hard-to-understand zeros in the strip 0<σ<10 < \sigma < 1. If ρ=β+iγ\rho = \beta+i\gamma is a zero, then it contributes a residue of xρρ-\frac{x^\rho}{\rho}. We only pick up the zeros with γ<T\left\lvert \gamma \right\rvert < T in our rectangle, so we get a term

    ρ,γ<Txρρ\sum_{\substack{\rho, \left\lvert \gamma \right\rvert < T}} \frac{x^\rho}{\rho} In what follows, ρ\rho always denotes a zero in the critical strip, written ρ=β+iγ\rho = \beta + i \gamma.

Putting this all together, we obtain that

ψ(x)=12πilimUlimT(x12log(1x2)log(2π)ρ,γ<Txρρ+I(x,U,T)) \psi(x) = \frac{1}{2\pi i} \lim_{U \rightarrow \infty} \lim_{T \rightarrow \infty} \left( x - \frac{1}{2} \log(1-x^{-2}) - \log(2\pi) - \sum_{\substack{\rho, \left\lvert \gamma \right\rvert < T}} \frac{x^\rho}{\rho} + I(x,U,T) \right)

where I(x,U,T)I(x,U,T) is the integral along the three sides of the rectangle. With some effort, you can show that in fact I(x,U,T)0I(x,U,T) \rightarrow 0; but the speed of convergence depends on xx. To avoid this, we can take UU \rightarrow \infty but leave TT intact, which gives the explicit formula:

Theorem 3 (Explicit Formula for the Chebyshev Function)

For x2x \ge 2

ψ(x)=xρ,γ<Txρρlog(2π)12log(1x2)+R(x,T) \psi(x) = x - \sum_{\rho, \left\lvert \gamma \right\rvert < T} \frac{x^\rho}{\rho} - \log(2\pi) - \frac{1}{2}\log(1-x^{-2}) + R(x,T)

where R(x,T)R(x,T) is some error term, with the property that R(x,T)0R(x,T) \rightarrow 0 as TT \rightarrow \infty for any fixed xx.

(It’s necessary to use a truncated sum since the series ρ1ρ\sum_{\rho} \frac{1}{\left\lvert \rho \right\rvert} actually converges only conditionally, and not absolutely.)

The error term is ugly enough that I haven’t included it in the formula, but if you want to know I’ll at least write it down: it is

R(x,T)=O(xlog(xT)2T+logxmin{1,xT<x>}) R(x,T) = O\left( \frac{x \log(xT)^2}{T} + \log x \min \left\{ 1, \frac{x}{T\left<x\right>} \right\} \right)

where <x>\left<x\right> is the distance from xx to the nearest prime power (x\ne x). Clearly, for any fixed xx, R(x,T)0R(x,T) \rightarrow 0 as TT \rightarrow \infty. What I will say is that it’s not our main concern:

4. The Prime Number Theorem

Here’s the deal: we know ψ(x)x\psi(x) \approx x, and want something as close as possible. We get the xx right off the bat, so we want everything else to be small.

The term 12log(1x2)-\frac{1}{2}\log(1-x^{-2}) is tiny, as is the constant. The R(x,T)R(x,T) can be handled as long as TT is big enough: the price we pay is that we introduce more zeros into the sum over ρ\rho. The sum of the zeros: well, what about the sum?

We know that for any zero ρ=β+iγ\rho = \beta + i \gamma, we have β<1\beta < 1. But if β=1ε\beta = 1 - \varepsilon, we’d be very upset, because now our sum has a term of size xρ=xβ=x1ε\left\lvert x^\rho \right\rvert = x^\beta = x^{1-\varepsilon} in it, which is bad since the thing we’re shooting for is ψ(x)x\psi(x) \approx x. So what we’d like is to be able to force β\beta to be less than something, like 0.90.9. Then we’d have an error term around x0.9x^{0.9}, which is not spectacular but better than x1εx^{1-\varepsilon}.

In fact, we believe that β=12\beta = \frac12, always – the Riemann Hypothesis. With some messing around with the value of TT, we would then get an error term of O(x12+ε)O\left( x^{\frac12 + \varepsilon} \right) which is pretty good, and actually near best possible (one can show that O(x)O(\sqrt x) is not achievable).

Unfortunately, we can not even show β<0.999\beta < 0.999. The most we know is that β<1clogγ\beta < 1 - \frac{c}{\log \gamma}, which gives some pretty crummy bounds compared to what we think is true: using this bound, the best we can do is ψ(x)=x+O(xexp(clogx))\psi(x) = x + O\left( x \exp\left( -c\sqrt{\log x} \right) \right) which is worse than O(x0.999)O(x^{0.999}). That’s the current state of affairs.

That ζ\zeta function is pretty mysterious.

References: Kedlaya’s 18.785 notes and Hildebrand’s ANT notes.