Why Yitang Zhang's proof is probably far less fundamental than the claim
Yitang Zhang worked at Subway before he would land a mathematics job. And when he did, he wasn't publishing almost anything for years before he would offer a proof of something rather important weeks ago. That turned the name of the popular math instructor in New Hampshire into one of the most well-known names of number theorists in the world.
Some increasingly popular links are:
\liminf_{n\to\infty} (p_{n+1}-p_n) \lt 70,000,000.
\] The operator above is called the limit inferior which is just\[
\liminf_{n\to\infty}x_n := \lim_{n\to\infty}\Big(\inf_{m\geq n}x_m\Big)
\] If you think about this limit of the infimum for a while, you will understand that the limit inferior in the claim proved by Zhang is just the smallest gap between the adjacent primes that is realized infinitely many times (for infinitely many pairs). In other words, there exists at least one number – a potential gap between adjacent primes – that is realized infinitely many times.
Because of some technical properties that probably depended on many personal choices that Zhang has made while attacking the problem, the upper bound in the inequality turns out to be a high number, namely 70 million. It is such a high number that for all practical purposes, the proposition proven by Zhang is de facto equivalent to\[
\liminf_{n\to\infty} (p_{n+1}-p_n) \lt \infty
\] i.e. to the claim that there exists a finite number that is realized as the gap between adjacent primes in infinitely many pairs.
On the other hand, as I will argue, the actually correct (but rigorously unproven) claim stronger than Zhang's theorem says\[
\liminf_{n\to\infty} (p_{n+1}-p_n) = 2
\] which means that even twin primes – pairs of primes that differ by two – are realized infinitely many times: there are infinitely many pairs of twin primes. This claim is the famous twin prime conjecture. In some sense, the assertion proven by Zhang is 35 million times weaker than the twin prime conjecture. Note that the first twin primes are\[
(3, 5), (5, 7), (11, 13), (17, 19), (29, 31), (41, 43), (59, 61), \\
(71, 73), (101, 103), (107, 109), (137, 139), \dots
\] and there doesn't seem to be the tiniest reason to think that the list should terminate at some point. The largest currently known twin prime pair is \(2,003,663,613\cdot 2^{195,000}\pm 1\), two similar numbers that have 58,711 digits (each).
I don't plan to study the proof in detail because it looks very complicated and "non-unique" to me. The proven statement is slightly interesting but the proof is probably less interesting – there's just a small chance that I am wrong – and there's less "profound message" to learn from it. It's like if you are interested in the Moon and someone asks you to study O-rings in the Apollo spacecraft.
Moreover, I am not too interested in the claim that has been proved. But there is one more key reason: I feel certain that the proposition is true. The reason behind this certainty is the validity of a much stronger claim – a not quite rigorously defined one – that implies the twin prime conjecture, Zhang's proof, and many and many other much weaker corollaries. The claim is that
In this analogy, "patterns that are easy to prove" are analogous to the "symmetries or other principles forbidding certain things". May I explain what I mean in the case of primes?
A pattern that is easy to prove is, for example, that if \(n\) is a prime, then \(n+1\) and \(n-1\) are not primes, assuming that \(n\gt 3\). It's because except for \(n=2\), only odd numbers may be prime. Similarly, among six consecutive integers greater than \(12\), just to be sure, at most two numbers may be primes. It's because only three numbers among the six are odd; and one of them is a multiple of three. One could continue with many examples of this kind.
Similarly, using the Gell-Mann totalitarian principle, one may demonstrate that the twin prime conjecture and its generalizations hold. There doesn't seem to be any reason why the difference between primes shouldn't be equal to two (or some other allowed even numbers) – there are many examples in which it is two, in fact – so there must exist infinitely many examples for the probability that \(n\) and \(n+2\) are both primes to be nonzero. Of course, it's hard to prove that "there is no reason why twin primes should stop at some point", either, but at least, one may prove that there exist no "reasons of the well-known types".
A TRF-based heuristic proof of the prime number theorem
Now, the density of primes around \(n\) asymptotically goes like \(1/ \ln n\). This is the right estimate for \(n\to\infty\), including the right numerical prefactor (the relative error goes to zero in the limit). This statement is known as the prime number theorem and it is a severely weakened sibling of the Riemann Hypothesis which may be equivalently stated as the (easy-to-prove) assertion that the roots of \(\zeta(s)\) only exist for \(s\in\RR\) or in the critical strip \(0\leq {\rm Re}(s)\leq 1\).
I can offer you a supersimple, Lumoesque argument why the density of primes goes like \(1/\ln(n)\). Call the functional dependence of the density \(\rho(n)\); it's really the probability that the number around \(n\) is prime. A number \(n\) is prime if it is not divisible by any prime smaller than or equal to \(\sqrt{n}\). These are statistically independent conditions. So\[
\rho(n) = P_{n\in{\rm primes}} = \prod_{p\leq \sqrt{n}}^{p\in{\rm primes}} (1-1/p)=\dots
\] because the probability that a random large \(n\) isn't a multiple of \(p\) equals \(1-1/p\). But the product may be written as the exponential of the sum of logarithms\[
\dots = \exp\sum_{p\leq \sqrt{n}}^{p\in{\rm primes}} \ln (1-1/p) = \dots
\] and the sum over primes \(p\) may be approximated by the sum over all integers \(i\) weighted by the probability \(\rho(i)\) that \(i\) is prime:\[
\rho(n) = \dots = \exp \sum_{i\leq\sqrt{n}} \rho(i) \ln (1-1/i).
\] Now, the sum over \(i\) may be approximated by an integral when \(\rho(i)\) is smoothened. Take the logarithm of the identity above (with the sum replaced by the integral)\[
\ln\rho(n) = \dots = \int_1^{\sqrt{n}} \dd i\, \rho(i) \ln (1-1/i)
\] and differentiate it with respect to \(n\) to get\[
\frac{\rho'(n)}{\rho(n)} = \frac{1}{2\sqrt{n}} \rho(\sqrt{n}) \ln(1-1/ \sqrt{n})\sim -\frac{\rho(\sqrt{n})}{2n}
\] where \(1/2\sqrt{n}\) came from \(\dd(\sqrt{n})/\dd n\) and where \(\ln(1-x)\sim -x\) because \(x\to 0^+\). One may easily verify that \(\rho(n)\sim 1 / \ln(n)\) satisfies the identity above; both sides are equal to \(-1/ n\ln(n)\) in that case. Among uniformly, nicely decreasing functions \(\rho(n)\), this solution may be seen to be unique. Even the coefficient in front of the logarithm or, equivalently, the base of the logarithm (\(e\)) may be seen to be determined by the (nonlinear) condition above.
You may check how Terence Tao imagines a heuristic proof of the prime number theorem. I leave you to decide who among the two of us is the cumbersome overworked craftsman and who is the seer. ;-)
At any rate, the heuristic proofs above aren't rigorous but one may rigorously prove the prime number theorem. One may also prove other things. As you can see by comparing various proofs sketched by various people – or the same people at various moments – there are many strategies that may be used to attack similar problems. When we're rigorously proving something like that in mathematics, we often work with lots of inequalities – not only the final one that e.g. Zhang has proved; but also with many inequalities in the intermediate steps. And the inequalities are usually ad hoc. We want to find an object that is "good enough to achieve a certain next step" but how good this good enough object has to be isn't quite determined. What the next step has to be isn't quite determined, either. There's simply a lot of freedom when one designs a proof.
It's very likely that some other mathematicians will improve Zhang's proof so that they will reduce the constant 70 million to something smaller. Such proofs may be perhaps obtained as "modest mutations" of Zhang's machinery. However, it's unlikely that someone will reduce the constant 70 million to a constant smaller than 6 while keeping the bulk of Zhang's proof intact because certain tools become inapplicable for such small gaps (see the Math Overflow summary of the proof).
The proof of the actual twin prime conjecture will probably have to be completely different than Zhang's proof. It's nice that he has achieved a rigorous proof of a theorem that is a weaker version of the twin prime conjecture but I doubt that one can learn a lot by studying the details of his proof. There had to be so much freedom when he designed it. So it's like a NASA rocket engineer's decision to study every detail of a Soyuz aircraft. I don't think that this is the most important activity needed to conquer the outer space. Much like the Soyuz spaceships, Zhang's proof probably have many idiosyncrasies reflect the Russians' and the Chinese-American man's suboptimal approach to problems.
In mathematics and theoretical physics, when something is just being proved, we often encounter two different situations: in one subclass, the methods needed to prove something give us such new insights that these insights – methods, auxiliary structures that were used to complete the proof, and so on – are actually more valuable than the statement that has been proven. But I tend to think that Zhang's proof belongs to the opposite class of situations – in which the proof is less important than the assertion because it's composed of many idiosyncratic steps and tricks that are probably inapplicable elsewhere and that may be replaced by completely different "building blocks" to prove even the desired proposition.
Of course that I can't be quite sure about this pessimistic appraisal of the proof's methodology if I haven't actually mastered the proof. But because of general reasons and experience, I believe it's the case, anyway. Moreover, I tend to believe that the theorem proved by Zhang – and even the twin prime conjecture that may be proved in the future – is extremely weak relatively to some rigorous formulations of Gell-Mann's totalitarian principle applied here which says something like "the distribution of primes is random except for [simple divisibility-based] patterns that may be easily demonstrated". I tend to believe that such a principle will ultimately be formulated in a rigorous way and proved by a rather simple yet ingenious method, too.
You should understand that if I believe that this elegant goal is a legitimate, finite, \({\mathcal O}(1)\) task for some future mathematicians, it's also reasonable for me to believe that the assertion by Zhang and its seemingly cumbersome proof is a nearly infinitesimal fraction of what mathematicians will achieve sometime in the future. Zhang's proof represents a kind of the cutting edge that the mathematicians are able to prove about similar propositions today. But do I really care about the cutting edge? This cutting edge, much like most cutting edges in mathematics, is made terribly modest by the mathematicians' uncompromising insistence on complete rigor. If one is actually interested in the truth and is satisfied with arguments suggesting that something is true at the 5-sigma or 10-sigma confidence level, in some counting, the cutting edge is elsewhere – it's much further.
So of course that the hunt for strictly rigorous proofs that has defined mathematics after its divorce with physics is a legitimate goal – a constraint worshiped by a large group of professionals, the mathematicians in the modern sense. However, the strict rules of this hunt inevitably imply that in many cases, these professionals place themselves miles beneath the actual cutting edge of knowledge as I understand it.
And that's the memo.
Yitang Zhang worked at Subway before he would land a mathematics job. And when he did, he wasn't publishing almost anything for years before he would offer a proof of something rather important weeks ago. That turned the name of the popular math instructor in New Hampshire into one of the most well-known names of number theorists in the world.
Some increasingly popular links are:
Bounded gaps between primes (Zhang's technical paper)If \(p_1,p_2,p_3,\dots =2,3,5,\dots\) denotes the \(n\)-th prime, the statement proven by Zhang may be phrased in a very simple way:\[
Philosophy behind the proof (Math Overflow)
First proof that... (Nature)
Prime number breakthrough by unknown professor (Telegraph)
\liminf_{n\to\infty} (p_{n+1}-p_n) \lt 70,000,000.
\] The operator above is called the limit inferior which is just\[
\liminf_{n\to\infty}x_n := \lim_{n\to\infty}\Big(\inf_{m\geq n}x_m\Big)
\] If you think about this limit of the infimum for a while, you will understand that the limit inferior in the claim proved by Zhang is just the smallest gap between the adjacent primes that is realized infinitely many times (for infinitely many pairs). In other words, there exists at least one number – a potential gap between adjacent primes – that is realized infinitely many times.
Because of some technical properties that probably depended on many personal choices that Zhang has made while attacking the problem, the upper bound in the inequality turns out to be a high number, namely 70 million. It is such a high number that for all practical purposes, the proposition proven by Zhang is de facto equivalent to\[
\liminf_{n\to\infty} (p_{n+1}-p_n) \lt \infty
\] i.e. to the claim that there exists a finite number that is realized as the gap between adjacent primes in infinitely many pairs.
On the other hand, as I will argue, the actually correct (but rigorously unproven) claim stronger than Zhang's theorem says\[
\liminf_{n\to\infty} (p_{n+1}-p_n) = 2
\] which means that even twin primes – pairs of primes that differ by two – are realized infinitely many times: there are infinitely many pairs of twin primes. This claim is the famous twin prime conjecture. In some sense, the assertion proven by Zhang is 35 million times weaker than the twin prime conjecture. Note that the first twin primes are\[
(3, 5), (5, 7), (11, 13), (17, 19), (29, 31), (41, 43), (59, 61), \\
(71, 73), (101, 103), (107, 109), (137, 139), \dots
\] and there doesn't seem to be the tiniest reason to think that the list should terminate at some point. The largest currently known twin prime pair is \(2,003,663,613\cdot 2^{195,000}\pm 1\), two similar numbers that have 58,711 digits (each).
I don't plan to study the proof in detail because it looks very complicated and "non-unique" to me. The proven statement is slightly interesting but the proof is probably less interesting – there's just a small chance that I am wrong – and there's less "profound message" to learn from it. It's like if you are interested in the Moon and someone asks you to study O-rings in the Apollo spacecraft.
Moreover, I am not too interested in the claim that has been proved. But there is one more key reason: I feel certain that the proposition is true. The reason behind this certainty is the validity of a much stronger claim – a not quite rigorously defined one – that implies the twin prime conjecture, Zhang's proof, and many and many other much weaker corollaries. The claim is that
except for patterns that may be easily proved, the prime integers are distributed randomly and independently with \(1/ \ln n\) being the probability that a random number close to \(n\) is a prime.This general – somewhat vague but still very important – claim has many consequences, including the Riemann Hypothesis. In fact, the character of the proposition above is more or less a special case of Gell-Mann's totalitarian principle in physics:
Everything that isn't forbidden is mandatory.By this quote that generalizes the experience of the people suffering under totalitarian regimes such as communism and Nazism (there's no freedom: they tell you what to do and what not to do) to all of physics, Gell-Mann meant that the coefficient of every interaction in a Lagrangian or the probability of any resulting complicated process is nonzero unless one may use a symmetry or another rock-solid principle to prove that the coefficient is zero (because it violates the symmetry or another sacred principle).
In this analogy, "patterns that are easy to prove" are analogous to the "symmetries or other principles forbidding certain things". May I explain what I mean in the case of primes?
A pattern that is easy to prove is, for example, that if \(n\) is a prime, then \(n+1\) and \(n-1\) are not primes, assuming that \(n\gt 3\). It's because except for \(n=2\), only odd numbers may be prime. Similarly, among six consecutive integers greater than \(12\), just to be sure, at most two numbers may be primes. It's because only three numbers among the six are odd; and one of them is a multiple of three. One could continue with many examples of this kind.
Similarly, using the Gell-Mann totalitarian principle, one may demonstrate that the twin prime conjecture and its generalizations hold. There doesn't seem to be any reason why the difference between primes shouldn't be equal to two (or some other allowed even numbers) – there are many examples in which it is two, in fact – so there must exist infinitely many examples for the probability that \(n\) and \(n+2\) are both primes to be nonzero. Of course, it's hard to prove that "there is no reason why twin primes should stop at some point", either, but at least, one may prove that there exist no "reasons of the well-known types".
A TRF-based heuristic proof of the prime number theorem
Now, the density of primes around \(n\) asymptotically goes like \(1/ \ln n\). This is the right estimate for \(n\to\infty\), including the right numerical prefactor (the relative error goes to zero in the limit). This statement is known as the prime number theorem and it is a severely weakened sibling of the Riemann Hypothesis which may be equivalently stated as the (easy-to-prove) assertion that the roots of \(\zeta(s)\) only exist for \(s\in\RR\) or in the critical strip \(0\leq {\rm Re}(s)\leq 1\).
I can offer you a supersimple, Lumoesque argument why the density of primes goes like \(1/\ln(n)\). Call the functional dependence of the density \(\rho(n)\); it's really the probability that the number around \(n\) is prime. A number \(n\) is prime if it is not divisible by any prime smaller than or equal to \(\sqrt{n}\). These are statistically independent conditions. So\[
\rho(n) = P_{n\in{\rm primes}} = \prod_{p\leq \sqrt{n}}^{p\in{\rm primes}} (1-1/p)=\dots
\] because the probability that a random large \(n\) isn't a multiple of \(p\) equals \(1-1/p\). But the product may be written as the exponential of the sum of logarithms\[
\dots = \exp\sum_{p\leq \sqrt{n}}^{p\in{\rm primes}} \ln (1-1/p) = \dots
\] and the sum over primes \(p\) may be approximated by the sum over all integers \(i\) weighted by the probability \(\rho(i)\) that \(i\) is prime:\[
\rho(n) = \dots = \exp \sum_{i\leq\sqrt{n}} \rho(i) \ln (1-1/i).
\] Now, the sum over \(i\) may be approximated by an integral when \(\rho(i)\) is smoothened. Take the logarithm of the identity above (with the sum replaced by the integral)\[
\ln\rho(n) = \dots = \int_1^{\sqrt{n}} \dd i\, \rho(i) \ln (1-1/i)
\] and differentiate it with respect to \(n\) to get\[
\frac{\rho'(n)}{\rho(n)} = \frac{1}{2\sqrt{n}} \rho(\sqrt{n}) \ln(1-1/ \sqrt{n})\sim -\frac{\rho(\sqrt{n})}{2n}
\] where \(1/2\sqrt{n}\) came from \(\dd(\sqrt{n})/\dd n\) and where \(\ln(1-x)\sim -x\) because \(x\to 0^+\). One may easily verify that \(\rho(n)\sim 1 / \ln(n)\) satisfies the identity above; both sides are equal to \(-1/ n\ln(n)\) in that case. Among uniformly, nicely decreasing functions \(\rho(n)\), this solution may be seen to be unique. Even the coefficient in front of the logarithm or, equivalently, the base of the logarithm (\(e\)) may be seen to be determined by the (nonlinear) condition above.
You may check how Terence Tao imagines a heuristic proof of the prime number theorem. I leave you to decide who among the two of us is the cumbersome overworked craftsman and who is the seer. ;-)
At any rate, the heuristic proofs above aren't rigorous but one may rigorously prove the prime number theorem. One may also prove other things. As you can see by comparing various proofs sketched by various people – or the same people at various moments – there are many strategies that may be used to attack similar problems. When we're rigorously proving something like that in mathematics, we often work with lots of inequalities – not only the final one that e.g. Zhang has proved; but also with many inequalities in the intermediate steps. And the inequalities are usually ad hoc. We want to find an object that is "good enough to achieve a certain next step" but how good this good enough object has to be isn't quite determined. What the next step has to be isn't quite determined, either. There's simply a lot of freedom when one designs a proof.
It's very likely that some other mathematicians will improve Zhang's proof so that they will reduce the constant 70 million to something smaller. Such proofs may be perhaps obtained as "modest mutations" of Zhang's machinery. However, it's unlikely that someone will reduce the constant 70 million to a constant smaller than 6 while keeping the bulk of Zhang's proof intact because certain tools become inapplicable for such small gaps (see the Math Overflow summary of the proof).
The proof of the actual twin prime conjecture will probably have to be completely different than Zhang's proof. It's nice that he has achieved a rigorous proof of a theorem that is a weaker version of the twin prime conjecture but I doubt that one can learn a lot by studying the details of his proof. There had to be so much freedom when he designed it. So it's like a NASA rocket engineer's decision to study every detail of a Soyuz aircraft. I don't think that this is the most important activity needed to conquer the outer space. Much like the Soyuz spaceships, Zhang's proof probably have many idiosyncrasies reflect the Russians' and the Chinese-American man's suboptimal approach to problems.
In mathematics and theoretical physics, when something is just being proved, we often encounter two different situations: in one subclass, the methods needed to prove something give us such new insights that these insights – methods, auxiliary structures that were used to complete the proof, and so on – are actually more valuable than the statement that has been proven. But I tend to think that Zhang's proof belongs to the opposite class of situations – in which the proof is less important than the assertion because it's composed of many idiosyncratic steps and tricks that are probably inapplicable elsewhere and that may be replaced by completely different "building blocks" to prove even the desired proposition.
Of course that I can't be quite sure about this pessimistic appraisal of the proof's methodology if I haven't actually mastered the proof. But because of general reasons and experience, I believe it's the case, anyway. Moreover, I tend to believe that the theorem proved by Zhang – and even the twin prime conjecture that may be proved in the future – is extremely weak relatively to some rigorous formulations of Gell-Mann's totalitarian principle applied here which says something like "the distribution of primes is random except for [simple divisibility-based] patterns that may be easily demonstrated". I tend to believe that such a principle will ultimately be formulated in a rigorous way and proved by a rather simple yet ingenious method, too.
You should understand that if I believe that this elegant goal is a legitimate, finite, \({\mathcal O}(1)\) task for some future mathematicians, it's also reasonable for me to believe that the assertion by Zhang and its seemingly cumbersome proof is a nearly infinitesimal fraction of what mathematicians will achieve sometime in the future. Zhang's proof represents a kind of the cutting edge that the mathematicians are able to prove about similar propositions today. But do I really care about the cutting edge? This cutting edge, much like most cutting edges in mathematics, is made terribly modest by the mathematicians' uncompromising insistence on complete rigor. If one is actually interested in the truth and is satisfied with arguments suggesting that something is true at the 5-sigma or 10-sigma confidence level, in some counting, the cutting edge is elsewhere – it's much further.
So of course that the hunt for strictly rigorous proofs that has defined mathematics after its divorce with physics is a legitimate goal – a constraint worshiped by a large group of professionals, the mathematicians in the modern sense. However, the strict rules of this hunt inevitably imply that in many cases, these professionals place themselves miles beneath the actual cutting edge of knowledge as I understand it.
And that's the memo.
Heuristic ideas about bounded prime gaps
Reviewed by DAL
on
May 28, 2013
Rating:
No comments: