You are currently browsing the category archive for the ‘Uncategorized’ category.
When one considers the sequence of values of some reasonable arithmetic function, like the divisor function
, or the function
the number of prime factors of
, one may notice that excluding a few abnormal values, the sequence looks very much like a neat, recognizable sequence. Or at least, smoothing the sequence by averaging it produces a sequence which looks neat.
Thus has normal order
: almost all numbers
have about
; this is the theorem of Hardy-Ramanujan. It’s even better than this : for almost all numbers,
is within
of its model
. In fact the values of
follow a Gaussian distribution, as the Erdös-Kac theorem reveals.
In contrary the divisor function doesn’t have any normal order, but it does have an average order which is quite good-looking, namely
(while
has a normal order which, surprisingly isn’t
but a bit a less, namely
, showing that a few exceptionally highly divisble numbers are able to make the average deviate substantially).
Now of course on the primes and
collapse violently, being stuck at respectively 2 and 1. The question is whether just after such a shocking value, at
, one recovers more normal arithmetic indices, whether there are traces of the past catastrophe (or signs of the coming castrophe just before it).
Of course isn’t absolutely any number, for instance it’s surely even; but then is
just any number? It must also be said that it has higher chances to be mutiple of 3 than a generic number, in fact one in two chance instead of one in three, because
whith equal probability.
From the point of view of
There the answer is: yes, perfectly. Indeed, Heini Halberstam established that the Erdös-Kac theorem holds when is constrained to range in the shifted primes
. That is,
lies within
of
for a positive proportion
of primes, the density being again gaussian.
From the point of view of
Not quite. For this consider the Titchmarsh divisor problem, consisting in estimating . If
was replaceable by
as we suppose, then this sum would be asymptotic to
by Mertens formula. It turns out that it is in fact asymptotic to
, the constant prefactor being well over 1,9, so that it can be said that
has almost twice as many divisors as banal numbers of his size. Now remember it’s always even; now there are good heuristic reasons to believe that
, so that in average
is close to
, but 1,9 is still much larger sensibly larger than 4/3. Here the higher chances of
to be divisible by 3, 5 etc. in comparison to a banal number weigh enough to make the average deviate.
It would be interesting to determine whether has a normal order, in the same vein as Halberstam’s result.
Another criterion of banality: the probability of being squarefree
As we know, the density of squarefree numbers is . It is then reasonable to wonder if among numbers of the forme
, the proportion of square free is the same. It’s clear to see that it can’t quite be so: indeed,
has one on two chance of being divisible by the square 4 (while a generic number has of course one on four chance), one on six chances of being divisible by 9, one on
chance of being divisibly by the prime
. So one guesses that
is a little less likely to be squarefree than any number; indeed, it’s been proven by Mirsky that the proportion of squarefree numbers among numbers of the form
is
, to compare with
.
The property of being squarefree can be replaced by the property of being -free for instance, the density among shifted primes being then
, to compare with
.
Appendix: average of the divisor function on odd and even numbers
We know that
And let us suppose that and
.
Now the sum on even numbers can be decomposed into a sum on numbers of 2-adic valuation 1,2,3… But numbers of 2-adic valuation k are numbers which are of the form 2^km with odd, so
for fixed
. Pretending this asymptotic also holds when
is allowed to grow with
, we infer from
that
. Hence
and
.
It might be that this intuition has been proven more rigorously, I’d like to see where!
The genuine prime number theorem says that there are about prime numbers between 1 and N. The prime number theorem in function fields is rather about the number
of irreducible monic polynomials of degree
in the polynomial ring
on a finite field
. It is remarkable that the result is of the same shape
, which is
if you let
be the number of monic polynomials of degree
. But the proof, as often with the polynomial analogues of number theory, is much neater and even yields a nice exact formula for
First proof
This one requires a good deal of knowledge of the theory finite fields.
We prove that is the product
of all monic irreducible polynomials whose degree divide
, which immediately provides the above formula by comparison of the degrees. To prove this, notice that the roots of
are all the elements of the extension
of
, so that
, as an element of
can be factored as
. On the other hand, if
is an irreducible polynomial of degree
, it splits into linear factors in
. So
has a root
and all its other roots are images of
under the Frobenius element, more precisely,
so that (at first sight in
, but also in
because they are both in
– it is well known that the gcd is preserved under field extensions). Moreover, all monic irreducible
are coprimes, so that their product still divides
. Finally, we need to prove the converse divisibility, either in
or in
, which is equivalent. So let’s do it in the latter; take a factor
with
. Let
be the order of the Frobenius element on
, so that
and no smaller integer has this property; then
which is indeed an irreducible of degree
, hence a factor of
. Given the factorization of
and the fact that the factors
are coprime, their product
still divides
Second proof
This one proves exactly what we want and not more and uses only power series. We need the notation .
Let
be the zeta function (for s of real part >1). It is easy to see that by looking at the sum expression, and
by looking at the product expression. Putting
, both sides become nice power series so that we get the equality
and doing classical operations (logarithmic derivation and expansion in power series followed by termwise identification in particular) on both sides yield the result.
This proof seems to use about zero input of algebra, so its success looks like a miracle. The only step where a (tiny) bit of knowledge seems to be used is the transformation of the sum into an Euler product – it needs the existence of unicity of the decomposition of a monic polynomial into product of irreducibles.
A kind of sieve method ?
It could be reasonable to hope that the sieve of Eratosthenes would work better in the realm of polynomial rings. We’ll try to adapt it here. Notice that in this case we have count primes up to a certain large degree, and not of exactly certain large degree (both quantities are far from equivalent in fact !).
Let’s define to be the number of monic irreducible (let’s say primes) polynomials of degree at most n, and
. We also use Moebius’ function for monic polynomials, also defined in terms of the number of prime factors. The Moebius inversion formula
still holds in the polynomial world. We also write
Thus
and if we forget about we obtain exactly
this looks quite nice, but the error term, just like in the integer case, can’t easily be bounded by anything better than which we know will be vastly bigger than the main term. By the way this product can very easily be shown to be
but the exact asymptotic is not any easier than the PNT itself… So this sieve seems to have very little interest, given the convenience of the other methods.
Two posts in french have already appeared on this topic, showing that the maximal number of unit vectors having pairwise negative (resp. non-positive) inner products grow linearly wit the dimension of the space they live in. This is good fun; however, as mentioned in a comment, it becomes even better fun when one bothers asking, for
how big a family of unit vectors
can be if…
- … for all
, one requires
;
- … for all
, one requires
;
Let’s deal with the first question. This involves the remarkable fact that on the sphere of very large dimension, almost all the area is concentrated on a small band near the equator (or any equator). This is the phenomenon called concentration of measure ; so pick a vector . To continue your family, you have to pick a vector outside the spherical ball
. So the concentration of measure implies that the forbidden area is of order
. When you have picked already
vectors, the forbidden area for the choice of the next one is at most
which is strictly less than 1 until
. Then not all the sphere is forbidden, so you can accommodate certainly
: exponentially (in the dimension) many instead of linearly.
The second one is easier. Indeed,
which implies so we get a constant bound, certainly much lower than
for large dimensions.
I devised yesterday a false calculation of the area of the sphere . The idea is quite natural. Decompose the sphere in circles
of radius
. Thus it seems that the area of a sphere should be
. However, by the change of variable
, we easily notice that
, so we get
, instead of
, the well known value we should find. What’s wrong?
Well why should it be true on the first place? The intuitive reason why this should be true consists in approximating our sphere by a ziggurat of cylinders, of height 1/n and radius , thus getting something like
for the area of our tower. This Riemann sum does converge to
, but there’s no reason why it should be the area of the sphere. There’s no reason why it should be the area of anything definite either… Notice by cutting this picture by a plane, it amounts to approximating a (quarter of a) circle by a bunch of vertical segments… whose sum of lengths is of course constantly equal to 1, instead of converging to
.
In the same vein, take a half circle based on the segment [-1,1], and then two smaller half circles under it, and split again each arc into two arcs. This yields a sequence of curves which converge in the
norm to the straight line segment, but which have a constant length
which doesn’t converge to the length of the limit curve. Indeed, the length of a curve is not a continuous function, only semi-continous.
So we have to think hard about the notion of surface area of a (say convex) body in the space. Certainly the right way to proceed is via Minkowski content (in other words in this case: derivative of the volume of the sphere) or Hausdorff measure.
In the Green-Tao method appear quite often expressions such as for some system of affine linear forms
where the
are some positive integers. The most basic example would be
where the last equality is here only to insist heavily on the form a general statement will have to take. A bit more generally, one easily sees that if
is d-periodic, then its average until
is basically its average until
. Following the ideas from Green-Tao, in fact we can show that for our general problem, the asymptotic should be analogously
where
is the gcd of the d’s.
Thus, applying this method, one has frequently to deal with such « local densities », the intersection of the zero sets of affine forms. More generally again, one may have to deal with things like for some polynomial
. Say
where the
are affine linear forms, so that
is of degree
. Suppose also that no two of them are affinely dependant, even when reduced modulo
. Then to compute
, notice that in order for
to satisfy
, either two of the forms must be divisible by
in n, or one must be divisible by
there.
To analyse the first possibility, reduce the forms modulo p, so they are as seen affine forms on the field . They are non-zero form, so the image of each linear part is of dimension
and the kernels of dimension
; so the zero locus of each of our affine forms is a hyperplane, and the hypothesis of no affine dependance ensures that the intersection of any two of these hyperplanes is of dimension
. So there are only
common zero for each pair of forms. But each point of
gives rise to
antecedents in
under the map of reduction modulo p. Finally, this possibility provides a summand of at most
to
.
To compute now , notice that any such 0 must arise from a 0 mod p. Now there are
0 mod p, because the set of zeros is a hyperplane. Then we can use a form of Hensel’s lemma in several variables. To obtain such a statement, we rely on the following simple identy for any
:
which shows that if , so
latex \text{grad}(Q)\neq 0\text{ mod }p$ then the vector
is constrained to live in a hyperplane if
is to vanish. So there are only
elements
with
and
. Of course if
is affine then the gradient is simply the linear part. Thus the number of roots modulo
is
. By the union bound, we see that:
. In the end,
.
To find more about this calculation and similar ones and the reason why I am interested in them, look at this article here.
What I claim is that, letting being the nth prime
in other words the proportion of the gaps between primes up to N which are smaller than is tending to 0. Thus although we know that there are infinitely many bounded gaps, we see here that they are almost all of size at least logarithmic.
To prove it, we recall a classical application of Selberg’s sieve:
where the implied constant is absolute (does not depend on m).
Now this product is obviously less than than which by Mertens’s theorem and the obvious bound
implies that is at most
. But now the numerator of our fraction is obviously
The last term is simply which is indeed negligible compared to the number of primes until N, whence the result.
Notice that we could have argued a bit differently: indeed, (instead of
previously). Thus, the number of primes until N such that the next prime is at distance at most M is bounded by
, which is
as soon as
.
As a conclusion, gaps of size have density 0.
There are between 1 and N about perfect squares, which makes this set
a quite sparse set, of density
. Given that
, it could be reasonable to think that
, with some luck, cover all of
or at least have positive density. In fact it’s not the case.
To estimate the size of , notice first that it contains all the primes congruent to 1 mod 4, which is a well-known theorem of Fermat, which is about one half of the primes, so that
. This already a huge lot bigger than the squares alone.
In fact there are much more sums of two squares than these primes, because all their products are also sums of two squares. In fact, it contains all the products of such primes. A similar heuristics to the one leading to the prime number theorem suggests that the density of the numbers having no prime factor congruent to 3 mod 4 should be
, and the log of this product should be roughly
so we get a density . While this indeed is no proof, it should be believable. In general, a number n is a sum of two squares if and only if it is of the form
where a is product of primes congruent to 1 mod 4 and b is a product of primes congruent to 3 mod 4. Thus, the total number of sums of two squares seem to be bounded above (using the union bound) by something like
where the denotes a sum restricted over number whose prime factors are all 3 mod 4. But this should be of the order of
and the sum is bounded, so certainly . In fact, Landau has proven in 1908 that the number of sums of two squares until capital N is asymptotic to
which fits quite well with what the heuristics indicate (the product over the restricted primes corresponds perfectly to the restricted sum).
Repartition of the sums of squares in residue classes
We see that the primes are analogous to the sums of two squares: indeed, the primes are the numbers which have no factor, and the sums of two squares are the numbers which have no factor congruent to 3 mod 4 (plus other ones of course, the products of such numbers by 9, 49, 121 etc. of which we saw that it simply multiplied the size of the set by a constant). The same heuristic provides a density which then turns out to be correct. So why not, as Dirichlet did for the primes, ask for the repartition of the sums of squares in residue classes?
A sum of two squares can only be congruent to 0,1 or 2 mod 4, because squares are 0 or 1 mod 4. Let’s show that these values are not equiprobable in fact. First notice that
can be written as
where
is also a sum of two squares. So given the number
of
until
which are sums of two squares, we get
and
. At least, the probabilities of being even or odd are the same. Another instance is that the sums of two squares are twice as dense in
than in
. Also notice that everybody in
has
so nobody there can be a sum of two squares. The final word is by Prachar. Putting
the constant appearing above, he found
with
when ; thus the density is the same as the general density in APs, except when the modulus
is divisible by 4 (which causes a multiplication by 2 of the density) and/or it has factors congruent to 3 mod 4.
Additive quadruples in the squares
Recall that by Cauchy-Schwarz, where
is the number of additive quadruples, i.e. of quadruples of square
with
, which is also the number of quadruples
with
. Thus, we have at least
additive quadruples.
We can try to do better. Indeed, rewrite the equation as
(we replaced by N here). It amounts to
We’ll count the positive solutions here, the other solutions are easily obtained from them. We partition such quadruples depending on the gcd . So
. Then
for some
. So the number of add quadruples should be
By Cauchy-Schwarz, the inner sum ought to be
Now the second sum is not much bigger than
so the total number of additive quadruples should be
. Returning to the previous definition, we have the inequalities
Notice that in a random set of this density you would expect less solutions: of the order of magnitude of N. So if are squares,
has increased probability to be a square.
Anyway the heuristics that for
of density
implies that that
which is not the case here.
Sums of two squares in
We know that in , there are
squares (including 0; here
is a prime). This suffices to show, as a simple case of Cauchy-Davenport, that everybody is congruent mod p to a sum of two squares. This of course doesn’t extend to non-prime
, as for instance not everybody is a sum of two squares modulo 4. Let us denote by
the number of representations of a as sum of two squares mod d. Then we can see that for
a prime congruent to 1 mod 4 (so that
is a square mod p),
, while for any other residue
, we must have
. On the other hand
for
, and hence
for all other
.
By the Chinese Remainder theorem, we know that . But it is far from obvious to estimate the
. In fact it happens that
where is the only non-trivial character mod 4, which is also Legendre symbol mod 4.
This post is an echo to Davide Ravotti’s brilliant talk one week ago.
Equidistribution has already been discussed here on this blog (during the good old epoch of francophony…) and it is a fascinating property. It must be observed that it isn’t the property of a countable set but of a rational sequence; in other words it depends crucially on the order in which the terms are considered.
The rationals between 0 and 1 are countable for instance. The usual way to see it is, as Cantor did, to naturally embed in
and to order the couples of positive integers
by increasing
. But number theorists know that the right way to order rationals is by increasing denominator. Indeed, in diophantine approximation, the denominator determines the quality of the approximation. The fractions
, at fixed q, follow each other at distance
so these produce sequences which are of course evenly distributed and denser and denser. So let’s introduce the Farey sequence which is the most relevant ordering of the rationals.
The Farey sequence
We denote for
the fractions
, where
are the integers coprime to and smaller than
. This yields an enumeration of the rationals:
… In other words, our sequence is defined by
where
is the only integer such that
and
. We may want to add
. The Farey sequence of order n is
. Every rational appears in
because every rational can be put in lowest terms
with
.
The Farey sequence of order n yields a dissection of the unit interval (or rather of the circle if you want to identify both ends, which we do in the end if we use Weyl’s characterisation of equidistribution) in the following way: we look at all the mediants of successive pairs. The intervals between two mediants will be our Farey arcs and each fraction of the Farey sequence will be contained in one arc. For instance 0=1=1/1 is between the points
and
so that the mediants formed with them and 1/1 are
and
, hence the arc containing 0 is
. It is easy to see that the arc containing
is made of two parts (one at each side of the Farey fraction conatined in it) both of length between
and
. This denotes a certain uniformity. It also provides again another proof of Liouville’s theorem by the way (see Hardy & Wright).
Neville indeed proved in 1949 that the Farey sequence satisfies Weyl’s criterion and hence is indeed equidistributed. In fact the equidistribution remains true however we order the people inside each
sequence.
Connection to Riemann’s hypothesis
Jérôme Franel observed in 1924 that Riemann’s hypothesis would imply (and is equivalent to the fact) that on average the i-th fraction of
is very close to
where
. So the Farey sequence is quite close to the evenly spaced sequence of fractions of the same length. Precisely, putting
, Riemann’s hypothesis is equivalent to
for all
.
Other sequences of rationals and generalisations to higher dimensions
Ordering the elements amounts to ordering the elements
(in both case we want to restrict to
which we call primitive tuples). Here an « optical » description is useful, as strikingly it is already in the theory of Farey sequences to prove that if two elements
are consecutive, then
.
As the picture of the Farey sequence of order 8 above shows, one can imagine an eye looking from the origin into to the first quadrant. He only sees the first integer point on each line; hence 2/8 is hidden from its sight by 1/4, its primitive form. Two fractions are successive if and only if there’s no point in the parallelogram between both vectors, which happens only if it has area 1, which implies
.
To order the elements, we use a piecewise linear function with integer coefficients. Thus the Farey case is
and the « Cantor » case is
. Given such a function, we order the integral points according to the value of the function on them; in order words we partition the integral points of the cone into level sets of
. So we need that each of this level set have a compact intersection with the cone, so that we have a finite number of integral points on each level set…
If this is proportional to the volume of , this means that we have equidistribution of the sequence. Indeed it’s the generalisation of case
, where simplices are simply intervals.
From now, we restrict without loss of generality to a linear . The key step is to compute
for integer, where
. We can in fact drop the primitivity condition, and retrieve the cardinality searched by Moebius inversion.
Ehrhart’s theory
This is precisely the scope of Ehrhart theory, dealing with things such as for
a rational polyhedron (=with rational vertices).
Notice that Eugène Ehrhart was an alsacian high-school teacher who didn’t complete his PhD before his 60th birthday…
Examples
then
- if P is the standard simplex then a classical counting argument yields
- if P is a convex polygon of the plane with integral vertices then Pick’s formula says that
This looks polynomial and it isn’t a coincidence.
Ehrhart’s theorems:
- If
is integral polyhedron of dimension d, then
is a rational polynomial of degree
, whose dominating coefficient is the relative volume
of the polyhedron.
- If
is integral polyhedron of dimension d, then
is a rational quasi-polynomial of degree
, i.e. there exist functions
which are periodic, of periods divising the lcm of the denominators of the vertices such that
.
The second statement comprises the first. The relative volume is defined as follows: there exists an affine isomorphism
such that
and then
, which doesn’t depend on the choice of
, because two possible choices differ by a transformation of determinant 1, hence give same volume.
Notice that for a general convex set of non-zero volume, the arguments from Tao-Vu’s book Additive combinatorics or Green-Tao’s Linear equations in primes (appendix A) give that
. It is for instance quite well known (Gauss circle problem) how many integral points there are asymptotically in a big disc of Radius R: about
.
Finally, performing the Moebius inversion:
and finally where
according to some number theory (cf the exercises of Murty).
In the end of the day, we get that . To be quite honest, it isn’t exactly the case when
is not integral but simply rational. Indeed, the leading coefficient
in Ehrhart’s (quasi)-polynomial of a polyhedron
is not always constantly equal to the relative volume
. To see this, take
. This is a nice rational polyhedron of dimension 1 and relative volume 1 (morally it’s the volume when seen in its « natural environment », its affine hull). Then
contains an integral point if and only if
is even, so
. So instead of
, we should get here
where b is the first integer so that
contains an integral point.
When do we get equirepartition? Probably essentially when for all simplices S. This is certainly only the case when
all the time, which seems to imply that
, in other words the Farey case.
Given , we would like to find solutions, probably rather approximate solutions, to the diophantine equation
(so the unknown
are constrained to live in
; we may also require them to be coprime, and the denominator shall always be positive). This equation has an exact integral solution if and only if
is rational. Finding an approximate solution amounts to finding a fraction
approximating
. We want to define a notion of best approximation. Of course there are rationals arbitrarily close to
, but their denominators are then unbounded; so we fix an amount we are ready to pay, or rather a bound to the denominators, say
, and we restrict our attention to fractions
with
. We will say that
with
is a best approximation (of first kind) of order Q if
for all
. The second kind would be to demand
to be minimal; it happens that both awards aren't always awarded to the same candidate.
How well can a number be approximated ?
Given a bound on the denominators, how close can be fractions with denominator
to
? Let
be the distance to the nearest integer, so we are looking at
. A simple application of the pigeonhole principle ensures that there exists
with
, so in particular
and there exists a sequence of fractions
with
. Note that the result is not constructive. This exponent 2 is nice, but can we do better? Can we be sure to find an unbounded sequence of
with
for instance?
The answer is no: Liouville proved that if is an irrational algebraic number of degree at most d (so that there exists a unitary rational polynomial of degree
such that
, then there exists a constant
such that for any p,q,
. Thus for quadratic irrationals (irrationals which are algebraic of degree 2), the exponent 2 is really the best possible. But the discussion concerning the optimal constant C is still open.
So if you define then
for algebraic irrationals of degree
.
This theorem seems paradoxical because it states that the irrationals which seem to be the least irrational, the algebraic ones of low degree, are actually the most difficult to approximate. The paradox vanishes (or at least fades) when one remarks that the rationals are actually the most difficult to approximate: indeed, if is in lowest terms, and if
is a rational in lowest terms with
, then
so we get an exponent 1. Notice the trick that a non-zero integer is at least 1, which is also together with the intermediate value theorem, the trick to prove Liouville’s theorem. Liouville’s theorem is also the easiest gateway to exhibiting transcendental numbers; indeed Liouville found numbers with arbitrarily high
.
In fact, as far as the exponent is concerned, algebraic irrationals of high degree are not any easier to approximate than quadratic irrationals, as has been revealed by the theorem of Thue-Siegel-Roth, saying that for an algebraic irrational and any
, there exists a constant
such that for all integers
,
.
The best exponent is 2 in general, but what about the constant ?
We can always find a sequence with
, and for quadratic irrationals,
so between
and 1, what is the optimal constant? The truth, as Hurwitz’s theorem says, is that
for infinitely many
. And for the golden ratio and its conjugates,
is optimal; for all other numbers, one can do better, certainly arbitrarily better (arbitrarily big constants) for algebraic irrationals of degree more than 2, and merely a better constant for other quadratic irrationals. We’ll come back to this later.
How to construct good approximations (exponent 2)?
The existence of a sequence with
is always granted but how to construct one? If you naively look at the decimal expansion of your number, you only get an approximation of quality
.
So the correct way to proceed is to use continuous fraction expansion. So take , and look at the map
. Put
and
. Then
. If you write
the kth convergent, then you can prove that
, which is the quadratic approximation we desired.
Best approximation
Not only do continuous fractions yield a natural way towards quadratic approximation, they also provide the « best approximations » in the following sense. We fix a number (let’s say for convenience that it’s irrational); we say a fraction
is a best approximation if every other fraction
with
satisfies
.
Then every best approximation is a convergent and any convergent is a best approximation
Which numbers are the toughest to approximate ?
We have seen that for every number , there exists a convergent sequence
with
. The only numbers for which
is optimal are the golden ratio
and his conjugate, solution of the quadratic equation
. We know that
and
, so
and by injecting the description of
in the right-hand side, we get
so the convergents are plainly
. In particular it is periodic.
We could try to characterize the numbers for which the (called the partial quotients) are periodic. The same question is easy to answer for the other pleasant type of rational approximations, the decimal approximations: for which numbers is the sequence of digits eventually periodic, i.e.
where the bar means that the sequence is repeated ad vitam? The answer is very easy: it is the rationals. So for which numbers do we have
? Lagrange proved that these were exactly the quadratic irrationals.
Now from Liouville’s theorem, and Thue-Siegel-Roth, we suspect that the toughest numbers to approximate are the quadratic irrationals, so the numbers which have periodic partial quotients. The answer is slightly more general: it is the numbers which have bounded partial quotients. Indeed, we can formulate the following alternative:
For any irrational number and integer
, either there are infinitely many positive
such that
or
for all sufficiently large indices
. This is a generalization of Hurwitz' theorem.
Other neighbouring questions
What we have just seen is a characterization of the real numbers such that
: they are the real numbers with bounded partial quotients
. Borel has shown that this set has Lebesgue measure 0. But of course they do exist, the golden ratio for instance. Now what if you take a second real number
and ask for
: will you always find 0? Littlewood’s conjecture is that it is the case.
Happy new year to all readers! This year I hope will be full of mathematical reflexions here, alas certainly again in English, in order to reach new audience (I’ve just noticed I’m referenced on Terry Tao’s weblog for instance).
I have already discussed what this same Tao calls the dichotomy principle (« a massive lot or very few ») for polynomials. It was essentially the case of polynomials in two variables, as in the traditional Bezout’s theorem: two coprime polynomials of degrees
have at most
common zeros,
where can in fact be replaced by any field, and
with the inequality being an equality in the algebraic closure, if one counts points at infinity and multiplicities of intersection. This can be proven either with the resultant, or by some argument of dimensions.
Lines in curves, planes in surfaces
For instance it implies that if a polynomial of degree vanishes on
distinct points of the line
, then
is a factor of this polynomial. However it’s not relevant at all in this case to invoke Bezout’s theorem, because the statement is quite immediate in the case where the line has equation
– expand according to the powers of
, so latex
and notice that this line is in fact fairly general (by change of variables).
What about a polynomial in three variables now, so an algebraic surface of the space instead of an algebraic curve in the plane? An algebraic surface can very well include a full line and still be irreducible, as show ruled surfaces like the hyperbolic paraboloid . Lines are not the zero set of a truly irreducible polynomial anymore; indeed a line is defined by two equations of degree 1
and
. But this system of equations is in
synonymous with
, which involves a polynomial of degree 2 which is irreducible over
but not over
. But an algebraic surface cannot contain a plane, without the degree 1 equation polynomial defining the plane to factor out in it.
Intersection points, intersection lines
In dimension , a reasonable generalization of Bezout’s theorem must involve enough hypersurfaces (of codimension 1) in order for the intersection to be expected to be points, so of dimension 0, hence you need
hypersurfaces. If their degrees are
, it is reasonable to guess
intersection points. But the example of the hyperbolic paraboloid (degree 2) with a line included in it, defined by two equations of degree 1, yields infinitely many points of intersections, instead of the expected 2… Although all the equations are coprime. But the correct statement is that if the number of intersections is finite, it’s at most the product of the degrees.
However, a variant of Bezout’s theorem involving two surfaces in $\mathbb{R}^3$ is still possible. Surely, you won’t bound the number of points of intersection, but you can bound the number of lines of intersection: the intersection of two coprime surfaces of degree and
contains at most
points. A proof of this can be seen on Larry Guth’s lecture notes on polynomial method lecture 13.
Monge-Cayley-Salmon: flecnodal points, ruled surfaces
Salmon’s theorem, hinted at by Monge earlier and also stated by Cayley, says that if an algebraic surface is such that at every point
there exists a line
tangent up to order 3 to
(flecnodal point), then for all points
, the line
is in fact fully included and so the surface is ruled. As a line tangent up to order 3, on a surface of degree 2 has to be included, we suppose that
is squarefree and of degree at most 3. Then flecnodal points of
are determined by a polynomial
of degree
. Together with Bezout’s theorem, it yields the following dichotomy for a surface of degree
:
1) Either ruled, so one line in through each point
2) Or only on
.
This is one of the basic ingredients for Guth-Katz solution of the Erdös distinct distance. problem.
Commentaires récents