No, we are not exaggerating. We are instead simplifying.
In fact, many of the phenomena associated with the main statistic models are better understood if studied, at least in the first measure, from the projective point of view and on an algebraically closed numerical field.
The main link between Algebraic Statistics and Projective Algebraic Geometry is based on the constructions of this chapter.
4.1 Motivations
We have seen as many models of interest in the fields of statistics are defined, in the space of the distributions of a system S, by (polynomials) algebraic equations of degree greater than one. To understand such models, a mathematical approach is to initially study the subsets of a space defined by the vanishing of polynomials. This is equivalent to studying the theory of solutions of systems of polynomial equations of arbitrary degree, which goes under the name of Algebraic Geometry.
The methods of Algebraic Geometry are based on various theories: certainly, on Linear and Multi-linear Algebra, but also on rings theory (in particular, on the theory of rings of polynomials) and on Complex Analysis. We will repeat here only a part of the main results that can be applied to statistical problems. It must be kept in mind, however, that the theory in question is rather developed, and topics that will not be introduced here, they could be important in the future also from a statistical point of view.
The first step to take is to define the ambient space in which we move. Being to study solutions of nonlinear equations, from an algebraic point of view, it is natural to pass from the real field, fundamental for applications but without some algebraic elements, to the complex field which, being algebraically closed, allows a complete understanding of the solutions of polynomial equations.
We will then have to expand, from a theoretical point of view, the space of distributions, to admit points with complex coordinates. These points, corresponding to distributions with complex numbers, will allow us a more immediate characterization of the sets of solutions of algebraic systems. Naturally, when we have to reread the results in the common statistical theory, we will have to return to models exclusively defined on the real field, that intersects with the real space contained in every complex space. This final step, which in general poses technical problems that are absolutely nontrivial, can, however, be overlooked in a first reading, where the indications that we will get on the complexes will still help us to understand of real phenomena.
Once this first enlargement is accepted, to arrive at an understanding even more in-depth of algebraic phenomena, it is opportune to operate a second step, perhaps apparently even more challenging: the passage from affine spaces to the associated projective spaces.
The reasons for this second enlargement are justified, from a geometric point of view, with the need to work in compact ambients. Compactness is indeed an essential property for our geometric understanding. In very descriptive terms, thanks to the introduction of points to infinity, we will avoid to lose the solutions when, passing to the limit, we should have phenomena of parallelism or however asymptotic phenomena. The possibility of following the reasonings passing to the limit is one of the winning cards that geometry in projective setting offers, compared to that in similar environments.
Of course, projective compactification, to make sense in statistical problems, it must be carried out properly, differentiating, for example, the passage to the limit of the various random variables.
For those who find too cumbersome the procedure to use homogeneous coordinates to describe distributions on random systems, perhaps it is worth remembering that a similar procedure, in statistics, is always present: the normalization. In practice, if we have a distribution D on a random variable x having states , then it is natural to replace D with the distribution obtained dividing each by the sampling (in the event that x is not neutral with respect to D). Note that, in doing so, in the space of distributions that concerns the variable x, we get to replace the point with the point . If, in the affine space, the point is changed, in the projective space, where the n -tuple represents homogeneous coordinates, passing to normalization, the point does not change! In effect, every point in the projective space can always be represented by homogeneous coordinates such that .
From another point of view, the classical statistical theory, in the space of the distributions of a random variable x as above, tended to shrink to the hyperplane defined by the equation , as we have seen, for example, in the statement of Varcenko’s Theorem. Instead, projective statistical theory works on distributions up to scaling, then does not need this restriction, since from a projective point of view, normalization, like any other scaling, is irrelevant.
Therefore, it is quite simple to convince ourselves that, in the end, these are two equivalent approaches. The difficulty of passing from one to another basically resides in habit. The advantage of using the projective language consists in being able to access directly to the vast literature on Algebraic Geometry which, in many ways, mainly uses this terminology.
During this chapter, we will use many topics from the Part III. We strongly recommend to the reader which is not expert in Algebraic Geometry to study such chapters before to start the treatment of projective algebraic models.
4.2 Projective Algebraic Models
What is the link of all of this with Algebraic Statistics?
When we meet a distribution , where S is a system of random variables, we are, in practice, collecting a set of observed data. To the aim of our interpretation of data, is usually irrelevant (within certain reasonable limits) the sampling of the variables.
If, for example, we are evaluating the efficiency of a medicine, to administer the medicine to 100 sick people and having 50 healed suggests that the medicine is efficient in about the 50% of cases. From this point of view, we obtain the same conclusion if we administer the medicine to 120 sick and record 60 healed.
So, in our setting, we will consider that a distribution gives the same information on the analyzed phenomenon than any of its scaling.
In Statistics, the problem is solved by choosing, among all possible scaling of a distribution, the associated probabilistic distribution, introduced in Definition 1.3.2. Such distribution is uniquely determined, given a distribution D, but it exists only when all variables have sampling different from 0 in D.
Definition 4.2.1
The elements of are thus equivalence classes, each of them containing a distribution D and all its scaling. In such a way we can link more easily (Projective) Algebraic Geometry with the study of significative statistical models.
In this new view, only the statistical models which are independent of the scaling (multiconi) in the space of distributions, are significative.
Because the overwhelming majority of important models (if correctly interpreted) are independent of the scaling, so they are cones, the depth investigation of our theory will not be affected.
Definition 4.2.2
A K-projective model on S is a subset of . A K-projective algebraic model on S is a model corresponding to a (multi)projective variety, i.e., defined by the vanishing of multi-homogeneous polynomials.
Example 4.2.3
The independence model can be thought as a projective algebraic model, because it is defined by many multi-homogeneous equations (see Theorem 6.4.13).
A linear model is a projective algebraic model when its defining linear polynomials do not have a constant term.
Example 4.2.4
The projective algebraic models on a system S with a single variable (as in the case of the total correlation) are strictly related with the cones of the vector space . Every cone defines a projective model on S. Vice versa, given a projective model on S, its preimage in the projection is a cone.
Example 4.2.5
Consider the system S formed by two ordinary dice. The projective space of distributions is . Instead, the projective space of distribution of is , corresponding to the space of matrices.
If S is a system where the two variables represent a die and a coin, the projective space of distributions is . In this case, it is easy to observe that the only variable in the total correlation has 12 states, hence the projective space of distributions of is .
If a system S has n Boolean variables, then its projective space of distributions is a product of n copies of , and the space of distributions of is .
4.3 Parametric Models
We are now able to define projective parametric models on A. For this aim, we will use Definition 9.3.1 of projective map, in Sect. 9.3.
Definition 4.3.1
If S, T are random systems, we call projective connection any projective map . Notice, in particular, that if is a projective connection, then the image of any scaling of a distribution D is a scaling of .
We say that a model M is projective parametric if it is the image of a projective connection .
Many interesting parametric models have a counterpart projective parametric.
Example 4.3.2
The independence model is projective parametric. As a matter of fact let S be a system with random variables and let be the number of states of the variable . Hence, the total correlation has a unique variable, with states.
It is evident from the same definition that the model of independence corresponds to a Segre variety (compare with Definition 10.5.9).
Notice that, in general, M is quite bigger than . For example, if and , then and the model corresponds to the Segre variety of embedded in .
Example 4.3.3
On a random system with three variables , the model without triple correlation of Example 3.2.5 it is not, strictly speaking, parametric projective.
In order to obtain our model as the image of a well-defined projective map, we must restrict to a subvariety of .
The fact that the image of a Segre map can be interpreted as (projective) model of independence of an random system, by Theorem 6.4.13, guarantees us that the Segre varieties are all projective varieties.
Let us see how, in general, there are projective parametric models which are not algebraic models.
Example 4.3.4
Consider two systems S, , both with a single random variable.
We identify both the projective spaces of distribution over , and , with . We can define a projective connection by posing .
It is easy to check that the image W of contains infinitely many points of . However, it does not contain all points: as a matter of fact, the point with homogenous coordinates is not in the image.
On the other hand, each projective variety in , being defined by the vanishing of a homogeneous polynomial in two variables, or coincides with , or it can only contain a finite number of points.
So W can not be a projective variety.
Example 4.3.5
Let us go back to the situation represented in Example 3.3.4.
Recall that the initial situation corresponds to a Boolean system S with a variable (of states A, B) while the final situation corresponded to a system with only one variable that could take the 3 values AA, AB, BB.
The connection , defined by , is clearly a projective map between and . The image corresponds to the subset defined by the points satisfying the equation .
It should be noted, however, that not all the homogeneous coordinates of these points can be obtained in the map. In fact the point P of coordinates (1, 2, 1) is in the image (we get it for , but no pair of gives , which are also coordinates of P.
From Chow’s Theorem (see Theorem 10.6.3), it immediately follows:
Theorem 4.3.6
Each projective parametric model is a projective algebraic model.
This theorem generalizes the situation already seen for the model of independence and explains how each projective parametric model can be defined by homogeneous polynomial equations.
The proof of Chow’s Theorem also explains theoretically how the homogeneous equations of a projective parametric model can be found.
In practice, as one can imagine, when the number of variables grows, it is not easy to follow step by step the directions, and find an effective set of equations, even with the help of computational tools. The use of Groebner bases, which we will introduce later, allows an optimization of the procedure (see Chap. 13).
The advantage of presenting a model with homogeneous equations (implicit equations), rather than through parametric equations, should instead be clear in the daily practice of Algebraic Statistics: to test whether a given phenomenon, i.e., a given distribution, falls within the model imagined by a theory (in more imaginative words: if an experiment confirms or less a theory), once the implicit equations are known, it is sufficient to check if the distribution satisfies them. A similar computation is elementary, for every single equation. In everyday practice, the complication derives only from the fact that normally each model is described by an astronomical number of equations, sometimes with approximate coefficients. However, these problems can be managed by methods of sample searching and error checking.
Instead, having to show that a given distribution belongs to a model of which only parametric equations are known, the problem moves to show the existence of parameters for which the parameterization returns the starting distribution. Such a problem of existence is extremely difficult to control, even in the presence of a few, precise equations. Imagine when the equations are thousands, with approximate coefficients!