A polynomial with coefficients in a field and of degree is determined by its evaluations at any distinct points. A common way to see this is via Lagrange interpolation. But what happens in the more general case where the coefficients come from a commutative ring with ? It’s easy to see that the statement fails. Consider e.g. , and let . Then vanishes everywhere on R (easy to check), despite having degree two. In particular, there are multiple polynomials of degree (viz. and the zero polynomial) that vanish at three distinct points e.g. .
The correct generalization of the statement can be derived by considering the Vandermonde matrix. Recall that, given points , the Vandermonde matrix is the x matrix consisting of powers of the . The matrix-vector product of and the vector of the coefficients of a polynomial then gives the vector of evaluations of at the points . Interpolation goes the other way, i.e. from evaluations to coefficients. So we’d like to be able to invert the Vandermonde matrix.
As it happens, a square matrix over a commutative ring with is invertible if and only if its determinant is invertible in (the construction of the inverse matrix in terms of the adjugate demonstrates this). The determinant of the Vandermonde can be shown (using only column operations and properties of the determinant) to be the product of the for . Thus we see that a polynomial of degree is determined by its evaluations at distinct points if the differences of these evaluation points are invertible in .
In fact, we can do much better: the differences don’t need to have inverses in , they just need to be invertible in a larger ring: it in fact suffices that the differences are not zero divisors in . For suppose that this is the case. Let be the multiplicative closure of the set of pairwise differences. Then contains no zero divisors, and so can be considered as a subring of its localization at the subset . Importantly, the pairwise differences have inverses in . Hence, by the above argument, any polynomial of degree with coefficients in is determined by its evaluations at our points, and this of course continues to hold when the coefficients (and evaluations) lie in the subring .
To return to the problematic example above: for any three distinct points in , either at least two of them are odd, or at least two of them are even, and in either case there will be a pair of distinct points whose difference is even and hence either zero or a zero divisor.
The Vandermonde matrix computes evaluations of polynomials from their coefficients via a matrix-vector product. The Vandermonde matrices are nested, i.e. each Vandermonde matrix is the principal submatrix of any larger Vandermonde matrix that uses (an extension of) the same sequence of evaluation points. For example, here is the (square) Vandermonde matrix for the evaluation points ; the Vandermonde matrix for is the x principal (i.e. top-left) submatrix:
If the inverse Vandermonde matrix exists, then the inverse computation (from evaluations to coefficients, i.e. polynomial interpolation), can also be performed as a matrix-vector product (for the inverse of the Vandermonde matrix to exist, it must be square and the evaluation points must be distinct). Unfortunately, the inverse Vandermonde matrices are no longer “nested” in the above sense. For example, here are the inverse Vandermonde matrices for evaluation points and .
Who cares? Well, it would be nice if they were nested since then one could pre-compute a sufficiently large inverse Vandermonde matrix, and be able to interpolate any polynomial that came along. But it just isn’t so! However, for the particular case where the evaluation points are the non-negative integers (and indeed in many more general cases), the neatness can be restored using a certain triangular decomposition of the Vandermonde matrix and its inverse.
Let’s write for the Vandermonde matrix for the evaluation points . Then can be expressed as a product of an lower-triangular, diagonal and upper-triangular matrices , , and i.e. . Thus . It turns out that all of these factors form nested families in the sense above, e.g. is the principal x submatrix of any , while is the principal x submatrix of , and so on. Thus if we pre-compute then we’ll be able to interpolate any polynomial of degree at most (by selecting the appropriate principal submatrix of each factor, and then performing three matrix-vector multiplications).
Furthermore, the entries of (also of , , ) are given by pleasing and useful recursive formulae, and these can be used to increase the size of your precomputed matrices as required. For instance, and the identity is easily adapted to a recurrence on the that allows us to extend as needed. The diagonal entries of are given by (recurrence obvious) while the entries of are given by the (signed) Sterling numbers of the first kind via . These quantities satisfy the recurrence with boundary conditions and for all .
These formulae were first derived in Vandermonde matrices on integer nodes (Eisinberg, Franzé, Pugliese; 1998). Another useful (and freely available) reference is Symmetric functions and the Vandermonde matrix (Oruç & Akmaz; 2004) which deals with the case of -integer nodes (take in their Theorem 4.1 to obtain the formulae above).