A General Theory of Vandermonde Matrices
A General Theory of Vandermonde Matrices
4 (1986), 193—215
MW © Bibliographisches Institut 1986
T.Y. Lam 1
§1 Introduction
In history books of mathematics, Alexandre Théophile Vandermonde (1735—1796)
is usually credited for his study of resolvents and the group of permutations of the
roots of an equation, and for his role in isolating the theory of determinants as a
subject for independent study. Nowadays he is perhaps best remembered through
the beautiful matrix
1 l 1 (I)
“{1 a; an
V,,(a1,...,an)=
ail—1 a; 1... a:—1
1) Supported by NSF.
2) Actually, it is not altogether clear that the matrix (1) had occurred explicitly in
Vandermonde’s work. According to Lebesgue, “Vandermonde matrix” might have been a
misnomer because someone had misread subscripts for exponents (cf. [Edwards: 1984,
p. 18]). However, V3(a, b, c) did appear in Vandermonde’s work and the generalization from
3 x 3 to n x n seems to be a natural step.
‘ 3) In this paper, we use the terms “nonsingular” and “invertible” interchangeably.
0723-0869/86/030193-23 $6.10
17-7 A . l . hill“
with finite Fourier series, the polynomial interpolation problem, in solving differ-
'ential and difference equations, in curve fitting, in coding theory and in the study
of recursively defined sequences. For a survey of some of these applications, see
[Kalmanz 1984]. Of course, the Vandermonde matrix is also important for many
theoretical considerations in abstract algebra, for instance, in ring theory, algebraic
number theory, group representation theory, invariant theory and the theory of
symmetric functions [MacDonald: 1979].
Like so many other good notions in mathematics, the Vandermonde matrix is
capable of interesting generalizations. For instance, one may form the more
general n x n matrix whose i"l row is (a;‘, ...,a;‘) where 0=r1 <r2 < <r,, are given
integers. The resulting determinant, called a “higher alternant” ([Muir: 1933,
Ch. 11]), is important for combinatorics [Stanley: 1971] and for the study of the
irreducible characters of the symmetric group [Mumagham 1938], [Littlewoodz
1950]. Another way to generalize V,,(a,,...,a,,) is to form certain “differentiated”
Vandermonde matrices [Aitken'r 1959, p.119], [Norton: 1984]; their deter-
minants, called “confluent alternants”, are useful for solving “osculatory” in-
terpolation problems [Kalmanz 1984] and for studying shift register sequences
[Grundofer-Liineburg: 1980]. Yet other generalizations appeared in [Hirsch:
1949], [Carlitzz 1955] and [Famelk 1959].
In this paper, we shall give an exposition on a new generalization of the Vander-
monde matrix. Our generalization differs from the aforementioned ones in that it
contains the following two new ingredients: (1) we allow the elements a1, ...,a,l in
the Vandermonde matrix to be elements from a division ring D, so that they may
no longer commute with one another, and (2) our new Vandermonde matrix takes
into account an arbitrarily given endomorphism a of the division ring D. More
precisely, for given elements a1, ...,a,,eD, and a field endomorphism aHa‘ of D,
we define the a-Vandermonde matrix to be
1 1 1
N192.) naz) NM
V,,"(a1,...,a,,)= * (3)
Ml-l(al) N_1(a2) “Ml-1(an)
where, for any aeD, we define N,(a) to be a“"...a‘2a"a. In the special case when
a is the identity automorphism of D (we write a=1 for short), we have N,(a)=a",
so in this case the a-Vandermonde matrix boils down to the ordinary Vander-
monde matrix V,,(a1 , . . . , a").
Note that our a—Vandermonde matrix represents a nontrivial generalization of the
ordinary Vandermonde matrix even in the case when D is a field. We shall
mention two examples among many interesting ones.
(1) If D is a field of characteristic p>0, we can take a to be the endomorphism of
D defined by a"=a’ (aeD). A 4 x 4 a-Vandermonde matrix, say, looks like
A general theory of Vandermonde matrices 195
l 1 1 l
a b c d
ap+1 bp+l cp+1 dp+l ' (4)
with a dilemma. In the matrix M, it would be nice to “extract” the factors b—a
and c— a. If these were right factors instead of left factors, we would be able to
scale the two columns simultaneously by right multiplication of M by
((b (:1)_ (c (2)4). Since we are stuck with b-a and c-a as left factors, the
— _ —1
next best thing is to left multiply M by ((6 (:1) (c _:)_1) to get
((c—a)'1(b —a) I)
. Another row operation on the latter leads to
(c—a)‘1(b —a)b c
( (c—a)‘1(b—a) . l)
(c—a)‘1(b—a)b—c(c—a)‘l(b—a) 0 '
—a)“(b—a—-)b
Thus, for distinct a, b, ceD, V3(,a b, c) is singular if and only if (c—
=c(c—a)'1(b—a), or equivalently
(b—a)b(b -a)‘ =(c-a)c(c-—a)'l. , (10a)
At this point, to give the reader the proper perspective, we should point out that
the computation above is in fact what is needed to determine the so-called
Dieudonné determinant of V3(a, b, c) over the division ring D (see [Artinz 1957],
p. 151). For an invertible matrix V over a division ring D, the Dieudonné de-
terminant, det V, of V takes value in the commutator quotient group D*/[D*,D*],
where [D*,D*] is the commutator subgroup of the multiplicative group D*
=D\{0}. (For singular M, it is customary to take detV=0-[D*,D*].) Since
elementary matrices must have determinant 1, and det(diag(u1, ...,u,,))
=u1...u,,[D*,D*], the computation above shows that det V5(a, b, c) is given by the '
coset of
—(c —a)2[(c —a)' 1(b —a)b —c(c —a)‘ 1(b —a)]
=(c —a) [(c -a)c(c —a)‘ 1(b —a) —(b —a)b].
To those readers who may be appalled by this seemingly ugly expression, it is
perhaps comforting to note that, in case a commutes with both b and c, this
simplifies to the usual Vandermonde determinant (c —a)(b —a)(c —b)!
The above computation leads us to make several interesting observations. First,
by symmetry, the necessary and sufficient condition in (10a) for the singularity of
V3(a, b, c) has two other parallel forms:
(a—b)a(a—b)‘1 =(c—b)c(c—-b)_1, (10b)
(a—c)a(q—c)'1=(b—c)b(b—c)‘1, g (10c)
and therefore, (10a), (10b), (10c) are in fact equivalent, which does not seem’to be
obvious at first sight. Secondly, a necessary condition for (10a), (10b), (10c) is, that
a, b, c be pairwise conjugate. In particular, we see that, if a, b, c are not in a single
conjugacy class of D, then the Vandermonde matrix is necessarily nonsingular. If
a, b, c are indeed in one conjugacy class, then V3(a, b, c) may be singular. For
.-- 1.1.uulll
instance, over the division algebra of real quatemions, a=i, b=j, c=k are con-
1 1 1
jugate to one another; here, V3(i,j, k)= i j k is clearly singular, so we
—1 —1 —1
can predict that (10a) holds. There also exist examples (though not over the real
quaternions) where a, b, c are in one conjugacy class, but (10a) does not hold. For
such triples, V3(a, b, c) will be nonsingular. (The existence of these examples will be
clear later on.)
At this point, we could not have failed to observe a certain similarity between the
two motivating examples treated above. In the first example, we have to dis-
tinguish complex numbers by their moduli, and in the second example, we have
to distinguish elements in a division ring by the conjugacy classes they determine.
It is not difficult to see how these two cases can be unified. For any division ring
D equipped with an endomorphism a, let us say that two elements a,b are a-
conjugate if b=d"ad‘l for some deD“. (We also say that b is a o-conjugate of a.)
As is easily verified, being a-conjugate is an equivalence relation on D, so D is
partitioned into its a-conjugacy classes. If 0:], we get back the usual conjugacy
classe_s. If D=C and a=“bar”, then the a-conjugacy equation amounts to b
=a-d/d for some (1603*. A simple argument shows that this means exactly that a
and b have the same modulus.
mean that D[t,a] is the set of all left polynomials 5) Zeit‘ which are added in the
usual way, and multiplied by using the distributive law together with the rule that
t- c =c" - t for any ceD. Thus, the coefficients need not commute with the variable t.
The fact that (c1c2)‘=c‘{ c; guarantees the associative law for polynomial multipli-
cations, so D[t,a] is a ring. This so-called skew polynomial ring is a basic object
of study in noncommutative ring theory [Jacobson: 1943, Ch. 3]; the idea of the
twist equation tc=c"t goes as far back as [Hilbertz 1899]. Note that any right
polynomial Zt‘bi gives an element in D[t,a], but not every element in D[t,a] can
be expressed as a right polynomial, unless a is an automorphism of D.
As is easily seen, the usual division algorithm prevails in R=D[t,a]: For f (t)eR
and g(t)eR\{0}, we can write uniquely f (t)=q(t)g(t)+r(t) where r(t)=0 or
deg r(t)<degg(t). In particular, R is a left PID (principal ideal domain), i.e. any
(nonzero) left ideal 19R has the form R~g; here, g is any polynomial in I of the
smallest degree. It can further be shown that, in a suitable sense, R has the
Unique Factorization property, so it is a (not necessarily commutative) UFD (see
[Cohnz 1973]). However, the notion of unique factorization in the noncom-
mutative setting is a rather sophisticated one. In keeping with the elementary
nature of this article, we shall therefore try to avoid using the unique factorization
property in D[t,a]. (The expert reader will notice that in one or two places, our
proofs could be somewhat streamlined by invoking such a property.)
We shall now introduce the notion of evaluating a skew polynomial f (t)eD[t, a]
at a point deD. We write f as a left polynomial, say Zcit‘, and then define
f(d)=ZciNi(d), (11)
where N,(d)=d""'...d‘d for i 2 1, and No(d)El. Note that, in case 0= 1, then N,(d).
=di and so f (d)=Zc,d‘ is the usual evaluation off at d. In general, we must take
a into account in the evaluation process; this is why we introduced the functions
Ni. When we evaluate at d, we have to exercise caution. For instance, if f happens
to be given in the form of a right polynomial f =Ztibi, we must first express f as
a left polynomial, viz. f =b't‘, before “plugging in” (1. Thus, f (d) is given by
Zbg’iNM), not ZN,(d)b,). If f (d)=0, we say that d is a (right) root (or zero) of f.
Unless otherwise specified, the word “root” shall mean right root in this paper. If
A is a subset of D, we say that f vanishes on A if f(A)=0, i.e. if f(d)=0 for every
deA. Note that a nonzero f may vanish on an infinite set in the noncommutative
setting. For instance, in the division algebra of real quaternions, for 0:1, the
quadratic polynomial t2+l vanishes on all conjugates of i, and it is easy to see
that i has infinitely many conjugates. (This is a special case of a theorem of
[Hersteinz 1956]; for a quick proof, see MAA Monthly, Vol. 90 (1983), p. 292.)
Roots of polynomial equations over division rings have been studied by various
authors. Some of the key papers on this topic are [Richardson: 1927], [Nivenz
5) The terminology refers to the fact that the coefficients are written on the left.of the
variable.
200 T.Y. Lam
tool is the theory of roots of skew left polynomials. We shall now establish the
fundamental facts about such roots for later use. We start with two easy formulas
for the functions M.
Lemma 1. For deD, aeD“, and integers i,j 20, we have
(1) M+j(d)=1vj(d)dilvi(d);
(2) M(a°da‘1)=a”i1\l}(d)a‘i.
Proof. The right side of (1) equals
(d‘j"...d"d)‘i(d“"...d"d)=d"”’.‘ ..=d"d N,+J(d).
The left side of (2) equals
(aada‘ 1)"""(a"da‘ l)""“...(a"da- l)”(a"da" 1)=a*"‘(d“" therma-1
=a”i1\l,-(d)a‘ 1.
This completes the proof.
Ingeneral, evaluating skew polynomials at a point deD may not be a ring
homomorphism from D[t,a] to D. In other words, if f (t)=g(t)h(t) in D[t,a], f (d)
may not be equal to g(d)h(d), due to the lack of commutativity. However, we have
the following remarkable result which, to a large extent, recovers the homomor-
phism property.
Theorem 2. Let f (t)= g(t)h(t)eD[t, a] and deD. If h(d)=0 then f(d)=0. If a==h(d)
=|=O, then f (d): g(a"da'1)h(d). In particular, if d is a root off but not a root of h,
then a a-conjugate of d is a root of g.
Proof. Let g(t)= 2b. z' h(t)= 2“,- ti sofa)... Zbic,”it1+: and
f(d)= ZJNTN.+101)
=zbici‘Nj(d)"‘Ni(d) (by Lemma 1(1))
i-i .
=Zbia°‘Ni(d)
A general theory of Vandermonde matrices 201
Proof. Let r=row rank (A) and c=column rank (A). Let B be an rxn matrix
whose r rows form a basis of the (left) row space. Expressing the rows of A as left
linear combinations of those of B, we get A=B'B for a suitable m xr matrix B’.
But the equation A=B’B may also be construed as expressing the columns of A
as right linear combinations of the r columns of B', hence cSr. A similar
argument shows that rsc. Q.E.D.
The common value of rowrank (A) and column rank (A) is, of course, called the
rank of A. If A is a square nxn matrix, then A is invertible iff rankA=n. In
general, rankA is the largest integer s such that A contains an invertible
sxs submatrix. (The easy proof is left to the reader.) As a word of caution, we
should point out that the rank of A need not be equal to the dimension of the
subspace spanned by the rows in the right n-tuple vector space (D")D. In other
words, rankA may not be equal to rankAT, where AT denotes the transpose of A. In
A general theory of Vandermonde matrices 203
fact, for a,beD, A= (‘11 abb) has rank 1, but AT= (1 a) has rank2 unless ab
b ab
=ba. Thus, rankA =rank AT holds for all matrices A iff D is a field.
We are now ready to compute the rank of the a-Vandermonde matrix V
= V,,"(a1, ...,an). Let A ED be the subset {a1,...,a,,}. We have then
Theorem 8. rank V = rank A.
Proof. Let r=row rank V, c=column rank V and 6=rank A. We shall show that
r<5<c. Then Theorem 7 gives the desired result.
11— 1
The basic observation needed here is that a polynomial g(t)= go b t' (of degree
Sn- 1) vanishes on A iff (b0, bn_1) V: 0 (cf. (12)). To show thato6<c, it suffices
to find a nonzero polynomial g(t)eD[t, a] of degree Sc such that g(A)= 0. Among the
columns C1, ..., C,I of V, c of them form a basis of the column space. Without loss
of generality, we may assume that these are the first c columns. Let g(t)=2biti be
the minimal polynomial of the set {an ...,ac}. Then by Proposition 6, deggScSn,
and (b0, ...,bu_ il'C1=0 for lSiSC. Since each C]. is a right linear combination of
C1,...,Cc, we also have (bo,...,b,,_1)-Cj=0 and so (bo,...,bn_1)-V=0, i.e. g(A)
=0. Next, to show that r 36, let f = fA(t) be the minimal polynomial of A. It
suffices to show that each row Ri of V is a left linear combination of the first 6
rows R1, ...,R5. By the (left) division algorithm, we can write
=q(t)f(t)+(do+d1t+...+d,_1t""), (14)
since 6=rankA is the degree of f. Evaluating at a} (and using Theorem 2), we
have
M(aj)=do+d1N1(aj)+ u. +d‘5_lN;,_1(aJ-).
Next we shall extend the above result to the case of a rectangular a-Vandermonde
matrix. To this end, let us write V;n(a1, ...,a,) for the m x n a-Vandermonde
matrix whose (i+1)"t row is (Ni(al), ...,M(a,,)).
Lemma 9. Suppose man. Then
rank V" (a1, ...,a,l)=rank V‘(al, ,a ).
Proof. Let f(t) and 6 be as in the proof of the theorem above. The last part of the
argument in this proof 1s applicable to any row ofV,” n.Hence
row rank V3”, 5 6 = row rank V:.
The reverse inequality is clear. Q.E.D.
A curious consequence of this lemma is that, if some nxn submatrix of V‘
(man) is invertible, then V: itself must must be invertible. In the case when D is
commutative, this means that if det V": 0, then any n x n submatrix ofV,1, (man)
204 T.Y. Lam
also has determinant zero. This is obvious when a is the identity, but seems to be
far from obvious otherwise (say when D= C and a=complex conjugation).
Combining Theorem '8 and Lemma 9, we have the following result, valid for any
pair (m, n).
Theorem 10. rank V;,,,(a1, . . . , a”) = min {m, rank A}
Proof. Let 6=rank A sn. We may assume that the first 6 columns of V,“ are right
independent. Hence a5(a1, ...,a,,) has rank 6. By Lemma 9, Vggal, ...,a,,) also has
rank 6. This implies that the 6 rows of 1453011, ...,a,,) are left independent, and so
the first 6 rows of K,°(a1,...,a,,) are also left independent. If msé, then clearly
rank V;,,,(a1,...,a,.)=m=min{m,6}. Next assume m_>_6. If in fact man, we are
done by Lemma 9. If msn, then rank a'nsrank V,,"=6. But the first 5 rows of
V5,, are left independent so rank V,,‘,’,,,=6=min{m,6}. Q.E.D. '
Of course, it also makes sense to talk about the infinite a-Vandermonde matrix
V;,n(a1, ...,a,,) whose i'h column is the infinite tuple (1,N1(ai),N2(a,.), ...)T. The row
rank and the column rank of V3,," are defined as before, and the proof of
Theorem 7 carries over verbatim to show that they are equal. By Theorem 10 (or
_ its proof), we see easily that rank V;,,,=rank A.
Finally, we note the following remarkable result:
Proposition 11. Let C, (resp. C?) be the i'h column of V,f(a1, ...,a,,) (resp.
V;'n(a1, ...,a,,)), and let b1, ...,bneD. IfZCibi=0, then Ecrb,=o.
Proof. This follows since, as we saw before, any row of V1,... is a left linear
combination of the n rows of V,,". The details will be left to the reader. Q.E.D.
Theorem 16. Let A ED be a a—algebraic set. Let {b,, ...,b,,,} be a P-independent set
in A, and {(11, ...,d,,} be a P-spanning set for A. Then, after a reindexing of the di’s,
{b1, ...,b,,,, dm+1,...,d,,} is a P-spanning set for A.
Proof. Assume, inductively, that {b1, ...,b,,d,+1, ...,d,,} is a P-spanning set for A,
where r<m. Suppose br H is P-dependent on {b1,...,b,, dil,...,di'}, where s is
chosen minimal. (Note that s21 by the P-independence of the bi’s.) By the
Exchange Lemma, d,l is P-dependent on {b1,...,b,, b,+1, di2,...,d,..}. Thus, the
latter set is a P-spanning set for A, and the proof proceeds by induction. Q.E.D.
Of course, the remarkable parallel of the theory of P—independence and the usual
theory of linear independence in vector spaces is not an accident. The exact
connection between these two theories can be seen by using a-Vandermonde
matrices, which make possible the following interpretation of P-dependence in
terms of linear dependence in linear algebra.
Proposition 17. For deD and any a-algebraic set A ED, the following statements are
equivalent:
(1) d is P-dependent on A;
(2) (1,N1(d),N2(d), ...)T is right linearly dependent on {(1,N1(a),Nz(a), ...)T: aeA};
(3) There exist a1,...,a,,eA such that (1,Nl(d),...,N _1(d))T is a right linear com-
bination of {(1,N1(ai), ...,N,,,_ l(ai))T: lsisn} for some m2n+ 1.
Proof. Note that d is P-dependent on A iff d is P-dependent on some finite subset
of A, iff rank({d}uA)=rank A. Using the Vandermonde matrix interpretation of
these ranks (Theorem 10), we get the equivalence of (1) and (3). Finally, the
equivalence of (2) and (3) follows from Proposition 11. Q.E.D.
The Proposition above unifies the theories of P-independence and linear inde-
pendence, and puts our earlier results on P-independence in perspective. Of
course, it would have been possible to deduce all these results from the corre-
sponding facts in linear algebra, by using Proposition 17. Nevertheless, the ad hoc
treatment of the theory of P-independence and P-bases given earlier in this
section seems to be of interest in its own right.
With the theory developed so far, we are able to write down effectively the
condition for a a-Vandermonde matrix V:(a1,...,a_) to be singular. In fact, we
can write down effectively the condition for V:(a1,...,a,,) to have rank equal to
any prescribed number r (1 srsn). By Theorem 8, this amounts to A ={a1, ...,a,,}
having rank r, which means that there exist r distinct elements ah,...,aireA
forming a P-basis of A, i.e. they are P-independent, and each a, is a root of the
minimal polynomial f of {a,1,...,air}. Since f can be computed effectively, we
know how to express the condition that V:(a1,...,a,,) has rank r by a finite
number of equations in D. For instance, if n=3, the condition for V3"(a, b, c) to be
singular is that either two or more of a, b, c are equal, or all three are distinct,
and c satisfies the minimal polynomial (t—(b—a)"b(b—a)'1) (t—a) of {a,b}, i.e.
A general theory of Vandermonde matrices 207
Corollary 24. rank V2 E1 min{2, IA |} The matrix V is invertible iff each square
submatrix V is invertible If no three of the elements of A are o-conjugate, then V'is
invertible. If V is invertible, and A: {d1,... ,dn}, then for any given elements
b1,...,b,,eD, there exists a unique polynomial f(t)eD[t, a] of degreeSn—l such
that f (d,)=b, for all i.
(For the last statement, recall the interpolation formula (12).)
We shall mention two interesting cases of Corollary 24 in which we get a direct
and complete computation of the rank of the a—Vandermonde matrix. The general
notation above will be fixed.
Application (A). For any generalized quaternion division algebra D over a field F of
characteristic =|=2 (cf. (16)), let 0= 1. Then rank V= 2 min{2,|A,|}.
l-l
Application (B). For D=C and a=“bar”, let 61,...,6,,, be the different values
among |d|: deA. Then rank V= Z r, where ri=1 if exactly one element in A has
modulus 6,- and ri=2 otherwise. "1
A general theory of Vandermonde matrices 211
c==(1+x)a(1+x)'l
=(1+x)a(l —x+x2)/3 (since x3=2)
=(a+bx)(l —x+x2)/3 (22)
= ((a + 2b) + (b —a)x + (a —b)x2)/3.
Noting that f (t) is the minimal polynomial f4 for A={a,b}, the fact that f (c)=0
implies that c is P-dependent on A. Thus, V= V3(a, b, c) has rank 2.
Of course, the word “P-dependent” above should have been written as “left P-
dependent”, since it is checked by using left polynomials. We have a parallel
notion of right P-dependence by using, instead, right polynomials. We claim that
the 3-element set {a, b, c} is right P-independent. If this is the case, then applying
our results to right polynomials, it follows that VT is invertible, i.e. rank VT=3.
Assume, instead, that {a,b,c} is not right P-independent. Then there exists a
monic quadratic right polynomial g(t) vanishing on {a,b,c} as left roots. Since a,b
commute, g(t) must be just
(t —a)(t —b)=t2 -t(a+b)+ab.
Thus, we have c2 —c(a + b) +ab=0 as well as c2 —(a+b)c +ab =0. This implies
that a+b commutes with c. Since K - x‘=x‘- K and D=K$K~x$K-x2, we see
from (22) that a +b commutes with all three “components” of c in (22). In
particular (b —a)x(a + b) = (a + b)(b —a)x, or (b —a)(b + b2 —2)x = (a + a2 —2)(b —a)x.
A general theory ‘of Vandermonde matrices 213
Epilogue
Just as this article was being typed up for publication, I received, with perfect
timing, the doctoral disseration of André Leroy at Université de l’Etat a Mons,
entitled “Derivations Algébriques.” In Chapter 3 of this dissertation, Leroy also
considered Vandermonde matrices over division rings D through the connection
with Wronskian matrices, and observed the fact that V,,"(a1, ...,a,,) is invertible iff
“the left LCM of t—a1,...,t—a,, in D[t,a] is of degree n.” In a letter to me dated
March 18, 1985, Leroy further made the following important observations.
For a division ring D, let a be a fixed endomorphism and let 6: D—vD be a a-
derivation, i.e. an additive homomorphism such that 6(ab)=6(a)b+a"6(b) for all
a,beD. One can form the Ore skew polynomial ring D[t, 0,6] whose elements are
left polynomials Zait‘, multiplied according to the rule ta=a"t+6(a), for any
aeD. (For 6:0, one gets back the D[t,a] used in the text.) For aeD, redefine
Nn(a) by taking No(a)=1 and inductively M+1(a)=l\ln(a)‘a+6(Nn(a)). Then define
the (a,6)-Vandermonde matrix K,""(al, ...,a,,) as in (3). For f (t)=2ait‘eD[t, a, 6],
define f (a)=2aiN,-(a) as before, and, for c,deD, define c to be a (a,6)-conjugate of
d if c=a"da‘1 +6(a)a'l for some aeD*. Using the techniques in his dissertation,
Leroy has proved the following generalization of our Theorem 2:
Theorem (Leroy). Let f (t)=g(t)h(t)eD[t, a, 6] and deD. If h(d)=0 then f (d)=0. If
a==h(d)=|=0, then f(d)=g(a"da'l +6(a)a'1)h(d).
In view of this new form of Theorem 2, Leroy observed (and I happily agree) that
all of the main results in this paper carry over, with the same proofs, to the more
general case of (a, 5)-Vandermonde matrices and (a,6)-polynomial dependence.
One nice consequence of this observation is the following. Given (D, a, 6), and
x1, ...,x,,eD*, let W(x1, ...,x,,) be the following Wronskian matrix
x1 x2 x,l
6(x,) 6(x2) 6(xn)
............................................. (23)
.............................................
Note added in Proof. In a footnote to (51, the author conjectured that, over the complex
numbers with a =complex conjugation, one has the determinantal inequality
Idet V,,"(a1, ..., an)|2|det V,,(|a1|, ..., |a,,|)|. This has recently been proved to be true by David
Richman et al. at the University of South Carolina.
References
A. Aitken, Determinants and Matrices, Interscience Publishers, New York. Reprinted 1959.
E. Artin, Geometric Algebra, Interscience Publishers, New York, 1957.
B. Beck, Sur les equations polynomiales dans les quatemions, L'Enseignement Math. 25
(1979), 195—201.
L. Brand, The roots of a quaternion. MAA Monthly 49 (1942), 519—520.
U. Bray and G. Whaples, Polynomials with coefficients from a division ring, Canad. J.
Math. 35 (1983), 509—515.
L. Carlitz, A special determinant, MAA Monthly 62 (1955), 242—243.
C. Cerlienco and M. Mureddu, A note on polynomial equations in quatemions, Rend. Sem.
Fae. Sci. Univ. Cagliari 51 (1981), 95—99.
P.M. Cohn, Unique factorization domains, MAA Monthly 80 (1973), l—18. .
P.M. Cohn, Skew Field Constructions, London Math. Society Lecture Notes Series, Vol. 27,
Cambridge University Press, 1977.
J. Dauns, A Concrete Approach to Division Rings, Heldermann Verlag, Berlin, 1982.
LE. Dickson, Algebras and their Arithmetics, 1923. (Reprinted by Dover Publications,
1960).
H. Edwards, Galois Theory, Graduate Text in Mathematics, Vol. 101, Springer-Verlag,
Berlin, Heidelberg, New York, 1984.
S. Eilenberg and I. Niven, The “Fundamental Theorem of Algebra” for quatemions, Bull.
Amer. Math. Soc. 50 (1944), 246—248.
A. Farnell, A special Vandermondian determinant, MAA Monthly 66 (1959), 564—569.
B. Fein and M. Schacher, Solutions of pure equations in rational division algebras. I., 'J.
Algebra 17 (1971), 83~93.
B. Gordon and T. Motzkin, On the zeros of polynomials over division rings, Trans. Amer.
Math. Soc. 116 (1965), 218-226. (Correction ibid. 122 (1966), 547.)
T. Grundhofer and H. Liineburg, Uber die Summation von Reihen aus linearen Schiebe-
registerfolgen, Archiv der Mathematik 34 (1980), 428—439.
I. Herstein, Conjugates in division rings, Proc. Amer. Math. Soc. 7 (1956), 1021—1022.
I. Herstein, Non-commutative Rings, Carus Monograph, Vol.15, Math. Assoc. America,
1968. '
A general theory of Vandermonde matrices 215
D. Hilbert, Grundlagen der Geometric, 1899. Leipzig and Berlin, Teubner, (9th ed., 1962,
Stuttgart)
K.A. Hirsch, A note on the Vandermonde determinant, J. London Math. Soc. 24 (1949),
144—145.
N. Jacobson, The Theory of Rings, Amer. Math. 800., Providence, R.I., 1943.
D. Kalman, The generalized Vandermonde matrix, Math. Magazine 57 (1984), 15—21.
T.Y. Lam, The Algebraic Theory of Quadratic Forms, Benjamin, Addison-Wesley, 1973.
(Second Printing with revisions, 1980.)
T.Y. Lam, Polynomial dependence and polynomial equations in division rings, to appear,
1985. .
D.E. Littlewood, The Theory of Group Characters, 2nd edition, Oxford University Press,
1950.
LG. MacDonald, Symmetric Functions and Hall Polynomials, Clarendon Press, Oxford,
1979.
T. Muir, A Theatise on the Theory of Determinants, 1933. (Reprinted by Dover Publications,
1960.)
F. Murnaghan, The Theory of Group Representations, 1938. (Reprinted by Dover Publi-
cations, 1963.)
I. Niven, Equations in quarternions, MAA Monthly 48 (1941), 654—661.
I. Niven, The roots of a quaternion, MAA Monthly 49 (1942), 386—388.
V. Norton, A generalized Vandermonde determinant (Problem E2942), solution in MAA
Monthly (1984), 586—587.
B. Pollak, The equation tat=b in a quaternion algebra, Duke Math. J. 27 (1960), 261—271.
A.R. Richardson, Equations over a division algebra, Messenger of Math. 57 (1927), 1-6.
H. Rohrl, On the zeros of polynomials over arbitrary finite dimensional algebras, Manu-
scripta Math. 25 (1978), 359—390.
K.C. Smith, Theory and Applications of Noncommutative Polynomials, University of Okla-
homa Lecture Notes, 1974.
R. Stanley, Theory and application of plane partitions. 1., Studies in Applied Math. 50
(1971), 167—188.
Received 21.5.85
Department of Mathematics
University of California
Berkeley CA 94720