Commutative Algebra/The Cayley–Hamilton theorem and Nakayama's lemma

From testwiki
Revision as of 21:55, 1 June 2016 by imported>JackBot (The theorems: using AWB)
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)
Jump to navigation Jump to search

Determinants within a commutative ring

We shall now derive the notion of a determinant in the setting of a commutative ring.

Template:TextBox

We shall later see that there exists exactly one determinant.

Template:TextBox

Proofs:

1. Let A=(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n), where the j-th column 𝐚j is the zero vector. Then by axiom 3 for the determinant setting c=1,

det(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n)=det(𝐚1,,𝐚j1,𝐚j𝐚j,𝐚j+1,,𝐚n)=det(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n)det(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n)=0.

Alternatively, we may also set c=1 and 𝐛j=𝐚j=𝟎 to obtain

det(𝐚1,,𝐚j1,𝐚j+c𝐛j,𝐚j+1,,𝐚n)=(1+c)det(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n),

from which the theorem follows by subtracting detA from both sides.

Those proofs correspond to the proofs for T0=0 for a linear map T (in whatever context).

2. If we set 𝐛j=𝐚j+1 or 𝐛j=𝐚j1 (dependent on whether we add the column left or the column right to the current column), then axiom 3 gives us

det(𝐚1,,𝐚j1,𝐚j+c𝐛j,𝐚j+1,,𝐚n)=det(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n)+cdet(𝐚1,,𝐚j1,𝐛j,𝐚j+1,,𝐚n),

where the latter determinant is zero because we have to adjacent equal columns.

3. Consider the two matrices A:=(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n) and B:=(𝐚1,,𝐚j1,𝐚j+1,𝐚j,,𝐚n). By 7.2, 2. and axiom 3 for determinants, we have

detB=det(𝐚1,,𝐚j1,𝐚j+1+𝐚j,𝐚j,,𝐚n)=det(𝐚1,,𝐚j1,𝐚j+1+𝐚j,𝐚j+1,,𝐚n)=det(𝐚1,,𝐚j1,𝐚j,𝐚j+1,,𝐚n)=detA.

4. We exchange the j-th and k-th column by first moving the j-th column successively to spot k (using |jk| swaps) and the k-th column, which is now one step closer to the j-th spot, to spot j using |jk|1 swaps. In total, we used an odd number of swaps, and all the other columns are in the same place since they moved once to the right and once to the left. Hence, 4. follows from applying 3. to each swap.

5. Let's say we want to add c𝐚k to the j-th column. Then we first use 4. to put the j-th column adjacent to 𝐚k, then use 2. to do the addition without change to the determinant, and then use 4. again to put the j-th column back to its place. In total, the only change our determinant has suffered was twice multiplication by 1, which cancels even in a general ring.

6. Let's say that the j-th column and the k-th column are equal, kj. Then we subtract column j from column k (or, indeed, the other way round) without change to the determinant, obtain a matrix with a zero column and apply 1.

7. Split σ into swaps, use 4. repeatedly and use further that sgn is a group homomorphism.

Note that we have only used axioms 2 & 3 for the preceding proof.

The following lemma will allow us to prove the uniqueness of the determinant, and also the formula det(AB)=detAdetB.

Lemma 7.3:

Let A=(ai,j)1i,jn and B=(bi,j)1i,jn be two n×n matrices with entries in a commutative ring R. Then

det(AB)=detAσSnsgn(σ)b1,σ(1)bn,σ(n).

Proof:

The matrix AB has k-th columns ν=1nbν,k𝐚ν. Hence, by axiom 3 for determinants and theorem 7.2, 7. and 6., we obtain, denoting AB=:C=(ci,j)1i,jn=(𝐜1,,𝐜n):

det(AB)=ν1=1nbν1,1det(𝐚ν1,𝐜2,,𝐜n)=ν1=1nν2nbν1,1bν2,2det(𝐚ν1,𝐚ν2,𝐜3,,𝐜n)==ν1,,νn=1nbν1,1bνn,ndet(𝐚ν1,,𝐚νn)=detAσSnsgn(σ)b1,σ(1)bn,σ(n)

Template:TextBox

Proof:

Let CRn×n be an arbitrary matrix, and set A=In and B=C in lemma 7.3. Then we obtain by axiom 1 for determinants (the first time we use that axiom)

detC=det(InC)=1σSnsgn(σ)c1,σ(1)cn,σ(n).

Template:TextBox

Proof:

From lemma 7.3 and theorem 7.4 we may infer

det(AB)=det(A)σSnsgn(σ)b1,σ(1)bn,σ(n)=det(A)det(B).

Template:TextBox

Proof:

First of all, In has nonzero entries everywhere except on the diagonal. Hence, if In=(ai,j)1i,jn, then a1,σ(1)an,σ(n) vanishes except σ(1)=1,,σ(n)=n, i.e. σ is the identity. Hence det(In)=1.

Let now A be a matrix whose j-th and j+1-th columns are equal. The function

f:SnSn,f(σ)=k{σ(k)k{j,j+1}σ(j)k=j+1σ(j+1)k=j

is bijective, since the inverse is given by f itself. Furthermore, since f amounts to composing σ with another swap, it is sign reversing. Hence, we have

det(A)=σSnsgn(σ)a1,σ(1)an,σ(n)=sgnσ=1a1,σ(1)an,σ(n)sgnσ=1a1,σ(1)an,σ(n)=sgnσ=1a1,σ(1)an,σ(n)sgnσ=1a1,f(σ)(1)an,f(σ)(n).

Now since the j-th and j+1-th column of A are identical, k,l:ak,σ(l)=ak,f(σ)(l). Hence detA=0.

Linearity follows from the linearity of each summand:

σSnsgn(σ)a1,σ(1)(aσ1(j),j+cbσ1(j),j)an,σ(n)=σSnsgn(σ)a1,σ(1)aσ1(j),jan,σ(n)+cσSnsgn(σ)a1,σ(1)bσ1(j),jan,σ(n).

Template:TextBox

Proof:

Observe that inversion is a bijection on Sn the inverse of which is given by inversion ((σ1)1=σ). Further observe that sgn(σ)=sgn(σ1), since we just apply all the transpositions in reverse order. Hence,

detA=σSnsgn(σ)a1,σ(1)an,σ(n)=σ1Snsgn(σ1)a1,σ1(1)an,σ1(n)=σSnsgn(σ)aσ(1),1aσ(n),n=detAt.

Template:TextBox

Proof 1:

We prove the theorem from the formula for the determinant given by theorems 7.5 and 7.6.

Let k{1,,n} be fixed. For each ν{1,,n}, we define

f:Sn1Sn,f(σ):=m{νm=νσ(m)m<νσ(m)<νσ(m)+1m<νσ(m)νσ(m1)m>νσ(m)<νσ(m1)+1m>νσ(m)ν.

Then

ν=1naν,kdetAν,k=ν=1n(1)ν+kaν,kσSn1sgn(σ)a1,f(σ)(1)ak1,f(σ)(k1)ak+1,f(σ)(k+1)an,f(σ)(n)=σSnsgn(σ)a1,σ(1)an,σ(n).

Proof 2:

We note that all of the above derivations could have been done with rows instead of columns (which amounts to nothing more than exchanging ai,j with aj,i each time), and would have ended up with the same formula for the determinant since

σSnsgn(σ)a1,σ(1)an,σ(n)=σ1Snsgn(σ1)a1,σ1(1)an,σ1(n)=σSnsgn(σ)aσ(1),1aσ(n),n

as argued in theorem 7.7.

Hence, we prove that the function Rn×nR given by the formula ν=1n(1)ν+kaν,kdetAν,k satisfies 1 - 3 of 7.1 with rows instead of columns, and then apply theorem 7.4 with rows instead of columns.

1.

Set A=In to obtain

ν=1naν,k(1)ν+kdetAν,k=(1)2kak,kdetAk,k=11=1.

2.

Let A have two equal adjacent rows, the j-th and j+1-th, say. Then

ν=1naν,k(1)ν+kdetAν,k=(1)j+kdetAj,k+(1)j+1+kdetAj+1,k=0,

since each of the Aν,k has two equal adjacent rows except for possibly ν=j and ν=j+1, which is why, by theorem 7.6, the determinant is zero in all those cases, and further Aj,k=Aj+1,k, since in both we deleted "the same" row.

3.

Define B:=(bi,j)1i,jn:=(𝐚1,,𝐚j1,𝐚j+c𝐛j𝐚j+1,,𝐚n)t, and for each ν,k{1,,n} define Cν,k as the matrix obtained by crossing out the ν-th row and the k-th column from the matrix C:=(ci,j)1i,jn:=(𝐚1,,𝐚j1,𝐛j𝐚j+1,,𝐚n)t. Then by theorem 7.6 and axiom 3 for the determinant,

ν=1nbν,k(1)ν+kdetBν,k=ν=1j1aν,k(1)ν+k(detAν,k+cdetCν,k)+(1)j+k(aj,k+cbj,k)detAj,k+ν=j+1naν,k(1)ν+k(detAν,k+cdetCν,k)=ν=1naν,k(1)ν+kdetAν,k+cν=1ncν,k(1)ν+kdetCν,k=detA+cdetC.

Hence follows linearity by rows.

For the sake of completeness, we also note the following lemma:

Lemma 7.9:

Let A be an invertible matrix. Then det(A) is invertible.

Proof:

Indeed, det(A)1=det(A1) due to the multiplicativity of the determinant.

The converse is also true and will be proven in the next subsection.

Exercises

  • Exercise 7.1.1: Argue that the determinant, seen as a map from the set of all matrices (where scalars are 1×1-matrices), is idempotent.

Cramer's rule in the general case

Template:TextBox

Proof 1:

Let j{1,,n} be arbitrary but fixed. The determinant of A is linear in the first column, and hence constitutes a linear map in the first column Lj:RnR mapping any vector to the determinant of A with the j-th column replaced by that vector. If 𝐚j is the j-th column of A, Lj(𝐚j)=det(A). Furthermore, if we insert a different column 𝐚k into Lj, we obtain zero, since we obtain the determinant of a matrix where the column 𝐚k appears twice. We now consider the system of equations

{a1,1x1++a1,nxn=b1an,1x1++an,nxn=bn,

where (x1,,xn)T is the unique solution of the system Ax=b, which exists since it is given by A1b since A is invertible. Since Lj is linear, we find an 1×n matrix (c1,,cn) such that for all 𝐯Rn

(c1,,cn)𝐯=Lj(𝐯);

in fact, due to theorem 7.8, ck=(1)j+kdet(Aj,k). We now add up the lines of the linear equation system above in the following way: We take c1 times the first row, add c2 times the second row and so on. Due to our considerations, this yields the result

det(A)xj=Lj(𝐛).

Due to lemma 7.9, det(A) is invertible. Hence, we get

xj=(det(A))1Lj(𝐛)=(det(A))1det(Aj)

and hence the theorem.

Proof 2:

For all j{1,,n}, we define the matrix

Xj:=(100x100010x20000xn01000xn001);

this matrix shall represent a unit matrix, where the j-th column is replaced by the vector (x1,,xn)𝐓. By expanding the j-th column, we find that the determinant of this matrix is given by det(Xj)=xj.

We now note that if A=(𝐚1,,𝐚n), then Xj=A1(𝐚1,,𝐚j1,A𝐛,𝐚j+1,,𝐚n)=A1Aj. Hence

xj=det(A1Aj)=det(A1)det(Aj)=det(A)1det(Aj),

where the last equality follows as in lemma 7.9.

Template:TextBox

Proof:

For j{1,,n}, we set 𝐛j:=ej=(0,,0,1,0,,0)T, where the zero is at the j-th place. Further, we set Lj to be the linear function from proof 1 of theorem 7.10, and Mj its matrix. Then adj(A) is given by

adj(A)=(M1Mn)

due to theorem 7.8. Hence,

adj(A)A=(M1AMnA)=(det(A)000det(A)00det(A)000det(A)),

where we used the properties of Lj established in proof 1 of theorem 7.10.

The theorems

Now we may finally apply the machinery we have set up to prove the following two fundamental theorems.

Template:TextBox

Note that the polynomial in ϕ is monic, that is, the leading coefficient is 1, the unit of the ring in question.

Proof: Assume that {m1,,mn} is a generating set for M. Since ϕ(M)IM, we may write

ϕ(mj)=k=1nbj,kxk,j{1,,n} (*),

where bkI for each k. We now define a new commutative ring as follows:

R~:={ϕk|k}R,

where we regard each element r of R as the endomorphism mrm on M. That is, R~ is a subring of the endomorphism ring of M (that is, multiplication is given by composition). Since ϕ is R-linear, R~ is commutative.

Now to every n×n matrix A with entries in R~ we may associate a function

A():MnMn,A((x1,,xn)T):=(k=1na1,k(x1),,k=1na1,k(x1)).

By exploiting the linearities of all functions involved, it is easy to see that for another n×n matrix with entries in R~ called B, the associated function of AB equals the composition of the associated functions of A and B; that is, (AB)(x)=A(B(x)).

Now with this in mind, we may rewrite the system (*) as follows:

A(x)=0,

where A has j,k-th entry δj,kϕbj,kR~. Now define B:=adj(A). From Cramer's rule (theorem 7.11) we obtain that

BA=Indet(A),

which is why

(detAx1,,detAxn)t=(BA)(x)=B(A(x))=B(0)=𝟎, the zero vector.

Hence, detAR~ is the zero mapping, since it sends all generators to zero. Now further, as can be seen e.g. from the representation given in theorem 7.4, it has the form

ϕn+an1ϕn1++a1ϕ+a0

for suitable an1,,a0I.

Template:TextBox

Proof:

Choose ϕ=IdM in theorem 7.12 to obtain for mM that

ϕn(m)+an1ϕn1(m)++a1ϕ(m)+a0m=(1+an1++a0)m=0

for suitable an1,,a0I, since the identity is idempotent.

Template:BookCat