Escolar Documentos
Profissional Documentos
Cultura Documentos
TSOGTGEREL GANTUMUR
Abstract. We review some concepts from linear algebra over R.
Contents
1.
2.
3.
4.
5.
6.
1
3
7
8
10
12
Rn ,
(1)
k = 1, 2, . . . , n, are defined as
e1 = (1, 0, 0, . . . , 0, 0),
e2 = (0, 1, 0, . . . , 0, 0),
... ...
(2)
en = (0, 0, 0, . . . , 0, 1).
In other words, the i-th component of the vector ek is given by
{
1 for i = k,
(ek )i = ik
0 otherwise.
(3)
(4)
for any scalars , R and vectors x, y Rn . In this context, Rn and Rm are called the
domain and the codomain of F .
Example 1.1. The identity map id : Rn Rn defined by id(x) = x, and the map F : R2 R
defined by F (x1 , x2 ) = x1 3x2 are linear.
Date: 2016/12/15 17:11 (EDT). Incomplete draft.
1
TSOGTGEREL GANTUMUR
(5)
Rm
ai,k xk .
(7)
k=1
(8)
and the whole collection A is called an m n matrix. When there is no risk of confusion, we
simply write a11 , aik , etc., instead of a1,1 and ai,k . Thus, any m n matrix generates a linear
map F : Rn Rm , and any linear map F : Rn Rm corresponds to an m n matrix. In
other words, denoting the space of m n matrices by Rmn , and the set of all linear maps
between Rn and Rm by L(Rn , Rm ), there is an identification L(Rn , Rm ) = Rmn .
Example 1.2. (a) For the identity map id : Rn Rn , we have id(ek ) = ek , k = 1, . . . , n,
and hence the matrix entries are [id(ek )]i = (ek )i = ik . This matrix
1
0 ... 0
0
1 ... 0
I = Inn =
(9)
. . . . . . . . . . . . ,
0
0 ... 1
is called the n n identity matrix.
(b) For the zero map F : Rn Rm defined by F (x) = 0 for all x Rn , the corresponding
matrix will obviously be the m n matrix whose entries are all 0. This matrix is called
the m n zero matrix, and denoted by simply 0, or by 0mn .
(c) For F : R2 R2 given by F (x1 , x2 ) = (x1 + 3x2 , x1 ), the corresponding matrix is
(
)
1 3
A=
.
(10)
0 1
(d) For F : R R2 given by F (t) = (2t, 5t), the corresponding matrix is
( )
2
A=
.
5
(11)
For A, C Rmn and R, the operations A C and A are defined in the same way
as for vectors, by identifying Rmn with Rmn . In addition, a multiplicative operation for
matrices arises in connection with composition of linear maps, as follows. Let F : Rn Rm
and G : R Rn be linear maps. Then the composition F G : R Rm defined by
(F G)(x) = F (G(x)) is linear, because
(
)
(
)
(
)
(
)
F G(x + y) = F G(x) + G(y) = F G(x) + F G(y) .
(12)
LINEAR ALGEBRA
Suppose that A = (aij ) Rmn and B = (bjk ) Rn are the matrices corresponding to F
and G, respectively. Then we have
(
)
F G(ek ) = F (b1k e1 + . . . + bnk en ) = b1k F (e1 ) + . . . + bnk F (en ),
(13)
and hence the (i, k)-th entry of the matrix corresponding to F G is
(
)
[F G(ek ) ]i = b1k [F (e1 )]i + . . . + bnk [F (en )]i = ai1 b1k + . . . + ain bnk .
(14)
In view of this, we define the product of A and B as the matrix AB Rm , whose entries
are given by
n
[AB]ik =
aij bjk = ai1 b1k + . . . + ain bnk .
(15)
j=1
Exercise 1.1. Prove that matrix multiplication is associative ((AB)C = A(BC)) and distributive (A(B + D) = AB + AD), but not commutative in general.
Recall that the components of F (x) for x Rn is given by (7). In light of the definition of
matrix multiplication, (7) tells us that F (x) is simply the product of the matrices A and x, if
we consider x as an n 1 matrix, i.e., as a column vector. Thus we define the (matrix-vector)
product between a matrix A Rmn and a vector x Rn as the vector Ax Rm , whose
components are given by
[Ax]i =
(16)
k=1
(17)
This set is also called the image of A or of F , or the column space of A. Furthermore, we
define the kernel of F or the kernel of A as
ker(F ) = ker(A) = {x Rn : Ax = 0} Rn ,
(18)
Other names for this set include the null space of A, and the zero set of F . The range
and the kernel of F contain important structural information on the map F . This can be
gleaned by considering the equation F (x) = b, where b Rm is given. It is clear that
F (x) = b has a solution if and only if b ran(F ). Moreover, if F (x) = F (y) = b, then
F (x y) = F (x) F (y) = 0, and hence x y ker(F ). On the other hand, if F (x) = b and
h ker F , then F (x + h) = F (x) + F (h) = b. This means that the extent of non-uniqueness
of the solutions of F (x) = b is precisely measured by the kernel of F . In particular, F (x) = b
has a unique solution if and only if b ran(F ) and ker(F ) = {0}.
Let x, y ker(A) and , R. Then we have
A(x + y) = Ax + Ay = 0,
(19)
TSOGTGEREL GANTUMUR
that is, x + y ker(A). Let , ran(A), that is, = Ax and = Ay for some x, y Rn .
Then for , R, we have
+ = Ax + Ay = A(x + y),
(20)
that is, + ran(A). The kernel and the range of A are examples of linear spaces.
Definition 2.1. A nonempty subset V Rn is called a linear space (or a linear subspace of
Rn ) if x + y V for all x, y V and , R.
Let us denote the columns of A by A1 , . . . , An Rm . Then for x Rn , the matrix-vector
product Ax can be written as
Ax = x1 A1 + x2 A2 + . . . + xn An .
(21)
In general, the right hand side is called the linear combination of the vectors A1 , . . . , An Rm ,
with the coecients given by x1 , . . . , xn R. Thus the range of A,
ran(A) = {x1 A1 + . . . + xn An : x Rn },
(22)
is simply the collection of all possible linear combinations of the columns of A, which explains
the alternative name column space. We say that the space ran(A) is generated, or spanned
by the collection {A1 , . . . , An }, or simply that ran(A) is equal to the span of {A1 , . . . , An },
with the latter defined by
span{A1 , . . . , An } = {x1 A1 + . . . + xn An : x Rn }.
(23)
The first step towards understanding ran(A) is to try to come up with an ecient description
of it, by removing redundant vectors from the collection {A1 , . . . , An }. In other words, we
want to have a minimal subcollection {V1 , . . . , Vk } {A1 , . . . , An }, with the property that
span{V1 , . . . , Vk } = span{A1 , . . . , An }.
(24)
The vector An can be removed from the collection {A1 , . . . , An } without changing its span if
and only if An can be written as a linear combination of the remaining vectors {A1 , . . . , An1 },
that is, if and only if there exist numbers x1 , . . . , xn1 R such that
An = x1 A1 + . . . + xn1 An1 .
(25)
We see that some vector Ai can be removed from the collection {A1 , . . . , An } without changing
its span if and only if there exist numbers x1 , . . . , xn R, not all equal to 0, such that
x1 A1 + . . . + xn An = 0.
(26)
In terms of A, the latter condition is the same as ker(A) = {0}. This also means that we
cannot remove any vector from the collection {V1 , . . . , Vk } without changing its span if and
only if the only way for the equality
x1 V1 + . . . + xk Vk = 0,
(27)
implies
x1 = . . . = xk = 0.
(28)
LINEAR ALGEBRA
(29)
(30)
for , Rk , then = . The reason is that we can write the preceding equality as
(1 1 )V1 + . . . + (k k )Vk = 0,
(31)
x x = 0 by convention, so the empty set is a basis of the trivial vector space {0}.
(b) The set {e1 , e2 , . . . , en } Rn is a basis of Rn , and called the standard basis of Rn .
Example 2.5. Consider the matrix
1 0 1
A = 0 1 1 ,
2 5 7
(32)
so that A1 = (1, 0, 2), A2 = (0, 1, 5), and A3 = (1, 1, 7). Since A1 + A2 A3 = 0, we can
express any of A1 , A2 and A3 in terms of the other two.
(a) Let us remove A3 , and consider the resulting collection {A1 , A2 }. This is linearly independent, because the first two components of the equation A1 + A2 = 0 already imply
= = 0. Thus we have
ran(A) = span{A1 , A2 },
(33)
and the collection {A1 , A2 } cannot be reduced any further, meaning that {A1 , A2 } is a
basis of ran(A).
(b) Let us remove A1 , and consider the resulting collection {A2 , A3 }. This is linearly independent, because the first two components of the equation A2 + A3 = 0 give = 0 and
+ = 0, yielding = = 0. Thus {A2 , A3 } is a basis of ran(A).
Suppose that we started with the collection {A1 , . . . , An }, and by removing redundant
vectors one after another, we ended up with the subcollection {V1 , . . . , Vk } that is a basis of
ran(A). We have seen in Example 2.5 that dierent choices in the intermediate steps can result
in a dierent subcollection, say, {W1 , . . . , Wr } {A1 , . . . , An }. It is also conceivable that a
basis {W1 , . . . , Wr } of ran(A) that is not necessarily a subcollection of {A1 , . . . , An } can be
obtained by some other means. In any case, we prove below that the number of elements in a
basis is a quantity that depends only on the linear space itself, rather than on what particular
vectors we have in the basis. Thus in our situation, it is necessarily true that r = k.
We first prove the following fundamental result.
Theorem 2.6 (Steinitz exchange lemma). Let {V1 , . . . , Vk } Rn be a linearly independent
set, and let {W1 , . . . , Wr } Rn be such that V1 , . . . , Vk span{W1 , . . . , Wr }. Then k r,
and after a possible reordering of the set {W1 , . . . , Wr }, we have
span{W1 , . . . , Wr } = span{V1 , . . . , Vk , Wk+1 , . . . , Wr }.
(34)
(35)
TSOGTGEREL GANTUMUR
1
1 V 1
2
1 W 2
...
r
1 Wr ,
(36)
(37)
1
2 V2
3
2 W3
...
r
2 Wr ,
(38)
and hence Y = span{V1 , V2 , W3 , . . . , Wr }. We can repeat this process, and end up with the
conclusion Y = span{V1 , . . . , Vk , Wk+1 , . . . , Wr }, which also shows that k r.
By applying the preceding theorem with {W1 , . . . , Wr } given by the standard basis {e1 , . . . , en }
of Rn , we get the following important corollary.
Corollary 2.7 (Basis completion). Let {V1 , . . . , Vk } Rn be a linearly independent set. Then
there exist vectors Vk+1 , . . . , Vn Rn such that span{V1 , . . . , Vn } = Rn .
In the Theorem 2.6, if {W1 , . . . , Wr } is a basis of span{V1 , . . . , Vk }, we can switch the roles
of {V1 , . . . , Vk } and {W1 , . . . , Wr } in the entire argument, and get r k as well.
Corollary 2.8 (Dimension theorem). Let {V1 , . . . , Vk } Rn and {W1 , . . . , Wr } Rn be
linearly independent sets such that span{V1 , . . . , Vk } = span{W1 , . . . , Wr }. Then k = r.
This corollary motivates the following definition.
Definition 2.9. If a linear space X has a basis with k elements, we call k the dimension of
X, and write k = dim X.
Example 2.10. (a) We have dim{0} = 0.
(b) For A from Example 2.5, we have dim ran(A) = 2.
(c) Since the standard basis {e1 , e2 , . . . , en } Rn has n elements, we have dim Rn = n.
(d) If X Rn is a linear space and k = dim X, then Corollary 2.7 implies that k n.
Intuitively, dim X is the number of degrees of freedom in X, or how many free variables
we need in order to describe a point in X. At this point, dim X is defined only when X admits
a basis. However, the following result guarantees that this does not pose any restriction.
Theorem 2.11 (Basis theorem). Every linear space X Rn admits a basis.
Proof. Since the empty set is a basis of {0}, we can assume that X = {0}. Hence there
is a vector V1 X with V1 = 0. If X = span{V1 }, we are done. If X = span{V1 }, then
there is a vector V2 X with V2 span{V1 }. Suppose that we continued this process, and
got X = span{V1 , V2 , . . . , Vk } for some k. By construction, the set {V1 , V2 , . . . , Vk } is linearly
independent, and so this set is a basis of X. Anticipating a contradiction, now suppose that we
never get X = span{V1 , V2 , . . . , Vk } for any k, so that we have an infinite sequence of vectors
V1 , V2 , . . . in X. By construction, {V1 , . . . , Vn+1 } is linearly independent, and so dim Y = n+1
for Y = span{V1 , . . . , Vn+1 }. However, since Vi Rn for all i, we have Y Rn , implying that
dim Y n, a contradiction.
Exercise 2.1. Let X Rn and Y Rn be linear spaces satisfying dim X = dim Y and
X Y . Show that X = Y .
LINEAR ALGEBRA
(39)
implies x1 = . . . = xn = 0. In other words, dim ran(A) = n if and only if ker(A) = {0}. This
suggests that there might be some law relating the dimensions of ran(A) and ker(A).
Suppose that the columns A1 , . . . , An are linearly dependent, and that we remove redundant
columns one by one to get a basis of ran(A). Let us consider the first step of this process.
Thus without loss of generality, we assume that
1 A1 + 2 A2 + . . . + n An = 0,
(40)
with 1 = 1. Taking this into account, we infer that x ker(A) if and only if
x1 A1 + x2 A2 + . . . + xn An = (x2 2 x1 )A2 + . . . + (xn n x1 )An = 0.
(41)
Let us denote by A the matrix consisting of the columns A2 , . . . , An , that is, the matrix that
results from removing the column A1 from A. Recall that ran(A) = ran(A ) by construction.
By the aforementioned argument, we also have
x2 2 x1
ker(A ).
...
(42)
x ker(A)
xn n x1
Hence in order to specify an arbitrary element x ker(A), we can choose x1 R freely,
and then choose the remaining components x2 , . . . , xn so that the second inclusion in (42)
holds. This suggests that dim ker(A) = dim ker(A ) + 1. To make this argument absolutely
convincing, let W1 , . . . , Wr Rn1 be a basis of ker(A ), and define V0 , V1 , . . . , Vr Rn by
1
0
0
2
[W1 ]1
[Wr ]1
,
V0 =
V
=
.
.
.
,
V
=
(43)
1
r
. . .
...
... ,
n
[W1 ]n1
[Wr ]n1
where [Wi ]j denotes the j-th component of the vector Wi Rn1 . Basically, what we have is
V0 = and Vi = (0, Wi ) for i = 1, . . . , r. The set {V0 , V1 , . . . , Vr } is linearly independent, and
AVi = 0 for i = 0, . . . , r. The latter implies that span{V0 , V1 , . . . , Vr } ker(A). Moreover, if
x ker(A), then by (42), for some coecients 1 , . . . , r R, we have
x2 2 x1
= 1 W1 + . . . + r Wr ,
...
(44)
xn n x1
or, equivalently,
x2
2
. . . = 0 . . . + 1 W1 + . . . + r Wr
xn
n
with
0 = x1 .
(45)
(46)
TSOGTGEREL GANTUMUR
Suppose now that we started with A(0) = A, and removed redundant columns one by one to get
a sequence of matrices A(0) , A(1) , . . . , A(k) , with A(k) Rm(nk) having linearly independent
columns. Since ran(A(k) ) = ran(A), we have dim ran(A) = n k. On the other hand, as A(k)
has linearly independent columns, its kernel is trivial, and so (46) yields
dim ker(A) = dim ker(A(1) ) + 1 = . . . = dim ker(A(k) ) + k = k.
(47)
This, at last, gives the sought relationship between dim ran(A) and dim ker(A), which turns
out to be the following beautiful equality:
dim ran(A) + dim ker(A) = n.
(48)
We did not spend much work to prove it, but it is one of most important results from linear
algebra, called the rank-nullity theorem. The name is explained by the following terminology.
Definition 3.1. The rank and the nullity of a matrix A Rmn are the dimension of its
range and of its kernel, respectively, and they are denoted by
rank(A) = dim ran(A),
(49)
(50)
This theorem can be very powerful, because it reduces the complexity of studying two quantities into that of studying either of those quantities. We will see many important applications
in what follows.
4. Invertible matrices
Let F : Rn Rm be a linear mapping, with the associated matrix A Rmn . If for every
Rm there exists a unique x Rn satisfying F (x) = , then we say that F is invertible, and
call the map F 1 : 7 x the inverse of F . As we have discussed in the preceding section,
F is invertible if and only if ran(F ) = Rm and ker(F ) = {0}. By the rank-nullity theorem,
this implies that m + 0 = n, or m = n. In other words, for a linear map to be invertible,
the domain and the codomain spaces must have the same dimensions. Once m = n has been
established, we see that the invertibility of F : Rn Rn is equivalent to either (and hence
both) of the conditions rank(A) = n and nullity(A) = 0.
Let F : Rn Rn be invertible. Then by definition, we have F (F 1 ()) = for all Rm .
Furthermore, for each x Rn , we have F 1 (F (x)) = x, because F (y) = F (x) implies y = x.
Hence for , Rn and , R, we have
+ = F (x) + F (y) = F (x + y),
where x =
F 1 ()
and y =
F 1 (),
(51)
implying that
F 1 ( + ) = F 1 () + F 1 ().
F 1
(52)
Rn
Rn
and
A1 A = I.
(53)
Remark 4.1. Suppose that A, B Rnn satisfy AB = I. This means that ABy = y for all
y Rn , and hence ran(A) = Rn or rank(A) = n. Therefore, the map F : Rn Rn associated
to A is invertible. Moreover, (53) implies that the matrix of the inverse map is given by
A1 = B, and that BA = I. We can apply the same argument to the equality BA = I, and
infer that B 1 = A.
LINEAR ALGEBRA
(54)
k=i
We want to show that ker(A) = {0}, and therefore A is invertible. Let x Rn with x = 0,
and let y = Ax. Pick an index i such that |xi | |xk | for all k = 1, . . . , n. Then we have
yi =
aik xk ,
or
aii xi = yi
aik xk .
(55)
k=i
k=1
k=i
and so
|aik |
(56)
k=i
n
(
)
(57)
k=i
4 1
1
1
0 3
1 1
A=
(58)
3 1 5 0 ,
2 2 1
6
is invertible.
10
TSOGTGEREL GANTUMUR
5. Duality
Our study of matrices gained a lot from writing the matrix-vector product Ax as
Ax = x1 A1 + x2 A2 + . . . + xn An ,
Rm
(59)
Rmn ,
Rn .
where A1 , . . . , An
are the columns of the matrix A
and x
Then a
natural question is: What if we think of a matrix as a collection of its rows, rather than a
collection of columns? This question will lead us to uncover a whole new structure known as
duality. Let A Rmn , and let 1 , . . . , m R1n be the rows of A. Then we can write
1 x
(60)
Ax = . . . .
m x
The elements of R1n are called row vectors.
Remark 5.1. It is convenient to introduce the notation Rn = R1n , and think of the
elements of Rn as n-tuples = (1 , . . . , n ) that behave like 1 n matrices with respect to
matrix-matrix and matrix-vector multiplication operations. That is, we have
[B]k = 1 b1k + . . . + n bnk ,
and
x = 1 x1 + . . . + n xn ,
(61)
for Rn , B = (bik ) Rnm , and x Rn . A more abstract point of view is that the space
Rn is the set of all linear maps f : Rn R. Thus (60) is simply the decomposition of a
linear map F : Rn Rm into its components, as in F (x) = (f1 (x), . . . , fm (x)).
Remark 5.2. Given a matrix
. . . b1,n
. . . b2,n
Rmn ,
... ...
. . . bm,n
b1,1 b1,2
b2,1 b2,2
B=
...
...
bm,1 bm,2
we define its transpose as
b1,1 b2,1
b1,2 b2,2
BT =
... ...
b1,n b2,n
For example, we have
)T (
)
(
1 0
1 2
=
,
2 3
0 3
(62)
. . . bm,1
. . . bm,2
Rnm .
... ...
. . . bm,n
( )T
(
)
3
= 3 2 ,
2
and
(63)
5 7
)T
( )
5
.
7
(64)
In particular, under transposition, row vectors become column vectors and vice versa.
Exercise 5.1. Show that (AB)T = B T AT for A Rmn and B Rnk .
The first thing to notice from (60) is that x ker(A) if and only if i x = 0 for all
i = 1, . . . , m. In other words, x ker(A) if and only if x = 0 for all coran(A), where
coran(A) = {1 1 + . . . + m m : 1 , . . . , m R} Rn ,
(65)
is called the corange of A (also called the coimage of A, or the row space of A). Since the
rows of A are the columns of AT , the corange of A is basically the range of A:
coran(A)
T ran(AT ).
(66)
(67)
LINEAR ALGEBRA
11
T ker(AT ),
(68)
that is, the cokernel of A is the kernel of AT under transposition. In particular, the dimension
of coker(A), or the left-nullity of A, is equal to the nullity of AT . Then the rank-nullity
theorem applied to AT yields
dim coran(A) + dim coker(A) = m.
(69)
(70)
(71)
(72)
Both V and W are linear spaces, and we have V = V and W = W . The following
result was derived earlier in this section, which we restate in terms of annihilators.
Theorem 5.4 (Corange and cokernel). For any matrix A Rmn , we have
coran(A) = ker(A) ,
and
coker(A) = ran(A) .
(73)
(74)
Proof. Let rank(A) = k, and let B1 , . . . , Bk Rm be a basis of ran(A). Then any column Ai
of A can be written as
Ai = ci1 B1 + . . . + cik Bk = BCi ,
(75)
where Ci = (ci1 , . . . , cik ) Rk , and B Rmk is the matrix with B1 , . . . , Bk as its columns.
In other words, we have
A = BC,
(76)
where C Rkn is the matrix with C1 , . . . , Cn as its columns. Denoting by 1 , . . . , m Rn
the rows of A, then we infer
j = bj1 1 + . . . + bjk k ,
(77)
where bj are the entries of B, and 1 , . . . , k Rn are the rows of C. This implies that
coran(A) span{1 , . . . , k }, and hence dim coran(A) k by the Steinitz exchange lemma.
Now we apply the preceding argument to AT , and infer that dim coran(AT ) dim ran(AT ).
Since dim coran(AT ) = dim ran(ATT ) = dim ran(A) = k and dim ran(AT ) = dim coran(A),
we get k dim coran(A), completing the proof.
12
TSOGTGEREL GANTUMUR
6. Determinants
As linear independence is such an important concept, it would be convenient if we were
able to detect whether or not a set of vectors is linearly independent simply by computing
a function of the vectors. That is, it would be great if we had a function of k arguments
V1 , . . . , Vk Rn , with the property that
(V1 , . . . , Vk ) = 0
(78)
From the outset, we restrict ourselves to the linear setting. Thus, we look for a function
: |Rn .{z
. . Rn} R,
(79)
k times
for any V Rn .
(81)
(82)
implying that if {V1 , . . . , Vk } linearly dependent, then (V1 , V2 , . . . , Vk ) = 0 for any alternating k-linear form . In the other direction, we will show later in Theorem 6.5 that if
{V1 , . . . , Vk } is linearly independent, then there is at least one alternating k-linear form
such that (V1 , V2 , . . . , Vk ) = 0.
When combined with multilinearity, (81) implies an antisymmetry condition as follows.
With the notation (V, W ) = (. . . , V, . . . , W, . . .), we have
(V + W, V + W ) = (V, V ) + (V, W ) + (W, V ) + (W, W )
= (V, V ) + (W, W ),
(83)
for any V, W Rn .
(84)
Exercise 6.1. Show that if as in (79) is linear in its first argument, and if it satisfies the
antisymmetry condition (84), then is an alternating k-linear form.
LINEAR ALGEBRA
13
i1 , . . . , ik = 1, . . . , n,
(85)
where ej Rn are the standard basis vectors. If Vi = vi1 e1 + . . . + vin en for i = 1, . . . , k, then
n
n
n
(
)
(V1 , . . . , Vk ) =
v1i1 ei1 , . . . ,
vkik eik =
i1 ...ik v1i1 vkik ,
(86)
i1 =1
ik =1
i1 ,...,ik =1
meaning that the coecients {i1 ...ik } determine the multilinear form completely. Furthermore, the antisymmetry condition (84) implies
...i...j... = (. . . , ei , . . . , ej , . . .), = (. . . , ej , . . . , ei , . . .) = ...j...i... ,
(87)
that is, i1 ...ik is antisymmetric in each pair of its indices. In particular, we have i1 ...ik = 0
whenever ia = ib for a = b.
Now suppose that {i1 ...ik : i1 , . . . , ik = 1, . . . , n} is a collection of coecients satisfying
...i...j... = ...j...i... .
(88)
Then : Rn . . . Rn R defined by
n
(V1 , . . . , Vk ) =
(89)
i1 ,...,ik =1
i,j=1
( )
n!
In general, the dimension of {i1 ...ik R : 1 i1 < i2 < . . . < ik n} is nk = k!(nk)!
.
n
A particular case of interest is alternating n-linear forms in R , where we have only one
independent component in the coecients i1 ...in , because there is only one possibility to have
n integers satisfying 1 i1 < i2 < . . . < in n. Thus, if (i1 , . . . , in ) is an even permutation
of (1, . . . , n), then i1 ...in = 1...n , and if (i1 , . . . , in ) is an odd permutation of (1, . . . , n),
then i1 ...in = 1...n . In other words, taking into account that 1...n = (e1 , . . . , en ), any
alternating n-linear form can be written as
(V1 , . . . , Vn ) = (e1 , . . . , en )
(91)
i1 ,...,in =1
where
0
if (i1 , . . . , in ) is not a permutation of (1, . . . , n).
(92)
14
TSOGTGEREL GANTUMUR
det(V1 , . . . , Vn ) =
(93)
i1 ,...,in =1
B = (bik ) Rnn .
(94)
i1 ,...,in =1
(96)
= b11 (b22 b33 b32 b23 ) b21 (b12 b33 b32 b13 ) + b31 (b12 b23 b22 b13 )
)
)
(
)
(
(
b12 b13
b12 b13
b22 b23
.
+ b31 det
b21 det
= b11 det
b22 b23
b32 b33
b32 b33
Example 6.4. Let B Rnn be an upper triangular matrix, in the sense that bik = 0
whenever i > k. This means that the indices of the nonzero terms in the sum (94) must
satisfy ik k. Since the only permutation (i1 , . . . , in ) of (1, . . . , n) with ik k is (1, . . . , n),
the determinant of a triangular matrix equals the product of the entries on the main diagonal:
det(B) = b11 b22 bnn .
(97)
(98)
for any alternating n-linear form . An important application of this is as follows. Given any
matrix A Rnn , the formula
(V1 , . . . , Vn ) = det(AV1 , . . . , BVn ),
(99)
defines an alternating n-linear form, with (e1 , . . . , en ) = det(A). Hence by (98), we infer
det(AV1 , . . . , AVn ) = det(A) det(V1 , . . . , Vn ),
(100)
(101)
or simply
for any A, B
Rnn .
1
det(A) .
LINEAR ALGEBRA
15
Proof. We have shown in Remark 6.2 that if {V1 , . . . , Vk } is a linearly dependent set and if is
any alternating k-linear form, then (V1 , . . . , Vk ) = 0. In the other direction, let {V1 , . . . , Vk }
be a linearly independent set. We want to show that there is at least one alternating k-linear
form such that (V1 , . . . , Vk ) = 0. We invoke Corollary 2.7 into complete the set {V1 , . . . , Vk }
to a basis {V1 , . . . , Vn } of Rn . This basis defines the invertible linear map F : Rn Rn as
F (x) = x1 V1 + . . . + xn Vn ,
(102)
(103)
to the coecient vector x Rn . Let A Rnn be the matrix associated to F 1 , and let
B Rkn be the matrix formed by the first k rows of A. In other words, BW Rk is equal
to the vector (x1 , . . . , xk ) in column format, for W Rn . Finally, we let
(W1 , . . . , Wk ) = det(BW1 , . . . , BWk ),
(104)
(105)
Since up to scaling, the determinant is the only alternating n-linear form in Rn , in view of
the preceding theorem, we conclude that the columns of A Rnn is linearly independent if
and only if det(A) = 0. In other words, A is invertible if and only if det(A) = 0.
Corollary 6.6. A matrix is invertible if and only if its determinant is nonzero.
Determinants do not only give us a way to check if a matrix invertible, but also are able
to provide an explicit formula for the inverse matrix. Let A Rnn be invertible, let b Rn ,
and consider the equation Ax = b, which can be written as
Ax = x1 A1 + x2 A2 + . . . + xn An = b,
(106)
det(b, A2 , . . . , An )
.
det(A)
(108)
i = 1, . . . , n.
(109)
(110)
To compute the entries of A1 , we start with the observation that AC = I means ACk = ek ,
k = 1, . . . , n, where Ck Rn is the k-th column of C = A1 , and ek is the k-th standard basis
vector in Rn . Thus Cramers rule yields
det(. . . , Ai1 , ek , Ai+1 , . . .)
cik =
,
i, k = 1, . . . , n,
(111)
det(A)
16
TSOGTGEREL GANTUMUR
for the entries of C. We define the adjugate of A, as the matrix adj(A) Rnn whose (i, k)-th
entry is given by
det(. . . , Ai1 , ek , Ai+1 , . . .).
(112)
Then we have
det(A)A1 = adj(A).
(113)