introduction of first and second order tensor with different operators.

© All Rights Reserved

13 visualizações

introduction of first and second order tensor with different operators.

© All Rights Reserved

- Statics- Force Vector 3
- Physics I Problems (6)
- 17 FY13CE Maths Detail Solutions
- Lin
- Mathematical Physics1
- IIT QUIZ
- ch1
- Lesson 1
- Moments in 3D.pdf
- 13885_Lecture Notes 2.9
- BEZIER
- Fall-2008-3[1]
- katznelson-linalgebra
- p4 davis alexander my vector life
- linfun2
- Algorithms
- At JCM H2 Topic12 Notes 37pages
- Hello World
- w1-C
- Vector Problems

Você está na página 1de 136

i i

1 Introduction to Tensors

In elementary physics, we often come across two classes of quantities, namely scalars and

vectors. Mass, density and temperature are examples of scalar quantities, while velocity

and acceleration are examples of vector quantities. A scalar quantitys value does not de-

pend on the choice of the coordinate system. Similarly, although the components of a vector

depend on a particular choice of coordinate system, the vector itself is invariant, and has

an existence independent of the choice of coordinate system. In this chapter, we generalize

the concept of a scalar and vector, to that of a tensor. In this general framework, scalars

are considered as zeroth-order tensors and vectors as first-order tensors. Tensor quantities

of order two and greater, similar to scalars and vectors, have an existence independent

of the coordinate system. Their components, however, just as in the case of vectors, de-

pend on the choice of coordinate system. We will see that the governing field equations

of continuum mechanics can be written as tensorial equations. The advantage of writing

the field equations in such coordinate-free notation is that it is immediately obvious that

these equations are valid no matter what the choice of coordinate system is. A particular

coordinate system is invoked only while solving a particular problem, whence the appro-

priate form of the differential operators and the components of the tensors with respect to

the chosen coordinate system are used. It must be borne in mind, however, that although

using tensorial notation shows the coordinate-free nature of the governing equations in a

given frame of reference, it does not address the issue of how the equations transform under

a change of frame of reference. This aspect will be discussed in greater detail later in this

book.

We now present a review of tensors. Throughout the text, scalars are denoted by light-

face letters, vectors are denoted by boldface lower-case letters, while second and higher-

order tensors are denoted by boldface capital letters. As a notational issue, summation over

repeated indices is assumed, with the indices ranging from 1 to 3. Thus, for example, ui vi

represents u1 v1 + u2 v2 + u3 v3 , and Tij n j represents Ti1 n1 + Ti2 n2 + Ti3 n3 . The quantity on

the right-hand side of a := symbol defines the quantity on its left-hand side. A function

on V V to V means that the function is defined in terms of two elements that belong to

V, and the result is also in V.

In what follows, we consider only real vector spaces. We denote the set of real numbers

by <. A vector space (or linear space) is a set, say V, equipped with an addition function

on V V to V (denoted by +), and a scalar multiplication function on < V to V, which

satisfy the following conditions:

Downloaded from https:/www.cambridge.org/core. Indian Institute of Technology Chennai IIT, on 08 Mar 2017 at 19:06:29, subject to the Cambridge Core terms of use,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

2 Continuum Mechanics

u + v = v + u.

( u + v ) + w = u + ( v + w ).

u + 0 = u.

4. Existence of negative elements: For each u V, there exists a negative element de-

noted u in V such that

u u = 0.

(u + v) = u + v.

6. Distributivity with respect to scalar addition: For all , <, and for all u V,

( + )u = u + u.

( u) = ()u.

1u = u.

Since a vector space is a group with respect to addition (see Section 1.12) with 0 and u

playing the roles of the neutral and reverse elements, respectively, all the results derived

for groups are applicable for vector spaces. In particular, the zero element of V, and the

negative element u corresponding to a given u V are unique. Note also that u + v

V for all , <, and all u, v V.

Perhaps, the most famous example of a vector space is the n-dimensional coordinate

space <n , which is defined by

For <n , addition and scalar multiplication are defined by the relations

u + v : = ( u1 + v1 , u2 + v2 , . . . , u n + v n ),

u := (u1 , u2 , . . . , un ).

Downloaded from https:/www.cambridge.org/core. Indian Institute of Technology Chennai IIT, on 08 Mar 2017 at 19:06:29, subject to the Cambridge Core terms of use,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 3

If p is a positive real number, then another example of a vector space is the space

Z 1

p p

L [0, 1] := f : | f | dx < ,

0

( f )( x ) := [ f ( x )], x [0, 1].

L p [0, 1]. This follows from the inequality

| f + g| p [2 max(| f | , | g|)] p 2 p | f | p + | g| p .

scalars 1 , 2 , . . . , m , not all zero, such that

1 u1 + 2 u2 + + m um = 0.

1 u1 + 2 u2 + + m um = 0,

zero element is always linearly dependent even when the subset {u1 , u2 , . . . , um } is linearly

independent, since the coefficient of the zero element can be taken to be nonzero.

A subset {e1 , e2 , . . . , en } of V is said to be a basis for V if

u V, then

u = u1 e1 + u2 e2 + + u n e n ,

where the scalars u1 , u2 , . . . , un are known as the components of u with respect to the

basis {e1 , e2 , . . . , en }.

If the bases have a finite number of elements, we have the following theorem:

Theorem 1.1.1. All bases for a given vector space contain the same number of elements.

Proof. Suppose that {e1 , e2 , . . . , en } and {e1 , e2 , . . . , em } are bases for a vector

space. Every ei , since it is an element of the vector space, can be expressed

in terms of the basis {e1 , e2 , . . . , en } as ij e j , ij <. Thus,

i ei = ij i e j = 0

Downloaded from https:/www.cambridge.org/core. Indian Institute of Technology Chennai IIT, on 08 Mar 2017 at 19:06:29, subject to the Cambridge Core terms of use,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

4 Continuum Mechanics

ij i = 0 i = 1, 2, . . . , m, j = 1, 2, . . . , n.

Let m > n. Then, the number of unknowns i is more than the number of equa-

tions, so that it is possible to find a nontrivial solution. Thus, there exist 1 , 2 ,

. . ., m , not all zero such that

i ei = 0,

Hence, m n. Next, suppose that m < n. Now reverse the roles of {ei } and

{ei } in the above argument to conclude that m n. Hence, m = n.

basis with n elements. If no such finite integer n exists, then the vector space is said to be

infinite-dimensional. For example, <n is finite-dimensional with

e1 = (1, 0, 0, . . . , 0),

e2 = (0, 1, 0, . . . , 0),

...

en = (0, 0, 0, . . . , 1),

vector space. Note that even in the finite-dimensional case, the basis need not be unique.

When V is finite-dimensional, using the linear independence of {e1 , e2 , . . . , en }, it is easy to

show that the components of any element u are unique.

Let V be an n-dimensional vector space. A subset { f 1 , f 2 , . . . , f m } of V, where

1. m > n cannot be a basis for V, since, as can be shown by following the same method-

ology as used in the proof of Theorem 1.1.1, it is linearly dependent.

2. m < n cannot be a basis for V, since by Theorem 1.1.1, any basis has n elements.

Although, the subset { f 1 , f 2 , . . . , f m } can be linearly independent in this case, an

arbitrary element of V cannot be expressed in terms of the elements of this subset.

following theorem is useful in finding if a given subset with n elements is a basis for an

n-dimensional vector space:

one of whose bases is given by {e1 , e2 , . . . , en }. Since { f 1 , f 2 , . . . , f n } is a subset of V,

each f i can be expressed as a linear combination of the basis vectors {e1 , e2 , . . . , en }, i.e.,

f i = ij e j , i, j = 1, 2, . . . , n, ij <. The following statements are equivalent:

1. { f 1 , f 2 , . . . , f n } is a basis.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 5

2. { f 1 , f 2 , . . . , f n } is linearly independent.

3. det [ ij ] 6= 0.

Proof. We prove that (i) = (ii) = (iii) = (i).

If { f 1 , f 2 , . . . , f n } is a basis, then it is linearly independent by definition.

To prove that (ii) = (iii), we show that not (iii) = not (ii). Thus, let

det [ ij ] = 0. From matrix algebra, it follows that there exist i , i = 1, 2, . . . , n,

not all zero, such that ij i = 0. Thus, there exist i , not all zero, such that

i f i = i ij e j = 0,

To prove that (iii) = (i), note that i f i = ij i e j = 0, implies, since

{e1 , e2 , . . . , en } is a basis (and hence a linearly independent set), that ij i = 0,

i, j = 1, 2, . . . , n, and since det [ ij ] 6= 0, it follows that all the i are zero,

thus showing that { f 1 , f 2 , . . . , f n } is linearly independent. In addition, since

det [ ij ] 6= 0, the matrix [ ij ] is invertible. Let [ij ] denote the components of this

inverse, i.e., ij jk = ik . Then

ij f j = ij jk ek = ik ek = ei .

ui ei , so that on using the above relation, we have

u = ui ei = ij ui f j .

Thus, the set { f 1 , f 2 , . . . , f n } satisfies the two conditions for qualifying as a basis.

bination of any two of its elements also lies in Vs , i.e., if (u + v) Vs for any arbitrary

u, v Vs , and , <. For example, <2 is a linear subspace of <3 . By assuming the op-

erations of addition and scalar multiplication to be the same as that for the parent space

V, it can be shown by verifying the defining axioms that a linear subspace is itself a vector

space.

The set of all linear combinations of a subset {u1 , u2 , . . . , um } of V is called the linear

span of {u1 , u2 , . . . , um }, i.e.,

Lsp{u1 , u2 , . . . , um } := {u : u = 1 u1 + 2 u2 + + m um , i <}.

From this definition, it immediately follows that Lsp{u1 , u2 , . . . , um } is a linear subspace

of V.

An inner product space (or Euclidean space) is a vector space V equipped with a function

on V V to <, denoted by (u, v), and called the inner product (also called scalar product or

dot product) of u and v, that satisfies the following conditions:

(u, v) = (v, u) u, v V, (1.1a)

(u + v, w) = (u, w) + (v, w) , < and u, v, w V, (1.1b)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

6 Continuum Mechanics

From the above definition, and the properties of the inner product it is obvious that

(1.2)

|u| 0 with |u| = 0 if and only if u = 0.

Both <n and L2 [0, 1] are inner product spaces. In the case of <n , the inner product can

be defined by

(u, v) := u1 v1 + u2 v2 + + un vn .

The choice of inner product is not unique. If S is a symmetric, positive definite n n matrix,

then another choice of inner product for <n is

n n

(u, v) := Sij ui v j .

i =1 j =1

Z 1

( f , g) := f ( x ) g( x ) dx.

0

Then

Proof. If (u, v) = 0, then the above inequality is obvious. If it is not zero, then u

and v are nonzero vectors, so that (u, u) and (v, v) are nonzero. By Eqn. (1.1c),

we have

(u v, u v) 0 <.

Expanding this equation using the properties of the inner product, we get

minimum at = (u, v)/(v, v). Substituting this value of into Eqn. (1.4), we

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 7

get Eqn. (1.3). Alternatively, one can use a modified form of Eqn. (5.128a) that

includes equality to directly obtain Eqn. (1.3).

If u and/or v is the zero element, then by the remark following the defini-

tion of linear independence, u and v are linearly dependent, and we also have

equality in Eqn. (1.3). If they are linearly dependent and nonzero, then there

exists < such that u = v, and equality in Eqn. (1.3) follows immediately.

Conversely, if there is equality in Eqn. (1.3), and if u and/or v is the zero ele-

ment, then they are linearly dependent. If there is equality, and both u and v are

nonzero, then by letting = (u, v)/(v, v), we see that

(u v, u v) = 0,

!2 ! !

n n n

ui vi u2i v2i ,

i =1 i =1 i =1

Z 1

2 Z 1

Z 1

2 2

f ( x ) g( x ) dx f ( x ) dx g( x ) dx .

0 0 0

|u + v| |u| + |v| u, v V.

Proof.

|u + v|2 = (u + v, u + v)

= |u|2 + 2(u, v) + |v|2

|u|2 + 2 |(u, v)| + |v|2

| u |2 + 2 | u | | v | + | v |2 (by the CauchySchwartz inequality)

2

= (|u| + |v|) .

A vector space V is a normed vector space, if we assign a nonnegative real number, kuk,

called the norm of u, to each u V such that

kuk 0 u V with kuk = 0 if and only if u = 0,

ku + vk kuk + kvk .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

8 Continuum Mechanics

One can show that L p [0, 1], p 1 is a normed vector space. From Eqns. (1.2) and the

triangle inequality, it is clear that an inner product space is a normed vector space with the

norm defined by

three-dimensional space, we now specialize the results of this section to the case n = 3,

and henceforth present results only for this case.

From now on, V denotes the three-dimensional Euclidean space <3 . Let {e1 , e2 , e3 } be a

fixed set of orthonormal vectors that constitute the Cartesian basis. We have

ei e j = ij ,

(

0 when i 6= j,

ij := (1.5)

1 when i = j.

The Kronecker delta is also known as the substitution operator, since, from the definition,

we can see that xi = ij x j , ij = ik kj , and so on. Note that ij = ji , and ii = 11 + 22 +

33 = 3.

Any vector u can be written as

u = u1 e1 + u2 e2 + u3 e3 , (1.6)

u = ui ei .

(u, v) = u v := ui vi = u1 v1 + u2 v2 + u3 v3 . (1.7)

ui = u ei . (1.8)

u = ( u ei ) ei . (1.9)

e j ek := eijk ei , (1.10)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 9

where eijk are the components of a third-order tensor E known as the alternate tensor

(which we will discuss in greater detail in Section 1.7), and are given by

e132 = e213 = e321 = 1

eijk = 0 otherwise.

Taking the dot product of both sides of Eqn. (1.10) with em , we get

em (e j ek ) = eijk im = emjk .

eijk = ei (e j ek ). (1.11)

If w denotes the cross product of u and v, then by using this property and Eqn. (1.10), we

have

w = uv

= (u j e j ) (vk ek )

= eijk u j vk ei . (1.12)

u v = v u.

Taking v = u, we get u u = 0.

The scalar triple product of three vectors u, v, w, denoted by [u, v, w], is defined by

[u, v, w] := u (v w).

(1.14)

If any two elements in the scalar triple product are the same, then its value is zero, as

can be seen by interchanging the identical elements, and using the above formula. From

Eqn. (1.13), it is also clear that the scalar triple product is linear in each of its argument

variables, so that, for example,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

10 Continuum Mechanics

u1 u2 u3

[u, v, w] = det v1 v2 v3 . (1.16)

w1 w2 w3

Using Eqns. (1.11) and (1.16), the components of the alternate tensor can be written in

determinant form as follows:

e i e1 e i e2 e i e3 i1 i2 i3

eijk = ei , e j , ek = det e j e1 e j e2 e j e3 = det j1 j3 . (1.17)

j2

e k e1 e k e2 e k e3 k1 k2 k3

Thus, we have

i1 i2 i3 p1 p2 p3

= det j1 j2 j3 det q1 q2

eijk e pqr q3

k1 k2 k3 r1 r2 r3

i1 i2 i3 p1 q1 r1

= det j1 j2 j3 det p2 q2 (since det T = det( T T ))

r2

k1 k2 k3 p3 q3 r3

i1 i2 i3

p1 q1 r1

= det j1 j2 j3 p2 q2 (since (det R)(det S) = det( RS))

r2

k1 k2 k3 p3 q3 r3

im mp im mq im mr

= det jm mp jm mq jm mr

km mp km mq km mr

ip iq ir

= det jp jq jr . (1.18)

kp kq kr

From Eqn. (1.18) and the relation ii = 3, we obtain the following identities (the first of

which is known as the e identity):

eijk eijm = 2km , (1.19b)

eijk eijk = 6. (1.19c)

(u v) (u v) = eijk eimn u j vk um vn

= (jm kn jn km )u j vk um vn

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 11

= (um vn um vn u j vm um v j )

= (u u)(v v) (u v)2 . (1.20)

Theorem 1.2.1. For any two vectors u, v V, u v = 0 if and only if u and v are

linearly dependent.

Proof. If u and v are linearly dependent, and u and/or v is zero, then it is obvious

that u v = 0. If they are linearly dependent and nonzero, then there exists a

scalar such that v = u. Hence

u v = u u = 0.

Conversely, if u v = 0, then

0 = (u v) (u v)

= (u u)(v v) (u v)2 , (by Eqn. (1.20))

But by Theorem 1.1.3, this implies that u and v are linearly dependent.

u with v w, and the cross product of u v with w, respectively, are different in general,

and are given by

(u v) w = (u w)v (v w)u. (1.21b)

u (v w) = eijk u j (v w)k ei

= eijk ekmn u j vm wn ei

= ekij ekmn u j vm wn ei

= (im jn in jm )u j vm wn ei

= ( u n w n v i u m v m wi ) e i

= (u w)v (u v)w.

For scalar triple products, we have the following useful result:

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

12 Continuum Mechanics

Theorem 1.2.2. The scalar triple product [u, v, w] is zero if and only if u, v and w are

linearly dependent.

Proof. Since

u = u1 e1 + u2 e2 + u3 e3 ,

v = v1 e1 + v2 e2 + v3 e3 ,

w = w1 e 1 + w2 e 2 + w3 e 3 ,

by Theorem 1.1.2, {u, v, w} is linearly independent if and only if (see Eqn. (1.16))

u1 u2 u3

[u, v, w] = det v1 v2 v3 6= 0,

w1 w2 w3

A second-order tensor is a linear transformation that maps vectors to vectors. We shall

denote the set of second-order tensors by Lin. If T is a second-order tensor that maps a

vector u to a vector v, then we write it as

v = Tu. (1.22)

T (0) = 0.

From the definition of a second-order tensor, it follows that the sum of two second-order

tensors defined by

( R + S)u := Ru + Su u V,

(T )u := ( Tu) u V,

are both second-order tensors. The two second-order tensors R and S are said to be equal

if

Ru = Su u V. (1.23)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 13

To see this, note that if Eqn. (1.23) holds, then clearly Eqn. (1.24) holds. On the other hand,

if Eqn. (1.24) holds, then using the bilinearity property of the inner product, we have

(v, ( Ru Su)) = 0 u, v V.

If we define the element Z as

Zu = 0 u V,

then we see that Z Lin, and is the zero element of Lin since

( T + Z )u = Tu + Zu = Tu u V,

T + Z = T.

Henceforth, we simply use the symbol 0 to denote the zero elements of both Lin and V.

With the above definitions of addition, scalar multiplication, and zero, Lin is a vector space,

and hence, all the results that we have derived for vector spaces in Section 1.1 are valid for

Lin.

If we define the function I : V V by

Iu := u u V, (1.25)

Choosing u = e1 , e2 and e3 in Eqn. (1.22), we get three vectors that can be expressed as

a linear combination of the base vectors ei as

Te1 = 1 e1 + 2 e2 + 3 e3

Te2 = 4 e1 + 5 e2 + 6 e3 (1.26)

Te3 = 7 e1 + 8 e2 + 9 e3 ,

Te j = Tij ei . (1.27)

The elements Tij are called the components of the tensor T with respect to the base vectors

e j ; as seen from Eqn. (1.27), Tij is the component of Te j in the ei direction. Taking the dot

product of both sides of Eqn. (1.27) with ek for some particular k, we get

ek Te j = Tij ik = Tkj ,

or, replacing k by i,

Tij = ei Te j . (1.28)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

14 Continuum Mechanics

By choosing v = ei and u = e j in Eqn. (1.24), it is clear that the components of two equal

tensors are equal. From Eqn. (1.28), the components of the identity tensor in any orthonor-

mal coordinate system ei are

Iij = ei Ie j = ei e j = ij . (1.29)

Thus, the components of the identity tensor are scalars that are independent of the Carte-

sian basis. Using Eqn. (1.27), we write Eqn. (1.22) in component form (where the compo-

nents are with respect to a particular orthonormal basis {ei }) as

vi ei = T (u j e j ) = u j Te j = u j Tij ei ,

which, by virtue of the uniqueness of the components of any element of a vector space,

yields

vi = Tij u j . (1.30)

Thus, the components of the vector v are obtained by a matrix multiplication of the com-

ponents of T, and the components of u.

The transpose of T, denoted by T T , is defined using the inner product as

Once again, it follows from the definition that T T is a second-order tensor. The transpose

has the following properties:

( T T )T = T,

(T )T = T T ,

( R + S)T = RT + ST .

If ( Tij ) represent the components of the tensor T, then the components of T T are

( T T )ij = ei T T e j

= Tei e j

= Tji . (1.32)

T T = T,

T T = T.

Any tensor T can be decomposed uniquely into a symmetric and an skew-symmetric part

as (see Problem 5)

T = T s + T ss , (1.33)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 15

where

1

Ts = ( T + T T ),

2

1

T ss = ( T T T ).

2

The product of two second-order tensors RS is the composition of the two operations R

and S, with S operating first, and defined by the relation

( RS)u := R(Su) u V. (1.34)

Since RS is a linear transformation that maps vectors to vectors, we conclude that the

product of two second-order tensors is also a second-order tensor. From the definition

of the identity tensor given by (1.25) it follows that RI = IR = R. If T represents the

product RS, then its components are given by

Tij = ei ( RS)e j

= ei R(Se j )

= ei R(Skj ek )

= ei Skj Rek

= Skj (ei Rek )

= Skj Rik

= Rik Skj , (1.35)

which is consistent with matrix multiplication. Also consistent with the results from matrix

theory, we have ( RS) T = S T R T , which follows from Eqns. (1.24), (1.31) and (1.34).

We now introduce the concept of a tensor product, which is convenient for working with

tensors of rank higher than two. We first define the dyadic or tensor product of two vectors

a and b by

( a b)c := (b c) a c V. (1.36)

Note that the tensor product a b cannot be defined except in terms of its operation on a

vector c. We now prove that a b defines a second-order tensor. The above rule obviously

maps a vector into another vector. All that we need to do is to prove that it is a linear map.

For arbitrary scalars c and d, and arbitrary vectors x and y, we have

( a b)(cx + dy) = [b (cx + dy)] a

= [cb x + db y] a

= c(b x) a + d(b y) a

= c[( a b) x] + d [( a b)y] ,

which proves that a b is a linear function. Hence, a b is a second-order tensor. Any

second-order tensor T can be written as

T = Tij ei e j , (1.37)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

16 Continuum Mechanics

where the components of the tensor, Tij are given by Eqn. (1.28). To see this, we consider

the action of T on an arbitrary vector u:

= ei T ( u j e j ) ei

= u j ei ( Te j ) ei

= (u e j ) ei ( Te j ) ei

= ei ( Te j ) (u e j )ei

= ei ( Te j ) (ei e j )u

= ei ( Te j ) ei e j u.

Hence, we conclude that any second-order tensor admits the representation given by

Eqn. (1.37), with the nine components Tij , i = 1, 2, 3, j = 1, 2, 3, given by Eqn. (1.28). The

dyadic products {ei e j }, i = 1, 2, 3, j = 1, 2, 3 constitute a basis of Lin since, as just men-

tioned, any element of Lin can be expressed in terms of them, and since they are linearly

independent (Tij ei e j = 0 implies that Tij = ei 0e j = 0).

From Eqns. (1.29) and (1.37), it follows that

I = ei ei , (1.38)

Eqn. (1.37), it follows from Eqn. (1.32) that the transpose of T can be represented as

T T = Tji ei e j . (1.39)

From Eqns. (1.37) and (1.39), we deduce that a tensor is symmetric (T = T T ) if and only

if Tij = Tji for all possible i and j. We now show how all the properties of a second-order

tensor derived so far can be derived using the dyadic product.

Using Eqn. (1.27), we see that the components of a dyad a b are given by

( a b)ij = ei ( a b)e j

= ei ( b e j ) a

= ai b j . (1.40)

are obtained by noting that

Convention regarding complex vectors: In what follows, we will often encounter the

dyadic product of two complex-valued vectors. We note here that we use the same def-

inition for the dyadic product of complex-valued vectors as that for real-valued vectors,

namely, the definition given by Eqn. (1.36). Using Eqn. (1.40), we get the components of

a b as ai b j . However, it is possible to follow another convention in defining the dyadic

product, namely ( a b)c = (b c) a (with the hat denoting complex conjugation), whereby

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 17

one obtains the components as ai b j . We will follow this rule of using the same definitions

for the real and complex case even when we define the dyadic products A B and A B

of two complex-valued second-order tensors A and B. The advantage of our convention is

that all relations that we derive for the real case are also valid for the complex case, since

the definitions are the same. If we used the alternate convention, one would need to re-

derive all the relations for the complex case, and this can be quite cumbersome, since we

will be deriving dozens of relations (including rules for differentiation) involving tensor

products. We emphasize that both conventions are correct, and, as long as one sticks to one

convention consistently in all the derivations, the particular choice made is just a matter of

convenience.

Theorem 1.3.1. If (u, v, w) and ( a, b, c) are two pairs of linearly independent vectors,

and T is an arbitrary tensor, then

[ Tu, v, w] + [u, Tv, w] + [u, v, Tw] [ T a, b, c] + [ a, Tb, c] + [ a, b, Tc]

= , (1.42a)

[u, v, w] [ a, b, c]

[ Tu, Tv, w] + [u, Tv, Tw] + [ Tu, v, Tw] [ T a, Tb, c] + [ a, Tb, Tc] + [ T a, b, Tc]

= , (1.42b)

[u, v, w] [ a, b, c]

[ Tu, Tv, Tw] [ T a, Tb, Tc]

= . (1.42c)

[u, v, w] [ a, b, c]

can be written as

ui v j wk Tei , e j , ek + ei , Te j , ek + ei , e j , Tek

LHS =

[u, v, w]

eijk ui v j wk {[ Te1 , e2 , e3 ] + [e1 , Te2 , e3 ] + [e1 , e2 , Te3 ]}

=

[u, v, w]

= [ Te1 , e2 , e3 ] + [e1 , Te2 , e3 ] + [e1 , e2 , Te3 ] .

Since the choice of vectors u, v and w is arbitrary, Eqn. (1.42a) follows. Equa-

tions (1.42b) and (1.42c) are proved in a similar manner.

three scalars, I1 , I2 and I3 such that

[ Tu, v, w] + [u, Tv, w] + [u, v, Tw] = I1 [u, v, w] , (1.43a)

[ Tu, Tv, w] + [u, Tv, Tw] + [ Tu, v, Tw] = I2 [u, v, w] , (1.43b)

[ Tu, Tv, Tw] = I3 [u, v, w] . (1.43c)

The above equations are valid even when [u, v, w] = 0. To see this, consider Eqn. (1.43a).

Following the proof of Theorem 1.3.1, we see that if [u, v, w] = 0, then

[ Tu, v, w] + [u, Tv, w] + [u, v, Tw] = [u, v, w] {[ Te1 , e2 , e3 ] + [e1 , Te2 , e3 ] + [e1 , e2 , Te3 ]} = 0.

Equations (1.43b) and (1.43c) can be proved analogously.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

18 Continuum Mechanics

The scalars I1 , I2 and I3 are called the principal invariants of T. The reason for calling

them as the principal invariants is that any other scalar invariant of T can be expressed in

terms of them. The first and third invariants are referred to as the trace and determinant of

T, respectively, and written as

I1 = tr T, I3 = det T.

Using the properties of the scalar triple product, it is clear that the trace is a linear operation,

i.e.,

Similarly, we have

tr T T = e1 T T e1 + e2 T T e2 + e3 T T e3

= e1 Te1 + e2 Te2 + e3 Te3

= tr T. (1.45)

tr ( a b) = a1 b1 + a2 b2 + a3 b3 = ai bi = a b. (1.46)

Using the linearity of the trace operator, and Eqn. (1.37), we get

tr T = tr Tij ei e j = Tij tr (ei e j ) = Tij ei e j = Tii ,

To prove

tr ( RS) = tr (SR),

= (Su, R T u) + (Sv, R T v) + (Sw, R T w).

Substituting

= (u, Ru)u + (u, Rv)v + (u, Rw)w,

tr ( RS) =(u, Ru)(u, Su) + (u, Rv)(v, Su) + (u, Rw)(w, Su)+

(v, Ru)(u, Sv) + (v, Rv)(v, Sv) + (v, Rw)(w, Sv)+

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 19

Interchanging R and S in the above equation gives an expression for tr (SR), which is the

same as that for tr ( RS), thus yielding the desired result. The proof using indicial notation

is left as an exercise (Problem 6).

Similar to the vector inner product given by Eqn. (1.7), we can define a tensor inner

product of two second-order tensors R and S, denoted by R : S, by

The fact that R : S satisfies the inner product conditions in Eqn. (1.1) can be easily verified.

Thus, Lin equipped with the above inner product is an inner product space, and enjoys all

the properties derived in Section 1.1. In particular, if the magnitude associated with the

inner product is given by

| T | = ( T : T )1/2 ,

then by the CauchySchwartz inequality, we get

|( R, S)| | R| |S| .

We have the following useful property:

R : (ST ) = (S T R) : T = ( RT T ) : S = ( T R T ) : S T , (1.48)

since

R : (ST ) = tr (ST ) T R = tr T T (S T R) = (S T R) : T = ( R T S) : T T

= tr S T ( RT T ) = ( RT T ) : S = ( T R T ) : S T .

From Eqn. (1.43c), it can be seen that

det I = 1,

det(T ) = 3 det T.

= det R [Su, Sv, Sw]

= (det R)(det S) [u, v, w] .

Choosing u, v and w to be linearly independent, the above equation leads us to the relation

By choosing u = e j ek = eijk ei , v = e j and w = ek in Eqn. (1.43c), so that [u, v, w] =

eijk eijk = 6 (by Eqn. (1.19c)), and using the fact that Tei = Tpi e p , we get

1 1

det T = eijk e pqr Tpi Tqj Trk = eijk e pqr Tip Tjq Tkr , (1.50)

6 6

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

20 Continuum Mechanics

where the second relation is obtained by interchanging the dummy indices. From Eqn. (1.50),

it is immediately obvious that (see Section 1.3.4 for a proof without using indicial notation)

By choosing u = e p , v = eq and w = er in Eqn. (1.43c), and using Eqn. (1.51), we also have

e pqr (det T ) = eijk Tip Tjq Tkr = eijk Tpi Tqj Trk . (1.52)

= det {[u e1 + v e2 + w e3 ] [e1 p + e2 q + e3 r ]}

= det [u p + v q + w r ] . (1.53)

Theorem 1.3.2. Given a tensor T, there exists a nonzero vector n such that Tn = 0 if

and only if det T = 0.

Proof. If det T = 0, then by Eqn. (1.43c),

which, by Theorem 1.2.2, implies that Tu, Tv, Tw are linearly dependent. This

means that there exist scalars , and , not all zero, such that

Tu + Tv + Tw = T (u + v + w) = 0.

Thus, Tn = 0, where n = u + v + w.

Conversely, if there exists a nonzero vector n such that Tn = 0, choose v and

w such that n,v and w are linearly independent, i.e., [n, v, w] 6= 0. Then, by

Eqn. (1.43c), we have

In order to define the concept of a inverse of a tensor, it is convenient to first introduce the

cofactor tensor, denoted by cof T, and defined by the relation

cof T (u v) := Tu Tv u, v V. (1.54)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 21

cof T obviously maps a vector to a vector. To prove linearity, note that every nonzero vec-

tor w can be expressed as u v for some nonzero u, v V (e.g., take v as a unit vector

perpendicular to w, and u = v w). We now show that if w1 and w2 are two arbitrary

nonzero vectors, they can be expressed as u1 v and u2 v. If w1 and w2 are linearly

dependent, then w2 = w1 , <, and then u2 = u1 . If w1 and w2 are linearly in-

dependent, then let v = (w1 w2 )/ |w1 w2 |2 . Using Eqn. (1.21a), we can show that

u1 = (w1 w2 )w1 + |w1 |2 w2 and u2 = |w2 |2 w1 + (w1 w2 )w2 yield u1 v = w1 and

u2 v = w2 . Noting that T Lin, we now have

= [ T (u1 + u2 )] Tv

= [Tu1 + Tu2 ] Tv

= ( Tu1 Tv) + ( Tu2 Tv)

= cof T (u1 v) + cof T (u2 v)

= cof T w1 + cof T w2 ,

Using Eqn. (1.54), we now prove the following explicit formula for the cofactor:

h i

tr T [ Tu, v, w] = T 2 u, v, w + [ Tu, Tv, w] + [ Tu, v, Tw] ,

h i

[ Tu, Tv, w] + [ Tu, v, Tw] = tr T [ Tu, v, w] T 2 u, v, w . (1.56)

h i

I2 [u, v, w] = [ Tu, Tv, w] + [ Tu, v, Tw] + (cof T ) T u, v, w . (1.57)

Substituting Eqn. (1.56) into (1.57), using the fact that u, v and w are arbitrary, and invoking

equality of tensors, we get Eqn. (1.55).

It immediately follows from Eqn. (1.55) that cof T corresponding to a given T is unique.

We also observe that

and that

1

(cof T )ij = e e Tmp Tnq . (1.60)

2 imn jpq

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

22 Continuum Mechanics

To prove this, let u = eq e j and v = eq in Eqn. (1.54), and use Eqn. (1.21b) to get u v =

2e j . On the right-hand side of Eqn. (1.54), use the relations eq e j = e pqj e p , Te p = Tmp em

and Teq = Tnq en . Taking the dot product with ei of both sides of the relation so obtained,

we get the desired result. Equation (1.60) when written out explicitly reads

T22 T33 T23 T32 T23 T31 T21 T33 T21 T32 T22 T31

[cof T ] = T32 T13 T33 T12 T33 T11 T31 T13 T31 T12 T32 T11 . (1.61)

T12 T23 T13 T22 T13 T21 T11 T23 T11 T22 T12 T21

Equation (1.54) can be used to get a simple expression for I2 , the second invariant of a

tensor. We first write Eqn. (1.43b) as

I2 [u, v, w] = [u, Tv, Tw] + [v, Tw, Tu] + [w, Tu, Tv] ,

and then use Eqn. (1.54) to write the above equation as

h i h i h i

I2 [u, v, w] = (cof T ) T u, v, w + (cof T ) T v, w, u + (cof T ) T w, u, v

h i h i h i

= (cof T )T u, v, w + u, (cof T )T v, w + u, v, (cof T )T w

= tr (cof T )T [u, v, w] ,

where the last step follows from Eqn. (1.43a). Using Eqn. (1.45), and choosing u, v and w

to be linearly independent (so that [u, v, w] 6= 0), we get

I2 = tr (cof T ). (1.62)

An alternative expression for the I2 can be found by directly taking the trace of both sides

of Eqn. (1.55). Using Eqns. (1.45) and (1.62), and the fact that the trace is a linear operator,

we get

I2 = 3I2 (tr T )2 + tr T 2 ,

which implies that

1h i

I2 = (tr T )2 tr T 2 . (1.63)

2

The equivalence of the two formulae for I2 given by Eqns. (1.62) and (1.63) can also be

shown using indicial notation (see Problem 10).

Substituting Eqn. (1.54) into Eqn. (1.43c), we get

( Tu, cof T (v w)) = det T ( Iu, v w) u, v, w V,

which implies that

Any arbitrary vector z can be expressed as v w for some v, w V (take w as a unit vector

perpendicular to z, and v = w z), so that from Eqn. (1.24), it follows that (cof T ) T T =

(det T ) I, which when combined with Eqn. (1.59) yields

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 23

Similar to the result for determinants, the cofactor of the product of two tensors is the

product of the cofactors of the tensors (see also Problem 9), i.e.,

This is because

= (det R)(cof S)T IS (by Eqn. (1.64))

= (det R)(det S) I (by Eqn. (1.64))

= det( RS) I, (by Eqn. (1.49))

The inverse of a second-order tensor T, denoted by T 1 , is defined by

T 1 T = I, (1.65)

exists, is unique.

Proof. Assuming T 1 exists, from Eqns. (1.49) and (1.65), we have

(det T )(det T 1 ) = 1, and hence det T 6= 0.

Conversely, if det T 6= 0, then from Eqn. (1.64), we see that at least one inverse

exists, and is given by

1

T 1 = (cof T )T . (1.66)

det T

it follows that ( T 11 T 21 ) T = 0. Choose T 21 to be given by the expression in

Eqn. (1.66) so that, by virtue of Eqn. (1.64), we also have T T 21 = I. Multiplying

both sides of ( T 11 T 21 ) T = 0 by T 21 , we get T 11 = T 21 , which establishes

the uniqueness of T 1 .

T 1 T = T T 1 = I. (1.67)

Tu = v,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

24 Continuum Mechanics

T 1 v = T 1 Tu

= Iu

=u

To prove the converse, note that if u1 and u2 are two solutions of the equation

Tu = v, then, by the assumed uniqueness, u1 = u2 . Also, by assumption, for

every v V, there exists a u such that Tu = v. Thus, the mapping T : V V is

one-to-one and onto, and hence, T 1 : V V exists.

Tu = v u = T 1 v, u, v V.

By the above property, T 1 clearly maps vectors to vectors. Hence, to prove that T 1 is a

second-order tensor, we just need to prove linearity. Let a, b V be two arbitrary vectors,

and let u = T 1 a and v = T 1 b. Since I = T 1 T, we have

I (u + v) = T 1 T (u + v)

= T 1 [ T (u + v)]

= T 1 [Tu + Tv]

= T 1 (a + b),

which implies that

T 1 (a + b) = T 1 a + T 1 b a, b V and , <.

The inverse of the product of two invertible tensors R and S is

( RS)1 = S1 R1 , (1.68)

S1 R1 RS = S1 IS = S1 S = I.

the transpose is given by

( T T ) 1 = ( T 1 ) T , (1.69)

since

( T 1 )T T T = ( T T 1 )T = I T = I.

Hence, without fear of ambiguity, we can write

T T : = ( T T ) 1 = ( T 1 ) T .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 25

Although, no easy expression for ( T 1 + T 2 )1 is known, the following Woodbury for-

mula gives an expression for a particular kind of perturbation on an invertible tensor T,

and which holds for any underlying space dimension n:

( T + UCV )1 = T 1 T 1 U (C 1 + V T 1 U )1 V T 1 . (1.70)

U A1 B1

To prove this formula, consider the inversion of the matrix VT

C 1

. If C 1 D1 represents

the inverse of this matrix, then the condition

" #" # " #

A1 B1 T U I 0

= ,

C 1 D 1 V C 1 0 I

A1 T + B1 V = I, (1.71a)

1

A1 U + B1 C = 0. (1.71b)

From Eqn. (1.71b), we get B1 = A1 UC, which on substituting into Eqn. (1.71a) yields

A1 = ( T + UCV )1 . (1.72)

A 1 = T 1 B 1 V T 1 , (1.73)

B 1 = T 1 U ( C 1 + V T 1 U ) 1 .

A 1 = T 1 T 1 U ( C 1 + V T 1 U ) 1 V T 1 . (1.74)

Comparing Eqns. (1.72) and (1.74), we get Eqn. (1.70). If C = 1, then we get the following

ShermanMorrison formula:

T 1 ( u v ) T 1

( T + u v ) 1 = T 1 . (1.75)

1 + v T 1 u

Let S Sym be an invertible tensor, and let u = S1 u and v = S1 v. Note that because

of the symmetry of S, we have u v = u v. By applying Eqn. (1.75) twice, we get

1

( S + u v + v u ) 1 = S 1 + [(u u)v v + (v v)u u (1 + u v)(u v + v u)] ,

D

where

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

26 Continuum Mechanics

As a corollary, for S = I, we get

1 h 2 i

( I + u v + v u ) 1 = I + |u| v v + |v|2 u u (1 + u v)(u v + v u) ,

D

where

D = (1 + u v )2 | u |2 | v |2 .

H = I + u v, (1.76)

and let the underlying space dimension be n. Let {e1 , e2 , . . . , en1 } be n 1 unit vectors that

are perpendicular to v, and mutually perpendicular to each other. Then from Eqn. (1.76),

it is immediately evident that (1, e1 ), (1, e2 ), . . . , (1, en1 ), (1 + u v, u/ |u|) are eigenval-

ues/eigenvectors of H. If u v 6= 0, then all the eigenvectors are linearly independent,

while if u v = 0, then the first n 1 eigenvectors are linearly independent, while the last

can be expressed in terms of them. The characteristic equation is

( 1)n1 ( 1 u v) = 0.

On applying the CayleyHamilton theorem (see Theorem 1.3.5), we get

[ H I ]n1 [ H (1 + u v) I ] = 0.

The minimal polynomial is

[ H I ][ H (1 + u v) I ] = 0,

which can be immediately verified by using Eqn. (1.76). Multiplying the eigenvalues of H,

we get

det H = 1 + u v.

Now consider the tensor that occurs in Eqn. (1.75), which is a generalization of the

tensor H, namely,

M = T + u v,

The result for arbitrary T is obtained by taking the limit in the above equation. We get

Consider a symmetric perturbation to I:

K = I + u v + v u.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 27

Let {e1 , e2 , . . . , en2 } be n 2 unit vectors that are perpendicular to u and v. Then (1, e1 ),

(1, e2 ), . . . , (1, en2 ), (1 + u v + |u| |v| , u/ |u| + v/ |v|), (1 + u v |u| |v| , u/ |u| v/ |v|)

are the eigenvalues/eigenvectors of K, so that

det K 1 + 2u v.

Let S be a symmetric positive definite tensor, and let U = S be its symmetric positive

definite square-root. If Z is defined by

Z := S + u v + v u,

Eqn. (1.77), we get

2

det Z = det S[ 1 + (S1 u) v (u S1 u)(v S1 v)].

Although our proof assumed S to be positive definite, actually, the above result holds for

any invertible S Sym.

If T is an arbitrary tensor, a vector n is said to be an eigenvector of T if there exists such

that

Tn = n. (1.78)

Writing the above equation as ( T I )n = 0, we see from Theorem 1.3.2 that a nontrivial

eigenvector n exists if and only if

det( T I ) = 0.

This is known as the characteristic equation of T. Using Eqn. (1.43c), the characteristic equa-

tion can be written as

[ Tu u, Tv v, Tw w] = 0,

for linearly independent vectors u, v and w. Using Eqns. (1.14) and (1.15), we can write the

characteristic equation as

3 I1 2 + I2 I3 = 0, (1.79)

where I1 , I2 and I3 are the principal invariants given by Eqns. (1.43). Since the principal

invariants are real, Eqn. (1.79) has either one or three real roots. If one of the eigenval-

ues is complex, then it follows from Eqn. (1.78) that the corresponding eigenvector is also

complex. By taking the complex conjugate of both sides of Eqn. (1.78), we see that the com-

plex conjugate of the complex eigenvalue, and the corresponding complex eigenvector are

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

28 Continuum Mechanics

complex, occur in complex conjugate pairs. If 1 , 2 , 3 are the roots of the characteristic

equation, then from Eqn. (1.79), it follows that

= 1 + 2 + 3 , (1.80a)

1h i T T12 T22 T23 T11 T13

I2 = tr cof T = (tr T )2 tr ( T 2 ) = 11 + +

2

T21 T22 T32 T33 T31 T33

= 1 2 + 2 3 + 1 3 , (1.80b)

1h i

I3 = det( T ) = (tr T )3 3(tr T )(tr T 2 ) + 2tr T 3 = eijk Ti1 Tj2 Tk3

6

= 1 2 3 , (1.80c)

where |.| denotes the determinant. The set of eigenvalues {1 , 2 , 3 } is known as the

spectrum of T.

If is an eigenvalue, and n is the associated eigenvector of T, then 2 is the eigenvalue

of T 2 , and n is the associated eigenvector, since

T 2 n = T ( Tn) = T (n) = Tn = 2 n.

and T are the same since their characteristic equations are the same.

An extremely important result is the following:

ing the same form as its characteristic equation, i.e.,

T 3 I1 T 2 + I2 T I3 I = 0 T. (1.81)

(cof T )T T = I2 T I1 T 2 + T 3 .

By taking the trace of both sides of Eqn. (1.81), and using Eqn. (1.63), we get

1h i

det T = (tr T )3 3(tr T )(tr T 2 ) + 2tr T 3 . (1.82)

6

From the above expression and the properties of the trace operator, Eqn. (1.51) follows.

We have

i = 0, i = 1, 2, 3 I T = 0 tr ( T ) = tr ( T 2 ) = tr ( T 3 ) = 0. (1.83)

The proof is as follows. If all the invariants are zero, then from the characteristic equation

given by Eqn. (1.79), it follows that all the eigenvalues are zero. If all the eigenvalues are

zero, then from Eqns. (1.80a)(1.80c), it follows that all the principal invariants I T are zero.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 29

principal invariants are zero. Conversely, if all the principal invariants are zero, then all

j

the eigenvalues are zero from which it follows that tr T j = 3i=1 i , j = 1, 2, 3 are zero.

Consider the second-order tensor u v. By Eqn. (1.60) or by writing u v as u v +

0 0 + 0 0 and using Eqn. (1.333), it follows that

cof (u v) = 0, (1.84)

so that the second invariant, which is the trace of the above tensor, is zero. Similarly, on

using Eqn. (1.53), we get the third invariant as zero. The first invariant is given by u v.

Thus, from the characteristic equation, it follows that the eigenvalues of u v are (0, 0, u

v). If u and v are perpendicular, u v is an example of a nonzero tensor all of whose

eigenvalues are zero.

Let W Skw and let u, v V. Then

On setting v = u, we get

(u, W u) = 0. (1.86)

v = e j , we see from the above results that any skew-symmetric tensor W has only three

independent components (in each coordinate frame), which suggests that it might be re-

placed by a vector. This observation leads us to the following result:

Theorem 1.4.1. Given any skew-symmetric tensor W, there exists a unique vector w,

known as the axial vector or dual vector, corresponding to W such that

Wu = w u u V. (1.87)

Conversely, given any vector w, there exists a unique skew-symmetric second-order

tensor W such that Eqn. (1.87) holds.

Proof. A skew-symmetric tensor W, like any other tensor, has at least one real

eigenvalue, say . Let p be the associated eigenvector. Then W p = p, and on

taking the dot product of both sides with p, we get = p (W p). By virtue

of Eqn. (1.86), = 0 (we show later in this section that this is the only real

eigenvalue). Let q, r be unit vectors that form an orthonormal basis with p. We

have

p = q r, q = r p, r = p q. (1.88)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

30 Continuum Mechanics

also have ( p, W q) = ( p, Wr ) = 0. Thus, using W = Wij ei e j referred to the

basis { p, q, r }, we get

W = (r q q r ),

arbitrary vector. Then

= [(u q)(r p q) (u r )(q + p r )]

= 0,

where the last step follows from Eqn. (1.88). Thus, associated with every skew-

symmetric tensor W, there is a vector w such that Eqn. (1.87) holds.

Conversely, given w, let q and r be unit vectors such that w/ |w|, q and r form

an orthonormal basis. Then

W = | w | (r q q r ), (1.89)

checked easily by verifying that W u = w u for any arbitrary vector u.

Given W, uniqueness of the axial vector w can be shown by assuming the

existence of two vectors w1 and w2 . Now we have W u = w1 u and W u =

w2 u for all vectors u. Hence,

(w1 w2 ) u = 0 u V.

w1 w2 , so that (w1 w2 ) u is nonzero. But this leads to a contradiction

with the above property. Hence w1 = w2 , and the uniqueness of the axial vector

is established.

Similarly, given w, uniqueness of the corresponding W can be established by

assuming the existence of two tensors W 1 and W 2 , such that W 1 u = w u and

W 2 u = w u for all vectors u. Now we have

(W 1 W 2 )u = 0 u V,

0, while conversely, if W u = w u = 0, then u is a scalar multiple of w by Theorem 1.2.1.

This result justifies the use of the terminology axial vector used for w. Also note that by

virtue of the uniqueness of w, the vector w, <, is a one-dimensional subspace of V.

Similarly, W 2 u = 0 if and only if u = w, <, since if W 2 u = 0, by the above result,

W u = w u = w, which is only possible when u = w and = 0.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 31

By choosing u = e j and taking the dot product of both sides with ei , Eqn. (1.87) can be

expressed in component form as

Wij = eijk wk ,

1 (1.90)

wi = eijk Wjk .

2

0 w3 w2

W = w3 0 w1 .

w2 w1 0

tr W T = tr W and det W = det W T = det W, we have tr W = det W = 0. The second

invariant is given by I2 = [(tr W )2 tr (W 2 )]/2 = (W : W )/2 = w w. Thus, from the

characteristic equation, we get the eigenvalues of W as (0, i |w| , i |w|), and the spectral

resolution as

W = i |w| (n n n n),

where n is the eigenvector corresponding to i |w|, and n is its complex conjugate. Note

that we have used the convention outlined at the end of Section 1.3.1 in writing the above

result.

A second-order tensor Q is said to be orthogonal if Q T = Q1 , or, alternatively by

Eqn. (1.67), if

Q T Q = QQ T = I, (1.91)

Theorem 1.5.1. A tensor Q is orthogonal if and only if it has any of the following

properties of preserving inner products, lengths and distances:

| Qu| = |u| u V, (1.92b)

| Qu Qv| = |u v| u, v V. (1.92c)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

32 Continuum Mechanics

By choosing v = u in Eqn. (1.92a), we get Eqn. (1.92b). Conversely, if

Eqn. (1.92b) holds, i.e., if ( Qu, Qu) = (u, u) for all u V, then

(( Q T Q I )u, u) = 0 u V,

By replacing u by (u v) in Eqn. (1.92b), we obtain Eqn. (1.92c), and, con-

versely, by setting v to zero in Eqn. (1.92c), we get Eqn. (1.92b).

As a corollary of the above results, it follows that the angle between two vectors u

and v, defined by := cos1 (u v)/(|u| |v|), is also preserved. Thus, physically speaking,

multiplying the position vectors of all points in a domain by Q corresponds to rigid body

rotation of the domain about the origin.

From Eqns. (1.49), (1.51) and (1.91), we have det Q = 1. Orthogonal tensors with de-

terminant +1 are said to be proper orthogonal or rotations (henceforth, this set is denoted

by Orth+ ). Since det( R) = (1)3 det R = 1 for R Orth+ , R is an orthogonal tensor

that is not a rotation. On the other hand, if Q is an orthogonal tensor that is not a rota-

tion, Q is a rotation. Thus, when the underlying vector space dimension is odd (which

includes n = 3), every orthogonal tensor is either a rotation, or the product of a rotation

with I:

Orth = { R; R Orth+ }.

0 1

This result is not true when the dimension n is even, as the example Q = 10 , which

cannot be written as R, where R Orth+ , shows. For Q Orth+ , we have

Theorem 1.5.2. Let {e1 , e2 , e3 } and {e1 , e2 , e3 } be two orthonormal bases. Then

Q = e1 e1 + e2 e2 + e3 e3 ,

is a proper orthogonal tensor. Conversely, every Q Orth+ can be represented in the

above manner.

Proof. If {e1 , e2 , e3 } and {e1 , e2 , e3 } are two orthonormal bases, then

QQ T = [e1 e1 + e2 e2 + e3 e3 ] [e1 e1 + e2 e2 + e3 e3 ]

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 33

Q = QI

= Q [e1 e1 + e2 e2 + e3 e3 ] (by Eqn. (1.38))

= e1 e1 + e2 e2 + e3 e3 ,

{e1 ,e2 , e3 } = { Qe1 , Qe2 , Qe3 } is an orthonormal basis since |ei | =

where

Qe = e = 1 for each i, and since, by virtue of Eqn. (1.94),

i i

orthogonal tensors, i.e.,

Lin = span{Orth+ }.

Proof. Let {e1 , e2 , e3 } denote the canonical Cartesian basis vectors. Then

1

e1 e1 = [ I + (e1 e1 e2 e2 e3 e3 )] ,

2

1

e1 e2 = [(e e2 + e2 e3 + e3 e1 ) + (e1 e2 e2 e3 e3 e1 )] ,

2 1

with similar expressions for the remaining combinations of dyadic products.

Thus, the tensor T = Tij ei e j can be expressed as i i Qi , where Qi

Orth+ .

If {ei } and {ei } are two sets of orthonormal basis vectors, then they are related as

ei = Q T ei , i = 1, 2, 3, (1.95)

where Q = ek ek is a proper orthogonal tensor by virtue of Theorem 1.5.2. The compo-

nents of Q with respect to the {ei } basis are given by Qij = ei (ek ek )e j = ik ek e j =

ei e j . Thus, if e and e are two unit vectors, we can always find Q Orth+ (not necessarily

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

34 Continuum Mechanics

unique), which rotates e to e, i.e., e = Qe. Let u and v be two vectors. Since u/ |u| and

v/ |v| are unit vectors, there exists Q Orth+ such that

v u

=Q .

|v| |u|

Thus, if u and v have the same magnitude, i.e., if |u| = |v|, then there exists Q Orth+

such that u = Qv.

We now study the transformation laws for the components of tensors under an or-

thogonal transformation of the basis vectors. Let ei and ei represent the original and new

orthonormal basis vectors, and let Q be the proper orthogonal tensor in Eqn. (1.95). From

Eqn. (1.9), we have

ei = (ei e j )e j = Q ji e j . (1.96b)

Using Eqn. (1.8) and Eqn. (1.96a), we get the transformation law for the components of a

vector as

In a similar fashion, using Eqn. (1.28), Eqn. (1.96a), and the fact that a tensor is a linear

transformation, we get the transformation law for the components of a second-order tensor

as

Conversely, if the components of a matrix transform according to Eqn. (1.98), then they all

generate the same tensor. To see this, let T = Tij ei e j and T = Tmn em en . Then

T = Tij ei e j

= Qim Q jn Tmn ei e j

= Tmn ( Qim ei ) ( Q jn e j )

= Tmn em en (by Eqn. (1.96b))

= T.

Due to this property, often, an alternate viewpoint is followed by many authors who take

the transformation law given by Eqn. (1.98) as the definition of second-order tensors.

We can write Eqns. (1.97) and (1.98) as

T

[ T ] = Q[ T ] Q . (1.100)

where [v] and [ T ] represent the components of the vector v and tensor T, respectively, with

respect to the ei coordinate system. Using the orthogonality property of Q, we can write

the reverse transformations as

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 35

e2

e1

e2

e1

Fig. 1.1 Example of a coordinate system obtained from an existing one by a rotation about the

3-axis.

[ T ] = Q T [ T ] Q. (1.102)

e1 cos sin 0

Q = e2 = sin cos 0 .

e3 0 0 1

6

The only real eigenvalues of Q Orth can be either +1 or 1, since if and n denote

the eigenvalue and eigenvector of Q, i.e., Qn = n, then

which implies that (n, n)(2 1) = 0. If and n are real, then (n, n) 6= 0 and = 1,

while if is complex, then (n, n) = 0. Let and n denote the complex conjugates of and

n, respectively. To see that the complex eigenvalues have a magnitude of unity observe

that (n, n) = ( Qn, Qn) = (n, n), which implies that = 1 since (n, n) 6= 0. Now,

using a similar argument as used in showing (n, n) = 0, we also have (n1 , n2 ) = 0 for n1

and n2 corresponding to distinct complex eigenvalues of Q (complex conjugates not being

6

considered as distinctthus, this result is relevant only when the dimension n > 3), and

(n, e) = (n, e) = 0, where e is the eigenvector of Q corresponding to the eigenvalue of +1

or 1.

If R 6= I is a rotation, then the set of all vectors e such that

Re = e (1.103)

forms a one-dimensional subspace of V called the axis of R. To prove that such a vector

always exists, we first show that +1 is always an eigenvalue of R, and 1 is always an

eigenvalue of an improper rotation. Since det R = 1,

= det( I R) = det( R I ),

sponding to the eigenvalue +1, then Re = e. Since every improper rotation is a product of

I times a proper orthogonal tensor, 1 is an eigenvalue of an improper orthogonal ten-

sor, with the same eigenvector as the proper orthogonal tensor from which it is generated.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

36 Continuum Mechanics

To show that the set of vectors e satisfying Eqn. (1.103) is a one-dimensional subspace of V,

premultiply both sides of this equation with R T to get R T e = e. Combining this relation

with Eqn. (1.103), we get ( R R T )e = 0. There are two possibilities: (i) R is unsymmetric,

or (ii) R is symmetric. If R is unsymmetric, then R R T is a nonzero skew-symmetric ten-

sor with e a scalar multiple of the axial vector of R R T . Since we have shown that a scalar

multiple of the axial vector of a skew-symmetric tensor constitutes a one-dimensional sub-

space, e, < is a one-dimensional subspace of V. Now consider the other possibility

where R is symmetric. By Theorem 1.6.1 all the eigenvalues of R are real, and as shown

above the only real eigenvalues can be 1. Taking into account that det R = 1, R 6= I

and that one of the real eigenvalues is +1, the set of eigenvalues in this case has to be

{+1, 1, 1}. Since the eigenvalue +1 is distinct from the other two, again invoking The-

orem 1.6.1, it follows that the corresponding eigenvector e is unique. Using Eqn.(1.123), it

also follows that R Orth+ Sym { I } has to be of the form 2e e I, where e is a unit

vector.

Conversely, given a vector w, there exists a proper orthogonal tensor R, and an im-

proper one R0 such that Rw = w and R0 w = w. To see this, consider the family of

tensors

1 1

R(w, ) = I + sin W + (1 cos )W 2 , (1.104)

|w| | w |2

where W is the skew-symmetric tensor with w as its axial vector, i.e., W w = 0. Using the

CayleyHamilton theorem, we have W 3 = |w|2 W, from which it follows that W 4 =

|w|2 W 2 . Using this result, we get

" #" #

T sin (1 cos ) 2 sin (1 cos ) 2

R R= I W+ W I+ W+ W = I.

|w| | w |2 |w| | w |2

Since R has now been shown to be orthogonal, det R = 1. However, since det[ R(w, 0)] =

det I = 1, by continuity, we have det[ R(w, )] = 1 for any . Thus, R is a proper orthogonal

tensor that satisfies Rw = w. Taking R0 = R, we get the required improper orthogonal

tensor. It is easily seen that Rw = w. Essentially, R rotates any vector in the plane per-

pendicular to w through an angle . By Eqn. (1.89), W 2 = w w |w|2 I. Thus, from

Eqn. (1.104), we get R(w, 0) = I and R(w, ) = 2e e I, where e = w/ |w|.

Using this result, we have the following representation theorem for proper orthogonal

tensors.

Theorem 1.5.4. Let W be a skew-symmetric tensor with an axial vector of unit magni-

tude, and let {e, q, r } be an orthonormal set of vectors. Then tensors of the form

are proper orthogonal. Conversely, there exist W Skw, and orthonormal vectors

{e, q, r }, such that every proper orthogonal tensor can be expressed in the above forms.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 37

Proof. The tensor R in Eqn. (1.105a) is just a special case of the tensor R in

Eqn. (1.104), which we have already shown to be proper orthogonal. Let

{e, q, r } be an orthonormal set of vectors. Choose W = r q q r, so that

W 2 = e e I. Substituting for W and W 2 into Eqn. (1.105a), we get the repre-

sentation in Eqn. (1.105b). Note that since Re = e, the vector e is the axis of R.

Conversely, we now show that every proper orthogonal tensor can be repre-

sented as Eqns. (1.105a) and (1.105b). Let e be the axis of R, and let q, r be such

that {e, q, r } form an orthonormal basis. Using the fact that Re = R T e = e, we

have ( Re, q) = ( Re, r ) = (e, Rq) = (e, Rr ) = 0. Thus, Rq and Rr lie in the plane

perpendicular to e. If ( Rq, q) = ( Rr, r ) = cos , then

R = RI

= R(e e + q q + r r )

= ( Re) e + ( Rq) q + ( Rr ) r

= e e + [( Rq, e)e + ( Rq, q)q + ( Rq, r )r ] q+

[( Rr, e)e + ( Rr, q)q + ( Rr, r )r ] r

= e e + cos (q q + r r ) + sin (r q q r )

= e e + cos ( I e e) + sin (r q q r ),

from which we get Eqn. (1.105a).

Using the results for the eigenvalues/eigenvectors of R, it is also possible to write the

following spectral resolutions (using the convention outlined at the end of Section 1.3.1):

I = e e + n n + n n,

R = e e + n n + n n, (1.106)

2 2 2

R = e e + n n + n n,

where = = 1 in case all eigenvalues of R 6= I are real (so that R = 2e e I), and

n n = n n = e n = e n = 0 and n n = 1 in case is complex.

Note from Eqn. (1.105b) that tr R = 1 + 2 cos . It is easily seen by multiplying the

expression

1 e e + 2 (q q + r r ) + 3 (r q q r ) = 0

successively by e, q and r that 1 = 2 = 3 = 0, which implies that {e e, q q + r

r, r q q r } are linearly independent. Using the representation given by Eqn. (1.105b),

we have

I 1 1 0 ee

R = 1 cos sin q q + r r .

The determinant of the square matrix in the above equation is 2 sin (1 cos ). Thus, from

Theorem 1.1.2 it follows that { I, R, R2 } are linearly independent for all except when (i)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

38 Continuum Mechanics

R2 = I (in this case { I, R} are linearly independent).

Using the representation in Eqn. (1.104), let

1 1

R1 ( w1 , ) = I + sin W 1 + (1 cos )W 21 ,

| w1 | | w1 |2

1 1

R2 ( w2 , ) = I + sin W 2 + (1 cos )W 22 ,

| w2 | | w2 |2

be two proper orthogonal tensors. If the axes of these two tensors coincide, i.e., if e1 =

w1 / |w1 | = w2 / |w2 | = e2 , then we have W 1 / |w1 | = W 2 / |w2 |, and hence R1 and R2

commute, i.e., R1 R2 = R2 R1 . However, the converse is not true since, for example, R1 =

diag[1, 1, 1] and R2 = diag[1, 1, 1] commute, but do not have the same axes. If e1

and e2 are the axes of R1 and R2 , respectively, then we assume the (unnormalized) axis

of R2 R1 to be of the form w = e1 + e2 + e1 e2 (since {e1 , e2 , e1 e2 } constitute

a basis when e1 and e2 are linearly independent) and find (, , ) using the condition

R2 R1 w = w. Using this procedure, we obtain the axis of R2 R1 as e = w/ |w|, where

of rigid-body dynamics, we shall need to consider the case when the angular velocity vec-

tor is a function of time, and which, in general, will not coincide with the axis of Q. In

such a situation, it is helpful to express Q Orth+ in terms of generalized coordinates.

Problem 26 shows that 3 parameters are required to characterize an orthogonal tensor in

three-dimensional space, and presents one such characterization. However, the most com-

monly used characterization is the one using the Euler angles , and . The representation

of Q in terms of these angles is

cos sin 0 1 0 0 cos sin 0

Q = sin cos 0 0 cos sin sin cos 0

0 0 1 0 sin cos 0 0 1

cos cos cos sin sin sin cos cos sin cos sin sin

= cos sin + cos cos sin sin sin + cos cos cos sin cos . (1.108)

(1 cos )(sin + sin )

e = (1 cos )(cos cos ) ( 6 = 0). (1.109)

sin [1 cos( + )]

We now present an expression for the eigenvalues of Q. Since from Eqn. (1.105b),

tr Q = 1 + 2 cos , and since we have shown that the magnitude of the eigenvalues is 1, the

eigenvalues of Q Orth+ are {1, ei , ei }. Since 1 cos 1, we have 1 tr Q 3,

with tr Q = 3 if and only if Q = I ( = 0 in Eqn. (1.105b)). The other bound tr Q = 1 is at-

tained when = (or when or + is equal to in Eqn. (1.108)), and from Eqn. (1.105b),

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 39

we see that Q = 2e e I Sym with eigenvalues {1, 1, 1}. When 1 < tr Q < 3,

two eigenvalues and the corresponding eigenvectors are complex.

Since for Q Orth+ , I2 = tr (cof Q) = tr [(det Q) QT ] = tr Q, by the Cayley

Hamilton theorem, we get

Q1 = a1 b1 + a2 b2 + a3 b3 ,

Q2 = c1 d1 + c2 d2 + c3 d3 ,

where { ai }, {bi }, {ci }, {di } are orthonormal sets of vectors, then there exist proper orthog-

onal tensors R1 = ck ak and R2 = dk bk such that Q2 = R1 Q1 R2T . However, if the

trace of Q1 and Q2 are the same (which implies that all their principal invariants are also

the same), the following stronger result holds.

Theorem 1.5.5. Two proper orthogonal tensors Q1 and Q2 have the same trace, i.e.,

tr Q1 = tr Q2 , if and only if there exists an orthogonal tensor Q0 (which means that

either Q0 or Q0 is a rotation) such that Q2 = Q0 Q1 Q0T .

Proof. If there exists Q0 Orth such that Q2 = Q0 Q1 Q0T , then using the same

method of proof as used to prove Eqn. (1.137), we see that all the principal in-

variants of Q1 and Q2 are the same. To prove the converse, note that if all the

principal invariants of Q1 and Q2 are the same, then their characteristic equa-

tions, and hence their eigenvalues are the same. Thus, using Eqn. (1.106), we can

write

Q1 = e e + n n + n n,

Q2 = f f + g g + g g.

The tensor

Q0 = f e + g n + g n

following Eqn. (1.106), we also see that Q2 = Q0 Q1 Q0T . In the above treatment,

we have assumed that two eigenvalues are complex. If (1, 1, 1) are the eigen-

values of Q1 and Q2 , then they have the representations Q1 = 2e e I and

Q2 = 2 f f I, so that any tensor that rotates e into f can be taken to be Q0 .

An alternative proof can also be given using Eqn. (1.105b) (recall that tr R =

1 + 2 cos ) by writing

Q1 = e e + cos ( I e e) + sin (r q q r ),

Q2 = f f + cos ( I f f ) sin (r q q r ),

where {e, q, r } and { f , q, r } are orthonormal sets of axes, and then taking Q0 as

f e + q q + r r.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

40 Continuum Mechanics

In this section, we examine some properties of symmetric second-order tensors. We first

discuss the properties of the principal values (eigenvalues) and principal directions (eigen-

vectors) of a symmetric second-order tensor.

We have the following result:

Theorem 1.6.1. Every symmetric tensor S has at least one principal frame, i.e., a right-

handed triplet of orthogonal principal directions, and at most three distinct principal

values. The principal values are always real. For the principal directions three possibili-

ties exist:

If all the three principal values are distinct, the principal axes are unique (modulo

sign reversal).

If two eigenvalues are equal, then there is one unique principal direction, and the

remaining two principal directions can be chosen arbitrarily in the plane perpen-

dicular to the first one, and mutually perpendicular to each other.

If all three eigenvalues are the same, then every right-handed frame is a principal

frame, and S is of the form S = I.

The components of the tensor in the principal frame are

1 0 0

S = 0 2 0 . (1.111)

0 0 3

(S I )n = 0. (1.112)

But this is nothing but an eigenvalue problem. For a nontrivial solution, we need

to satisfy the condition that

det(S I ) = 0,

or, by Eqn. (1.79),

3 I1 2 + I2 I3 = 0, (1.113)

where I1 , I2 and I3 are the principal invariants of S.

We now show that the principal values given by the three roots of the cubic

equation Eqn. (1.113) are real. Suppose that two roots, and hence the eigenvec-

tors associated with them, are complex. Denoting the complex conjugates of

and n by and n, we have

Sn = n, (1.114a)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 41

Sn = n, (1.114b)

where Eqn. (1.114b) is obtained by taking the complex conjugate of Eqn. (1.114a)

(S being a real matrix is not affected). Taking the dot product of both sides of

Eqn. (1.114a) with n, and of both sides of Eqn. (1.114b) with n, we get

n Sn = n n, (1.115)

n Sn = n n. (1.116)

Using the definition of a transpose of a tensor, and subtracting the second rela-

tion from the first, we get

S T n n n Sn = ( )n n. (1.117)

( )n n = 0.

The principal directions n1 , n2 and n3 , corresponding to distinct eigenvalues

1 , 2 and 3 , are mutually orthogonal and unique (modulo sign reversal). We

now prove this. Taking the dot product of

Sn1 = 1 n1 , (1.118)

Sn2 = 2 n2 , (1.119)

0 = ( 1 2 ) n1 n2 ,

where we have used the fact that S being symmetric, n2 Sn1 n1 Sn2 = 0.

Thus, since we assumed that 1 6= 2 , we get n1 n2 . Similarly, we have

n2 n3 and n1 n3 . If n1 satisfies Sn1 = 1 n1 , then we see that n1 also

satisfies the same equation. This is the only other choice possible that satisfies

Sn1 = 1 n1 . To see this, let r 1 , r 2 and r 3 be another set of mutually perpendicular

eigenvectors corresponding to the distinct eigenvalues 1 , 2 and 3 . Then r 1

has to be perpendicular to not only r 2 and r 3 , but to n2 and n3 as well. Similar

comments apply to r 2 and r 3 . This is only possible when r 1 = n1 , r 2 = n2

and r 3 = n3 . Thus, the principal axes are unique modulo sign reversal.

To prove that the components of S in the principal frame are given by

Eqn. (1.111), assume that n1 , n2 , n3 have been normalized to unit length, and

then let e1 = n1 , e2 = n2 and e3 = n3 . Using Eqn. (1.28), and taking into account

the orthonormality of e1 and e2 , the components S11 and S are given by

12

S11 = e1 Se1 = e1 (1 e1 ) = 1 ,

S12 = e1 Se2 = e1 (2 e2 ) = 0.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

42 Continuum Mechanics

Similarly, on computing the other components, we see that the matrix represen-

tation of S with respect to e is given by Eqn. (1.111).

If there are two repeated roots, say, 2 = 3 , and the third root 1 6= 2 , then

let e1 coincide with n1 , so that Se1 = 1 e1 . Choose e2 and e3 such that e1 -e3

form a right-handed orthogonal coordinate system. The components of S with

respect to this coordinate system are

1 0 0

S = 0 S22 .

S23 (1.120)

0 S23

S33

S22 + S33 = 22 ,

2

1 S22 + (S22 S33 (S23 ) ) + 1 S33 = 21 2 + 22 , (1.121)

h i

2

1 S22 S33 (S23 ) = 1 22 .

Substituting for 2 from the first equation into the second, we get

2 2

(S22 S33 ) = 4(S23 ) .

Since the components of S are real, the above equation implies that S23 = 0 and

2 = S22 = S33 . This shows that Eqn. (1.120) reduces to Eqn. (1.111), and that

Se2 = S12 e + S e + S e = e and Se = e (thus, e and e are eigen-

1 22 2 32 3 2 2 3 2 3 2 3

vectors corresponding to the eigenvalue 2 ). However, in this case the choice of

the principal frame e is not unique, since any vector lying in the plane of e2 and

e3 , given by n = c1 e2 + c2 e3 where c1 and c2 are arbitrary constants, is also an

eigenvector. The choice of e1 is unique (modulo sign reversal), since it has to be

perpendicular to e2 and e3 . Though the choice of e2 and e3 is not unique, we

can choose e2 and e3 arbitrarily in the plane perpendicular to e1 , and such that

e2 e3 .

Finally, if 1 = 2 = 3 = , then the tensor S is of the form S = I. To

show this choose e1 e3 , and follow a procedure analogous to that in the previous

case. We now get S22 = S = and S = 0, so that [ S ] = I. Using the

33 23

transformation law for second-order tensors, we have

[S] = Q T [S ] Q = Q T Q = I.

handed frame is a principal frame.

S = 1 e1 e1 + 2 e2 e2 + 3 e3 e3 , (1.122)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 43

which is called as the spectral resolution of S. The spectral resolution of S is unique since

If all the eigenvalues are distinct, then the eigenvectors are unique, and consequently

the representation given by Eqn. (1.122) is unique.

If two eigenvalues are repeated, then, by virtue of Eqn. (1.38), Eqn. (1.122) reduces to

S = 1 e1 e1 + 2 e2 e2 + 2 e3 e3

= 1 e1 e1 + 2 ( I e1 e1 ), (1.123)

If all the eigenvalues are the same then S = I.

Although the eigenvalues of S Sym are real, its eigenvectors can be complex (note

that in this case, real eigenvectors, obtained by combining the complex eigenvectors, also

exist).

For example, if S = diag [ 1, 2, 3 ] , then n = ( 1 + i )[ 1, 0, 0 ] / 2 and n = (1 i )[1, 0, 0]/

2 are eigenvectors corresponding

to the eigenvalue 1. If S = diag[1, 1, 2], then n =

[1, i, 0]/ 2 and n = [1, i, 0]/ 2 are eigenvectors corresponding to the repeated eigen-

value 1. Note that the eigenvectors in such a case occur in complex conjugate pairs (which

is proved simply by taking the complex conjugate of the relation Sn = n, so that (n, n)

are the eigenvectors corresponding to the eigenvalue ). Note, however, that although the

eigenvectors can be complex, the eigenprojections Pi as given by Eqn. (J.4) are real.

Theorem 1.6.2. Let S be given by Eqn. (1.122), and let {ek } denote the canonical basis

of V. Then Q = ek ek is a proper orthogonal tensor that diagonalizes S, i.e., the

matrix representation of QSQ T with respect to the canonical basis is diagonal.

Conversely, if Q Orth+ diagonalizes a tensor S, i.e., QSQ T = where is a

diagonal matrix, then S is symmetric, the diagonal matrix comprises the eigenvalues

of S, and the rows of Q are the eigenvectors of S.

Proof. By Theorem 1.5.2, Q = ek ek Orth+ , and since ei = Qei , we have

" #

QSQ T = Q i ei ei QT

i

= i ( Qei ) ( Qei )

i

= i ei ei ,

i

Conversely, if QSQ T = where is a diagonal matrix, then S = Q T Q is

clearly a symmetric tensor. By Theorem 1.6.10, S and have the same eigenval-

ues, and hence has the eigenvalues of S along the diagonal. If follows that

" #

S = Q T Q = Q T i ei ei Q= i ( Q T e i ) ( Q T e i ),

i i

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

44 Continuum Mechanics

which implies that S( Q T ei ) = i ( Q T ei ) (no sum on i); thus, Q T ei are the eigen-

vectors of S. Since Q = IQ, we have

Q = (ek ek ) Q = ek ( Q T ek ) = ek ek ,

S3 I1 S2 + I2 S I3 I = 0. (1.124)

We have already proved this result for any arbitrary tensor. However, the following sim-

pler proof can be given for symmetric tensors. Using Eqn. (1.341), the spectral resolutions

of S, S2 and S3 are

S = 1 e1 e1 + 2 e2 e2 + 3 e3 e3 ,

S2 = 21 e1 e1 + 22 e2 e2 + 23 e3 e3 , (1.125)

3

S = 31 e1 e1 + 32 e2 e2 + 33 e3 e3 .

Substituting these expressions into the left-hand side of Eqn. (1.124), we get

+ (33 I1 23 + I2 3 I3 )(e3 e3 ) = 0,

since 3i I1 2i + I2 i I3 = 0 for i = 1, 2, 3.

{S I1 I/3 : S Sym} constitutes the set of traceless symmetric tensors, while tensors

of the form ( a b + b a)/2, a, b V have determinant zero (a fact that can be verified

by writing ( a b + b a)/2 as ( a b + b a + 0 0)/2 and then using Eqn. (1.53)).

However, not all symmetric tensors with determinant zero can be represented as ( a b +

b a)/2 since, from the CauchySchwartz inequality it follows that the second invariant

of ( a b + b a)/2 computed using Eqn. (1.63) is always negative.

The following explicit expressions can be given for the eigenvalues of S Sym when it

is not of the form I [298]:

I1

1 = + 2 p cos ,

3

I

2 = 1 p(cos + 3 sin ),

3

I1

3 = p(cos 3 sin ),

3

where

1 I I

p= S 1I : S 1I ,

6 3 3

1 I

q = det S 1 I ,

2 3

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 45

p

1 p3 q2

= tan1 , 0 .

3 q

Alternative expressions for finding the eigenvalues are

I1 p

1 = 2 Q cos ,

3

I 2

2 = 1 2 Q cos +

p

, (1.126)

3 3

I1 p 2

3 = 2 Q cos ,

3 3

where

I12 3I2

Q= ,

9

1 h i

U= 9I1 I2 2I13 27I3 ,

54

!

1 U

= cos1 p .

3 Q3

If Q = 0 (i.e., if I12 = 3I2 ), then all the eigenvalues are repeated, 1 = 2 = 3 = I1 /3. If

Q 6= 0 and Q3 = U 2 (i.e., if I12 I22 4I23 4I13 I3 + 18I1 I2 I3 27I32 = 0), then 1 = I1 /3 2 Q

and 2 = 3 = I1 /3 + Q if U > 0, and 2 = I1 /3 + 2 Q and 1 = 3 = I1 /3 Q if

U < 0. If Q 6= 0 and Q3 6= U 2 , then the eigenvalues are distinct.

We saw in Theorem 1.6.1 that the component form of a symmetric tensor is a diagonal

matrix in a particular basis. The following theorem states the conditions under which the

matrix representation has all the diagonal terms zero [18], [27], [117]:

if and only if tr S = 0.

Proof. If Eqn. (1.127) holds then clearly tr S = n1 Sn1 + n2 Sn2 + n3 Sn3 = 0.

Thus, we just need to prove the converse. If tr S = 1 + 2 + 3 = 0, then the

spectral resolution of S can be written as

S = 1 (e1 e1 e3 e3 ) + 2 (e2 e2 e3 e3 ).

basis, and we assume that such a choice has been made. Note that when tr S = 0,

at most one eigenvalue can be zero, since if two or more are zero, the tensor S

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

46 Continuum Mechanics

is the zero tensor. Thus, we can assume, without loss in generality, that both 1

and 2 are nonzero, and 3 (1 + 2 ) can be either zero or nonzero. If we

choose

n1 = (e1 + e2 + e3 )/ 3, (1.128)

it is clear that n1 Sn1 = 0. We now try and find n2 and n3 in the plane perpen-

dicular to n1 such that n2 Sn2 = n3 Sn3 = 0. Since n1 n2 = 0, n2 and n3 are of

the form

1 e1 + 2 e2 (1 + 2 )e3

n2 = q ,

12 + 22 + (1 + 2 )2

(1.129)

(1 + 22 )e1 + (21 + 2 )e2 + (2 1 )e3

n3 = n1 n2 = q 2 ,

3 1 + 22 + (1 + 2 )2

and n3 Sn3 = 0 yield the equation

1 22 + 2(1 + 2 )1 2 + 2 12 = 0,

q

(1 + 2 ) 21 + 1 2 + 22

2 = 1 .

1

It suffices to consider only the solution with the positive sign, since it can be

shown that the solution obtained with the negative sign is simply a rearrange-

ment of the one obtained with the positive one, i.e., if (n1 , n2 , n3 ) is the solution

obtained with the positive sign, then (n1 , n3 , n2 ) is the solution obtained with

the negative one. Substituting for 2 into Eqns. (1.129), we get

g1 e1 + g2 e2 + g3 e3

n2 = q ,

g12 + g22 + g32

(1.130)

h1 e1 + h2 e2 + h3 e3

n3 = q ,

h21 + h22 + h23

where

g1 = 1 ,

q

g2 = 21 + 1 2 + 22 (1 + 2 ),

q

g3 = 2 21 + 1 2 + 22 ,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 47

q

h1 = 1 + 22 2 21 + 1 2 + 22 ,

q

h2 = 1 2 + 21 + 1 2 + 22 ,

q

h3 = 21 2 + 21 + 1 2 + 22 .

The solution that has been presented above is only one choice in an infinite

one-parameter family of solutions in terms of a parameter . To find this infinite

one-parameter family, assume a unit vector in the principal basis {ei } to be given

by

and also assume without loss in generality that 1 and 2 are chosen such

that 1 6= 2 . Using the fact that S in this coordinate system is given by

diag[1 , 2 , (1 + 2 )], and eliminating by using the condition n1 Sn1 = 0

yields

q 2

(1 +221)cos

2

2

q

n1 ( ) = (2 +21 ) cos2 1 .

2 1

cos

The second vector n2 also satisfies n2 Sn2 = 0, and hence has to be of the form

n1 (), where is found in terms of by imposing the condition that n1 n2 = 0.

Denoting t cos2 , we get

h h2

cos2 = 1 , (1.131)

D

where

Denoting the two roots for cos2 corresponding to the plus and minus signs

in Eqn. (1.131) by cos2 1 and cos2 2 , we get n2 = n1 (1 ) and n3 = n1 (2 ).

The plus or minus signs for the components of n1 can be chosen arbitrarily; the

corresponding signs for n2 are obtained by imposing the condition n1 n2 = 0,

and then n3 is found as n1 n2 .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

48 Continuum Mechanics

The result in Theorem 1.6.3 actually holds for an arbitrary traceless tensor T. This is

seen by splitting T into its symmetric and antisymmetric parts, and since the diagonal

elements of the antisymmetric part are zero in any basis, the basis found for the symmetric

part works. As a simple example of the above theorem, let the matrix representation of

S with respect to a basis be diag[1, 1, 0]. Then, by convention, we choose 1 and 2 to

be the nonzero eigenvalues, i.e., 1 = 1, 2 = 1. The corresponding eigenvectors are

e1 = (1, 0, 0), e2 = (0, 1, 0) and e3 = (0, 0, 1). From Eqns. (1.128) and (1.130), we get

n1 = (1, 1, 1)/ 3, n2 = (1, 1, 2)/ 6, and n3 = (1, 1, 0)/ 2. A more complicated

example is given in Section 3.7.

The decomposition of a matrix into two symmetric components, i.e., T = S1 S2 , where

S1 , S2 Sym, and where one of the factors can be taken to be invertible [26], can be found

as follows. Assuming S2 to be invertible, we get S1 = TS21 . Since S1 Sym, we get

S21 T T = TS21 , or, alternatively, T T S2 = S2 T. This results in at most n(n 1)/2 indepen-

dent equations for the n(n + 1)/2 components of S2 . For example, for dimension n = 2,

we get

T12 T21 T11 T22 S11

0 0 0 S22 = 0, (1.132)

0 0 0 S12

T12 T21 0 T11 T22 T31 T32 S11

T T31 T23 T21 T11 T33

0 S22

13

0 T23 T32 T13 T22 T33 T12 S33

= 0.

0 0 0 0 0 0 S12

0 0 0 0 0 0 S23

0 0 0 0 0 0 S13

( I )( I ).

In general, possible to express the factors S1 and S2 as a function of T. To see

it is not

this, let T = 0w w0 , whose determinant is nonzero, so that both S1 and S2 are invertible.

Using Eqn. (1.132), we get the most general solution as

" # " #" #

0 w 1 bw aw a b

= 2 , a, b <.

w 0 a + b2 aw bw b a

A second-order symmetric tensor S is positive definite if

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 49

We denote the set of symmetric, positive definite tensors by Psym. Since by virtue of

Eqn. (1.33), all tensors T can be decomposed into a symmetric part T s , and a skew-symmetric

part T ss , we have

= (u, T s u),

because (u, T ss u) = 0 by Eqn. (1.86). Thus, the positive definiteness of a tensor is decided

by the positive definiteness of its symmetric part. In Theorem 1.6.4, we show that a sym-

metric tensor is positive definite if and only if its eigenvalues are positive. Although the

eigenvalues of the symmetric part of T should be positive in order for T to be positive def-

inite, positiveness of the eigenvalues of T itself does

not ensure its positive definiteness

as the following counterexample shows. If T = 10 110 , then T is not positive definite

since u Tu < 0 for u = (1, 1), but the eigenvalues of T are (1, 1). Conversely, if T is pos-

itive definite, then by choosing u to be the real eigenvectors n of T, it follows that its real

eigenvalues = (n Tn) are positive.

1. S is positive definite.

2. The principal values of S are strictly positive.

3. The principal invariants of S are strictly positive.

Proof. We first prove the equivalence of (1) and (2). Suppose S is positive defi-

nite. If and n denote the principal values and principal directions, respectively,

of S, then Sn = n, which implies that = (n, Sn) > 0 since n 6= 0.

Conversely, suppose that the principal values of S are greater than 0. Assum-

ing that e1 , e2 and e3 denote the principal axes, the representation of S in the

principal coordinate frame is (see Eqn. (1.122))

S = 1 e1 e1 + 2 e2 e2 + 3 e3 e3 .

Then

Su = (1 e1 e1 + 2 e2 e2 + 3 e3 e3 )u

= 1 (e1 u)e1 + 2 (e2 u)e2 + 3 (e3 u)e3

= 1 u1 e1 + 2 u2 e2 + 3 u3 e3 ,

and

which is greater than or equal to zero since i > 0. Suppose that (u, Su) = 0.

Then by Eqn. (1.133), ui = 0, which implies that u = 0. Thus, S is a positive

definite tensor.

To prove the equivalence of (2) and (3), note that, by Eqn. (1.80), if all

the principal values are strictly positive, then the principal invariants are also

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

50 Continuum Mechanics

strictly positive. Conversely, if all the principal invariants are positive, then

I3 = 1 2 3 , is positive, so that all the i are nonzero in addition to being real.

Each i has to satisfy the characteristic equation

3i I1 2i + I2 i I3 = 0, i = 1, 2, 3.

If i is negative, then, since I1 , I2 , I3 are positive, the left-hand side of the above

equation is negative, and hence the above equation cannot be satisfied. We have

already mentioned that i cannot be zero. Hence, each i has to be positive.

Theorem 1.6.5. If S Psym, then there exists a unique H Psym, such that H 2 :=

HH = S. The tensor H is called the positive definite square root of S, and we write

H = S.

Proof. Before we begin the proof, we note that a positive definite, symmet-

ric tensor can have square roots that are not positive definite. For example,

diag[1, 1, 1] is a non-positive definite square root of I. Here, we are interested

only in those square roots that are positive definite. The short proof of unique-

ness given below is due to Stephenson [305].

Since

3

S= i ei ei ,

i =1

is positive definite, by Theorem 1.6.4, all i > 0. Define

3

i ei ei .

p

H := (1.134)

i =1

Since the {ei } are orthonormal, it is easily seen that H H = S. Since the eigen-

values of H given by i are all positive, H is positive definite. Thus, we have

shown that a positive definite square root tensor of S given by Eqn. (1.134) exists.

We now prove uniqueness.

With and n denoting the principal value and principal direction of S, we

have

0 = (S I )n

= ( H 2 I )n

= ( H + I )( H I )n.

Calling ( H I )n = n, we have

( H + I )n = 0.

This implies that n = 0. For, if not, is a principal value of H, which contra-

dicts the fact that H is positive definite. Therefore,

( H I )n = 0;

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 51

i.e., n

is also a principal direction of H with associated principal values . If

H = S, then it must have the form given by Eqn. (1.134) (since the spectral

decomposition is unique), which establishes its uniqueness.

Now we prove the polar decomposition theorem, which plays a crucial role in the char-

acterization of frame-indifferent constitutive relations.

it can be factored in a unique fashion as

F = RU = V R,

where R is an orthogonal tensor, and U, V are symmetric and positive definite tensors.

One has

p

U = FT F

p

V = FF T .

Proof. The tensor F T F is obviously symmetric. It is positive definite since

(u, F T Fu) = ( Fu, Fu) 0,

only if u = 0 (Fu = 0 implies that u = 0, since F is in-

with equality if and

vertible). Let U = F T F. U is unique, symmetric and positive definite by

Theorem 1.6.5. Define R = FU 1 , so that F = RU. The tensor R is orthogonal,

because

R T R = ( FU 1 ) T ( FU 1 )

= U T F T FU 1

= U 1 ( F T F )U 1 (since U is symmetric)

1 1

=U (UU )U

= I.

Since det U > 0, we have det U 1 > 0. Hence, det R and det F have the same

sign. Usually, the polar decomposition theorem is applied to the deformation

gradient F satisfying det F > 0. In such a case det R = 1, and R is a rotation.

Next, let V = FU F 1 = FR1 = RU R1 = RU R T . Thus, V is symmetric

since U is symmetric. V is positive definite since

(u, V u) = (u, RU R T u)

= ( R T u, U R T u)

0,

T T

with equality if and only if u = 0 (again since R is invertible). Note that FF =

2 T

V V = V , so that V = FF .

Finally, to prove the uniqueness of the polar decomposition, we note that since

U is unique, R = FU 1 is unique, and hence so is V .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

52 Continuum Mechanics

i ( Rei ), the pairs (i , f i ), where f i Rei are the eigenvalues/eigenvectors of V . Thus, F

and R can be represented as

3 3 3

F = RU = R i ei ei = i ( Rei ) ei = i f i ei ,

i =1 i =1 i =1

(1.135)

3 3

R = RI = R ei ei = ( Rei ) ei = f i ei .

i =1 i i =1

When F is invertible, the polar decomposition can be given in explicit form in terms of

the eigenvalues {1 , 2 , 3 } of U, and powers of C or B. First, one determines the eigen-

values {21 , 22 , 23 } of C = F T F (which are also the eigenvalues of B = FF T = RCR T ),

using Eqn. (1.126). Then, depending on whether the eigenvalues are repeated or distinct,

the relevant expressions for the component matrices in the polar decomposition are ([126],

[153], [326])

All eigenvalues distinct (1 6= 2 6= 3 6= 1 ):

ha a2 a3 i

R = F 1 I + C + C2 ,

a a a

a a2 a3

U 1 = 1 I + C + C 2 ,

a a a

1 a1 a2 a3 2

V = I+ B+ B ,

a a a

b1 b2 b3 2

U = I+ C+ C ,

b b b

b1 b2 b3 2

V = I+ B+ B ,

b b b

where

a1 = (1 + 2 + 3 )(21 22 + 22 23 + 21 23 ) + 1 2 3 (21 + 22 + 23 + 1 2

+ 2 3 + 1 3 ),

h i

a2 = (1 + 2 + 3 )(21 + 22 + 23 ) + 1 2 3 ,

a3 = 1 + 2 + 3 ,

a = 1 2 3 [(1 + 2 + 3 )(1 2 + 2 3 + 1 3 ) 1 2 3 ] ,

b1 = 1 2 3 (1 + 2 + 3 ),

b2 = 21 + 22 + 23 + 1 2 + 2 3 + 1 3 ,

b3 = 1,

b = (1 + 2 + 3 )(1 2 + 2 3 + 1 3 ) 1 2 3 .

ha a2 i

R = F 1I+ C ,

a a

1 a1 a2

U = I + C,

a a

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 53

a1 a2

V 1 = I+ B,

a a

b b2

U = 1I+ C,

b b

b1 b2

V = I+ B,

b b

where

a1 = 21 + 1 2 + 22 ,

a2 = 1,

a = 1 2 ( 1 + 2 ),

b1 = 1 2 ,

b2 = 1,

b = 1 + 2 .

1

R= F, U = V = I.

Rates of the stretch and rotation tensors are discussed in Section 1.9.4.

We shall frequently make use of the following theorem:

Theorem 1.6.7. Let X, Y and Z be normed vector spaces, and let f : X Z and

g : X Y be two mappings such that

{ x, x 0 X and g( x ) = g( x 0 )} = f ( x ) = f ( x 0 ).

f ( x ) = h( g( x )) x X.

A scalar function : V < is isotropic if (v) = ( Qv) for all v V and Q Orth+ .

We have the following characterization:

Theorem 1.6.8. is isotropic if and only if there exist a function such that

(v) = (|v|) v V.

Proof. If exists, then

( Qv) = (| Qv|) = (|v|) = (v) v V.

To prove the converse, we need to show that |u| = |v| implies (u) = (v),

and then the result follows from Theorem 1.6.7. As noted after the proof of

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

54 Continuum Mechanics

Theorem 1.5.2, if |u| = |v|, there exists a proper orthogonal tensor that rotates u

into v, i.e.,

v = Qu.

Orth+ . We have the following characterization:

Theorem 1.6.9. g is isotropic if and only if there exists a function such that

g (v) = (|v|)v v V.

orthogonal tensor with v/ |v| as its axis so that Qv = v. By the assumed isotropy,

we have

g (v) = (v)v. Again by the assumed isotropy, (v) = ( Qv), and the desired

result follows by Theorem 1.6.8.

In what follows, we restrict the domain of the functions under consideration to a sub-

set of Sym, and denote the list of principal invariants ( I1 , I2 , I3 ) by I . A scalar function

: Psym < is an isotropic function if it remains invariant under orthogonal transforma-

tions, i.e.,

Although we have restricted S to be in Psym in the above definitions, all the following

results will hold even if S lies in any subset A of Sym that is invariant under Orth+ , i.e., if

This condition is imposed in order that ( QSQ T ) and G ( QSQ T ) make sense. As is obvi-

ous from the above definition, Sym itself is invariant under Orth+ . To show that Psym is

invariant under Orth+ , let S Psym. By Theorem 1.6.4, all the eigenvalues of S are strictly

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 55

positive. The following theorem shows that the eigenvalues of QSQ T are the same as the

eigenvalues of S. Then, again appealing to Theorem 1.6.4, we conclude that QSQ T Psym.

= det[ Q(S I ) Q T ]

= [det( QQ T )] det(S I )

= det(S I ).

Thus, the characteristic equation for QSQ T and S is the same, leading to the

same principal invariants (and hence, also to the same eigenvalues).

Theorem 1.6.11. The tensors A, B Sym have the same principal invariants if and

only if there exists a proper orthogonal tensor Q such that A = QBQ T .

Proof. If there exists Q Orth+ such that A = QBQ T , then by Eqn. (1.137)

(which also holds when S Sym), A and B have the same principal invariants.

Conversely, if A and B have the same principal invariants, then by Eqn. (1.113)

they have the same eigenvalues. By the spectral resolution theorem, we can

write

A= i ei ei ,

i

B= i ei ei .

i

Since {ei } and {ei } are orthonormal sets of vectors, by Eqn. (1.95) there exists

Q = ek ek Orth+ such that ei = Qei . Using Eqn. (1.344), we now have

i i i

Theorem 1.6.12. A function : Psym < is isotropic if and only if there exists a

function : I(S) < such that

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

56 Continuum Mechanics

Eqn. (1.137).

To prove the converse, we need to show that

I A = I B = ( A) = ( B),

for A, B Psym, and then the desired result follows from Theorem 1.6.7. Since

A and B have the same invariants, by Theorem 1.6.11 there exists Q Orth+

such that A = QBQ T . Since is isotropic,

( A) = ( QBQ T ) = ( B).

The generalization of the above theorem to arbitrary tensors is given by the follow-

ing [246, 294]:

Theorem 1.6.13. A function : Lin < is isotropic if and only if there exists a

function such that

( T ) = (tr T, tr T 2 , tr T 3 , tr T T T , tr T 2 T T , tr T 2 ( T T )2 , tr T T T T 2 ( T T )2 )

T Lin (n = 3).

Now we state and prove the RivlinEricksen representation theorem for isotropic ten-

sor functions (the generalization to an isotropic symmetric tensor-valued function of two

symmetric tensors is stated in Eqn. (7.56)).

Psym Sym is isotropic if and only if there exist scalar functions 0 , 1 , 2 : I(S)

<, such that

Eqn. (1.137), we have

= QG (S) Q T ,

To prove the converse, assume that G is isotropic, i.e., G ( QSQ T ) = QG (S) Q T

for all Q Orth+ . We first prove that every eigenvector of S is an eigenvector

of G (S), or, alternatively, any matrix that diagonalizes S also diagonalizes G (S).

Let the spectral resolution of S be given by S = i i ei ei . If we form a

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 57

Theorem 1.6.2, Q diagonalizes S, and is proper orthogonal. Let

Q1 = e1 e1 e2 e2 e3 e3 ,

Q2 = e1 e1 + e2 e2 e3 e3 .

addition,

i

(1.140)

QG (S) Q T is diagonal, or, in other words, again appealing to Theorem 1.6.2, the

eigenvectors of S and G (S) are the same. Hence

G (S) = 1 e1 e1 + 2 e2 e2 + 3 e3 e3 , (1.141)

Let us now establish that G (S) can be written as

G ( S ) = 0 ( S ) I + 1 ( S ) S + 2 ( S ) S2 S Psym, (1.142)

where 0 , 1 , 2 are real-valued functions of S. Assume first that S has three dis-

tinct eigenvalues i , with associated eigenvectors ei . Then the two sets { I, S, S2 }

and {e1 e1 , e2 e2 , e3 e3 } span the same subspace of the vector space of

symmetric tensors. To see this, we first observe that e1 e1 , e2 e2 , and e3 e3

are symmetric tensors. They are also linearly independent, since if

e1 e1 + e2 e2 + e3 e3 = 0,

then carrying out the operation e1 ( )e1 , where ( ) denotes the left- and right-

hand sides of the above equation, and using the orthogonality of e1 , e2 and e3 ,

we get = 0; similarly, the operations e2 ( )e2 and e3 ( )e3 yield = 0 and

= 0. Thus, Lsp{e1 e1 , e2 e2 , e3 e3 } is a three-dimensional subspace of

the vector space of symmetric tensors, and, hence, is a vector space itself. Since,

by Eqns. (1.38) and (1.125),

I = e1 e1 + e2 e2 + e3 e3 ,

S = 1 e1 e1 + 2 e2 e2 + 3 e3 e3 ,

S2 = 21 e1 e1 + 22 e2 e2 + 23 e3 e3 ,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

58 Continuum Mechanics

1 1 1

det 1 2 3 = (1 2 )(2 3 )(3 1 ) 6= 0,

21 22 23

the result

Lsp{e1 e1 , e2 e2 , e3 e3 } = Lsp{ I, S, S2 }.

Hence, Eqn. (1.141) can be written in the form given by Eqn. (1.142).

Assume, next, that the tensor S has a double eigenvalue, say 2 = 3 6=

1 . Then, the two sets { I, S} and {e1 e1 , e2 e2 + e3 e3 } span the same

subspace, since, in this case, we can write

I = e1 e1 + (e2 e2 + e3 e3 ),

S = 1 e1 e1 + 2 (e2 e2 + e3 e3 ),

and

" #

1 1

det = 2 1 6= 0.

1 2

All the vectors in the subspace spanned by e2 and e3 are eigenvectors of S, and

hence those of G (S) as already proved. Thus,

G (S)e2 = 2 e2 , G (S)e3 = 3 e3 ,

G (S)(e2 + e3 ) = (e2 + e3 ),

G (S) = 1 e1 e1 + (e2 e2 + e3 e3 ),

as

Assume, finally, that S has three repeated eigenvalues. Now, the eigenvalues

of G (S) are also all repeated (proved in the same way as in the previous case),

and all unit vectors are eigenvectors of S, and hence also of G (S). Thus, by using

the spectral resolution of G (S), we conclude that it is a multiple I, i.e.,

G (S) = 0 (S) I,

proving the assertion in all cases. Note from the above discussion that 2 (S) = 0,

when there are two repeated eigenvalues, and 1 (S) = 2 (S) = 0, when all

eigenvalues are identical.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 59

By virtue of Theorem 1.6.12, it suffices to prove that 0 , 1 and 2 are isotropic

functions. This follows from the isotropy of G (S):

h i

G ( QSQ T ) = Q 0 ( QSQ T ) I + 1 ( QSQ T )S + 2 ( QSQ T )S2 Q T

h i

= QG (S) Q T = Q 0 (S) I + 1 (S)S + 2 (S)S2

Q T S Psym, Q Orth+ ,

{ I, S, S2 }, { I, S} or { I } according to which case is being considered. Thus, by

Theorem 1.6.12, there exist functions (IS ) such that

Note that the above theorem does not say that an isotropic function is a quadratic func-

tion of the elements of S, since the functions i (IS ), i = 0, 1, 2, can be arbitrary.

For linear functions, we have the following simpler result:

tions). A linear function G (S) : Sym Sym is isotropic if and only if there exist

scalars and such that

G (S) = (tr S) I + 2S S Sym.

Proof. Although one could try and specialize the result from the previous theo-

rem to the present case, it is easier to proceed directly; the following short proof

is due to Gurtin [108]. Let V1 be the set of all unit vectors. For e V1 , the tensor

e e has eigenvalues (0, 0, 1) (since there exists Q Orth+ such that e = Qe1 ,

which implies that e e = Q(e1 e1 ) Q T , so that by Eqn. (1.137), the principal

invariants, and hence the eigenvalues, of e e and e1 e1 are the same). Since

two eigenvalues of e e are the same, analogous to the case of repeated eigen-

values in the previous theorems proof, G (e e) also has a repeated eigenvalue,

and thus by Eqn. (1.123) we have

G (e e) = (e) I + 2(e)e e e V1 . (1.143)

Choose e, f V1 , and let Q be a proper orthogonal tensor such that Qe = f .

Since

Q(e e) Q T = f f ,

and since G is isotropic,

0 = QG (e e) Q T G ( f f ) = [(e) ( f )] I + 2[(e) ( f )] f f .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

60 Continuum Mechanics

f , it follows that = = 0, implying in turn that I and f f are linearly

independent; thus,

( e ) = ( f ), ( e ) = ( f ).

Therefore, and are scalar constants, and we conclude from Eqn. (1.143) that

G (e e) = I + 2e e. (1.144)

S= i ei ei ,

i

i

left as an exercise (see Problem 37).

A detailed discussion about the decomposition of higher-order tensors into irreducible ten-

sors, and their eigenvalues and eigentensors can be found in [8], [150], [151], [260], [304]

and [360].

Similar to the definition of a second-order tensor, a third-order tensor, which we repre-

sent by B, is defined as a linear transformation that transforms a vector v to a second-order

tensor T, i.e.,

Bv = T. (1.145)

In order to find the representation for a third-order tensor in terms of its components, we

define the dyadic or tensor product of three vectors a, b and c as

Since a b is a second-order tensor, and since the above transformation is linear, the tensor

product of three vectors is a linear transformation that transforms vectors into

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 61

second-order tensors, and hence is a third-order tensor. Just as a second-order tensor can

be expressed as the sum of 9 dyads (Eqn. (1.37)), a third-order tensor can be expressed as

the sum of 27 dyads as follows:

B = Bijk ei e j ek , (1.147)

where

Bu = ( Bu)ij (ei e j )

= ei ( Bu)e j (ei e j )

= ei [ B(uk ek )] e j (ei e j )

= ei [ B((ek u)ek )]e j (ei e j )

= ei ( Bek )e j (ek u)(ei e j )

= ei ( Bek )e j (ei e j ek )u .

Since u is arbitrary, we have proved that the tensor B can be represented by Eqn. (1.147),

with the components Bijk of this tensor given by Eqn. (1.148). Note that

Bem = Bijm ei e j .

In [360], using the identity e pjr e pbr = 2jb , the authors write

1

Bijk = e e B ,

2 pjr pbr ibk

then, assuming all eigenvalues to be distinct, use the spectral resolution of a fourth-order

tensor to write Bibk e pbr in the form 9m=1 m (Pm )ipkr , from which it follows that

9

1

Bijk = m e (Pm )ipkr .

2 pjr

(1.149)

m =1

Based on this representation, the authors claim that the i s (not all independent, since

Bibi e pbp = 9m=1 m = 0) are eigenvalues of B. However, since the eigenvalue problem for

a third-order tensor is never defined in this work, it is not clear how the eigenvalues of the

fourth-order tensor are also the eigenvalues of the third-order tensor.

In general, an nth order tensor is defined as a linear transformation that transforms a

vector to an n 1th order tensor. It can be proved by induction that an nth order tensor

can be represented by the sum of 3n dyads as

where

Ti1 i2 i3 ...in = ei1 ((( Tein )ein1 ) )ei2 .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

62 Continuum Mechanics

As in the case of second-order tensors, we define the sum of two nth order tensors S

and T as

(S + T )u := Su + Tu u V,

(T )u := ( Tu) u V.

It is easily seen from the definition of a tensor as a linear transformation that S + T and T

are nth order tensors.

The transformation law for third order tensors can be deduced from Eqn. (1.148) as

follows:

= Qil Q jm Qkn el ( Ben )em

= Qil Q jm Qkn Blmn .

The alternate tensor E whose components are given by Eqn. (1.11) is an example of a third-

order tensor. This can be proved by noting that

eijk = ei (e j ek )

= Qil Q jm Qkn el (em en )

= Qil Q jm Qkn elmn ,

which is nothing but the transformation law for the components of a third-order tensor. In

fact, it is an isotropic third-order tensor, as we show in Section 1.8. Note that [[E w]v]u =

[u, v, w].

Similarly, the transformation law for a fourth-order tensor is

Generalizing the above arguments, the transformation law for an nth order tensor is given

by

However, another alternative way that is very widely used is to construct it as a dyadic

product of two second-order tensors. This procedure is analogous to constructing second-

order tensors using the dyadic product of vectors. We now discuss this methodology,

which should be compared with the one for second-order tensors presented in Section 1.3.

A fourth-order tensor is defined as a linear transformation that maps second-order ten-

sors to second-order tensors. Thus, if C is a fourth-order tensor that maps a second-order

tensor A to a second-order tensor B, we write it as

B = CA. (1.152)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 63

C satisfies the property C(A1 + A2 ) = CA1 + CA2 , for all , < and A1 , A2 Lin.

The sum, scalar multiple and composition of fourth-order tensors is defined as

(C + D) A = CA + DA A Lin,

(C) A = (CA) A Lin,

(CD) A = C(DA) A Lin.

the notation Eij ei e j . We have

Analogous to the definition of the components of a second-order tensor (see Eqn. (1.27)),

we now define the components via Eqn. (1.152) as

Taking the dot product on either side with Emn and using Eqn. (1.154), we get Cmnkl =

Emn : CEkl , which, on replacing the indices (m, n) with (i, j) can be written as

CB = DB B Lin, (1.157)

or, alternatively, if

By choosing A Eij and B Ekl in the above equation, we see that equal tensors have

equal components.

Using Eqn. (1.155) and the linearity of fourth-order tensors, Eqn. (1.152) can be written

as

Similarly, we have

We now give some examples of fourth-order tensors. The fourth-order identity tensor I

is defined as one that maps a second-order tensor into itself. The transposer T, symmetrizer

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

64 Continuum Mechanics

S and skew-symmetrizer W are defined as ones that map a second-order tensor into its

transpose, symmetric and skew-symmetric part, respectively, i.e.,

IA = A A Lin,

T

TA = A A Lin,

1

SA = ( A + A T ) A Lin,

2

1

WA = ( A A T ) A Lin.

2

Note that

S = (I + T)/2, I = S + W,

(1.158)

W = (I T)/2, T = S W.

We have

1 1 1

ST = (I + T)T = (T + TT) = (T + I) = S = TS, (1.159)

2 2 2

1 1 1

WT = (I T)T = (T TT) = (T I) = W = TW,

2 2 2

1 1 1

SS = (I + T)S = (S + TS) = (S + S) = S,

2 2 2

1 1 1

WW = (I T)W = (W TW) = (W + W) = W,

2 2 2

1 1 1

SW = (I + T)W = (W + TW) = (W W) = 0 = WS. (1.160)

2 2 2

The components of I, T, S and W are obtained using Eqn. (1.156), and are given by

Iijkl = ik jl ,

Tijkl = il jk ,

1 (1.161)

Sijkl = ( + il jk ),

2 ik jl

1

Wijkl = (ik jl il jk ).

2

Using the above component form of S, we have

1h i

[SAS]ijkl = Aijkl + Aijlk + A jikl + A jilk , (1.162)

4

where A is any fourth-order tensor.

The transpose of C, denoted by CT is defined using the inner product as

The components of CT are obtained by taking A Eij and B Ekl in the above equation,

and are given by

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 65

(CT )T = C, (1.165a)

T T T

(CD) = D C . (1.165b)

Analogous to the dyadic product of two vectors, the dyadic product of two tensors A

and B can be defined as

( A B)(u v) = (u Bv) A.

A tensor C can be written as Cijkl Eij Ekl since, for an arbitrary A Lin, we have

CA = (CA)ij Eij

= [(CA) : Eij ] Eij

= Akl [ Eij : (CEkl )] Eij

= Cijkl ( A : Ekl ) Eij

= Cijkl [ Eij Ekl ] A.

C( A B) = (CA) B,

( A B )C = A (C T B ).

Another tensor product that is of great use is the square tensor product defined as

= (ei Aek )(e j Bel ) = Aik Bjl .

Since ( I I ) A = I AI T = A, for all A Lin, it is clear from Eqn. (1.157) that the fourth-

order identity tensor can be written as

I = I I, (1.167)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

66 Continuum Mechanics

The transpose of A B is given by A T B T since

( H I ) A = H A,

( I H T ) A = AH.

As an example of the use of these equations, note that the tensor equation H A + AH = 0

can be written as ( H I + I H T ) A = 0.

The composition rule is given by

( A B)(C D ) = ( AC ) ( BD ), (1.168)

since

( A B)(C D ) = A (C T BD ), (1.169)

T

( A B)(C D ) = ( ACB ) D, (1.170)

( A B)(u v) = ( Au) ( Bv), (1.171)

( a b ) ( u v ) = ( a u ) ( b v ). (1.172)

As per our convention outlined at the end of Section 1.3.1, the above relations are valid even

if the involved second-order tensors and vectors are complex-valued. The first relation is

proved as follows:

= [ A (C T BD )] E E Lin.

The transposer T commutes with the tensor product A B in the following way:

T( A B) = ( B A)T, (1.173)

since

S( A B + B A) = ( A B + B A)S = S( A B + B A)S = 2S( A B)S, (1.175)

S( A A) = ( A A)S = S( A A)S, (1.176)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 67

where the second relation in Eqn. (1.175) follows from the first by multiplying both sides

by S, and using the fact that SS = S, while the third relation in Eqn. (1.175) follows from

Eqn. (1.174). Equation (1.176) follows from Eqn. (1.175) by taking B = A.

The component form of ( A B)T is Ail Bjk while that of T( A B) is A jk Bil . A summary

of the component forms of the various dyadic products is

[ A B]ijkl = Aik Bjl ,

(1.177)

[( A B)T]ijkl = Ail Bjk ,

[T( A B)]ijkl = A jk Bil .

The tensor C is said to have the first minor symmetry if C = TC, or, in indicial notation,

if

and is said to have the second minor symmetry if C = CT, which in indicial notation reads

Cijkl = Cijlk .

Note that the first minor symmetry condition can be written as (I T)C = 2WC = 0,

while the second minor symmetry condition can be written as CW = 0. Equivalently,

using S = (I + T)/2, one can also write the two symmetry conditions as C = SC and

C = CS, respectively. One has

C = SC and C = CS C = SCS,

since given the conditions on the left, we directly get C = SCS, while, conversely, if C =

SCS, then SC = SSCS = SCS = C, and CS = SCSS = SCS = C. Thus a tensor C has both

the minor symmetries if and only if C = SCS.

The tensor C is said to have major symmetry if C = CT , or, in indicial notation (using

Eqn. (1.164)), if Cijkl = Cklij . The symmetric and skew-symmetric parts of C are given by

(C + CT )/2 and (C CT )/2. Using Eqn. (1.161) or directly, it can be shown that I, T, S

and W are all symmetric.

Assume that C has the first minor symmetry, i.e., C = TC. By transposing this equation,

and using Eqn. (1.165b) and the symmetry of T, we get

CT = CT T.

Thus, if C has the first minor symmetry, then CT has the second minor symmetry. It also

follows that if C has major symmetry and one of the minor symmetries, it also has the other

minor symmetry.

The eigenvalue problem for a fourth-order tensor A that does not possess minor sym-

metries, involves finding the eigenvalue and eigentensor pair (, A), A Lin, which satis-

fies

AA = A. (1.178)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

68 Continuum Mechanics

By applying the -mapping as per Eqn. (I.1e) to the above equation, we get

[(A) I ]( A) = 0.

This equation has a nontrivial solution ( A) 6= 0, if and only if the characteristic equa-

tion det[(A) I ] = 0 is satisfied. The characteristic equation can also be written in

expanded form as

I1 (A) = tr (A) = 1 + 2 + + 9 ,

1

I2 (A) = [(tr (A))2 tr ((A))2 ] = 1 2 + 2 3 + + 1 9 ,

2

...

I9 (A) = det (A) = 1 2 . . . 9 ,

are a strict subset of the complete set of the principal invariants of A (see the references

below for the case where A has minor symmetries, denoted by C there). Similar to the

result for second-order tensors,

det A = det (A) = det[(A1 )(A2 )] = [det (A1 )][det (A2 )] = (det A1 )(det A2 ).

(1.180)

dimension, are the (possibly complex-valued) eigenvalues/eigenvectors of diagonalizable

tensors A, B, then (i j , ui v j ), i, j = 1, 2, . . . , n, are the eigenvalues/eigentensors of

A B.

The theorem may not hold for non-diagonalizable tensors. For example, if A is non-

diagonalizable but invertible, and B = AT , then (1, I ) is an eigenvalue/eigentensor of

A AT , but ui v j (or their linear combinations) corresponding to i j = 1 may not

equal I. An application of the above theorem is presented in Theorem 1.7.2.

An orthogonal fourth-order tensor Q is defined as one for which

QT Q = QQT = I. (1.181)

orthogonal if and only if

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 69

the proof being identical to that for second-order orthogonal tensors. An example of an

orthogonal tensor is Q1 Q2 , where Q1 , Q2 Orth. This follows since

choosing Q1 = Q2 = Q, where Q Orth+ . Equation (1.150) can be expressed as

Since (I) = I 99 , we have det I = 1. It follows from Eqns. (1.179)(1.181) that

det Q = 1.

showing that the invariants that occur in the characteristic equation (and hence the eigen-

values) of A are independent (as they should be) of the choice of basis.

The CayleyHamilton theorem reads

0 = A9 I1 A8 + I2 A7 + I9 I

9

= (A i I).

i =1

k

q (A) = (A i I) mi ,

i =1

algebraic multiplicity of each eigenvalue i . If A is normal, i.e., if AT A = AAT , or, in

particular, symmetric, then each mi = 1.

The tensor A is said to be invertible if there exists another tensor, denoted by A1 , such

that

A1 A = AA1 = I.

(UV)1 = V1 U1 , (1.184a)

(C T ) 1 = (C1 ) T = : C T , (1.184b)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

70 Continuum Mechanics

From Eqn. (1.184b), it follows that if C has major symmetric then so does its inverse. The

inverse of A B (assuming that A and B are invertible) is given by

( A B ) 1 = A 1 B 1 , (1.185)

since, by Eqns. (1.167) and (1.168), we have

( A1 B1 )( A B) = ( A1 A) ( B1 B) = I I = I.

A fourth-order tensor A is invertible if and only if I9 (A) = 9i=1 i is nonzero, in which

case,

1 h 8 i

A1 = A I1 A7 + + I8 I .

I9

By virtue of the relation ( A T ) T = A, we have TTA = A for all A Lin, which implies that

TT = I. (1.186)

Thus, the inverse of the transposer tensor T is T itself.

If A has major symmetry (but not necessarily the minor symmetries), then [(A)] T =

(AT ) = (A), so that, similar to the case of second-order symmetric tensors, all the nine

eigenvalues are real, and the corresponding eigentensors Ai , i = 1, 2, . . . , 9 are orthogonal

in the sense that

( Ai , A j ) = Ai : A j = ij . (1.187)

Thus, we can write the spectral representation of A as

9 9

A= i Ai Ai , with Ai Ai = I.

i =1 i =1

If k is the number of distinct eigenvalues, then, similar to Eqn. (J.4), we have the following

explicit formula:

A I

kj=1 j , k > 1

i j

Ai Ai = j 6 =i (1.188)

I, k=1

9

1

A1 = i Ai Ai .

i =1

A is said to be positive definite if T : AT > 0 for all nonzero T Lin. Similar to symmetric

second-order tensors (see Theorem 1.6.4), one can show that A is positive definite if and

only if all the eigenvalues i (or all the invariants that occur in the characteristic equation)

are strictly positive. If A is positive semi-definite, the square root of A defined via the

relation HH = A is given by

9

H=

p

i Ai Ai .

i =1

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 71

Now consider the case when the fourth-order tensor C has both minor symmetries,

i.e., C = SCS. The transpose is still defined by Eqn. (1.163), except that A, B Sym.

Using this definition, one can show that when C and D have both minor symmetries, then

Eqns. (1.165) still hold. The eigenvalue problem given by Eqn. (1.178) is now written as

CS = S = SS, S Sym, or, alternatively, using Eqn. (I.14), as

h i

(C) H 1 I 66 (S) = 0.

Thus, C has only six eigenvalues, with the remaining three eigenvalues zero, and the cor-

responding eigentensors being linearly independent skew-symmetric tensors. The above

equation has a nontrivial solution (S) 6= 0, if and only if det[(C) H 1 I ] = 0. This

characteristic equation can be written in expanded form as

where

I1 (C) = tr [(C) H 1 ] = 1 + 2 + + 6 ,

1

I2 (C) = [(tr [(C) H 1 ])2 tr ((C) H 1 )2 ] = 1 2 + 2 3 + + 1 6 ,

2

...

are a strict subset of the set of principal invariants of C [4, 20, 141, 238] (the results of Ahmad

and Norris on the one hand, and Betten on the other regarding the number of quadratic

principal invariants seem to be contradictory). As an example,

If C has both the minor symmetries, then [C] = [S][C][S] and [C] = [S][C][S], so that

Eqn. (1.182) can be written as

where Q := [S][ Q Q][S], Q Orth+ . Now applying Eqns. (1.165b), (I.11c) and (I.11f),

and using the symmetry of S and H, we get

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

72 Continuum Mechanics

where A := (Q) H 1 is found by using Eqn. (I.5); the final expression for A is given by

Eqn. (I.12) with J Q T . Using Eqns. (1.165b) and (1.176), we get

det Q = 1.

= (det S) det(C S) = det(C S), (1.194)

showing that the invariants that occur in the characteristic equation (and hence the eigen-

values) of C are independent of the choice of basis.

The spectral decomposition of C reads

6 6

C= i Ai Ai , with Ai = AiT and Ai Ai = S.

i =1 i =1

C S

kj=1 j , k > 1

i j

Ai Ai = j 6 =i

S. k=1

C1 C = CC1 = S.

The necessary and sufficient condition for C to be invertible is that I6 (C) be nonzero. If this

condition is satisfied, then

6

1

C1 = i Ai Ai .

i =1

minor symmetries. Then, since ST = S, both relations given by Eqns. (1.184) hold. From

Eqn. (1.176), it follows (provided A is invertible) that

C is said to be positive semi-definite if S : CS 0 for all S Sym. If C is positive

semi-definite, its square root H is given by

6

H= with Ai = AiT .

p

i Ai Ai ,

i =1

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 73

Examples of some spectral resolutions are

9

I= Ai Ai ,

i =1

6 9

T= Ai Ai Ai Ai ,

i =1 i =7

6

S= Ai Ai , AiT = Ai

i =1

3

W= Ai Ai , AiT = Ai

i =1

I I

II =3 .

3 3

Let e be the axial vector of Q Orth+ \Sym, and let (, n) and (, n) denote the re-

maining eigenvalue/eigenvector pairs, with the hat denoting a complex conjugate. The

following result, which follows directly from Theorem 1.7.1, holds:

Orth+ \Sym, are (1, I ), (1, Q), (1, Q2 ), (, e n), (, n e), (, e n), (, n e),

(2 , n n), and (2 , n n).

denote the eigenvalue/eigenvector pairs of Q, then the eigenvalue/eigentensor pairs of

Q are (1, I ), (1, Q), (1, Q2 ), (1, Q3 ), (21 , n1 n1 ), (21 , n1 n1 ), (22 , n2 n2 ), (22 , n2

n2 ), (1 2 , n1 n2 ), (1 2 , n2 n1 ), (1 2 , n1 n2 ), (1 2 , n2 n1 ), (1 2 , n1 n2 ),

(1 2 , n2 n1 ), (1 2 , n1 n2 ) and (1 2 , n2 n1 ). In general, when the underlying space

dimension is n, the eigenvalue one is repeated n times, with the corresponding eigenten-

sors given by I, Q, . . ., Qn1 , and the remaining n(n 1) eigenvalue/eigentensor pairs

are formed by combinations of each complex eigenvalue/eigenvector of Q with itself and

with all the remaining eigenvalues/eigenvectors of Q barring the complex conjugate of

that eigenvalue/eigenvector.

We now discuss the solution of the tensor equation AX B + CX D = H, where A, B, C,

D and H are given second-order tensors. We write this equation as

[ A B T + C D T ] X = H.

where

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

74 Continuum Mechanics

1

T1111 T1122 T1133 T1112 T1123 T1131 T1121 T1132 T1113

X11 H11

X T T2222 T2233 T2212 T2223 T2231 T2221 T2232 T2213

H22

22 2211

X33 T3311 T3322 T3333 T3312 T3323 T3331 T3321 T3332 T3313

H33

X T T1222 T1233 T1212 T1223 T1231 T1221 T1232 T1213 H12

12 1211

X23 = T2311 T2322 T2333 T2312 T2323 T2331 T2321 T2332 T2313 H23 ,

X31 T3111 T3122 T3133 T3112 T3123 T3131 T3121 T3132 T3113

H31

X T T2122 T2133 T2112 T2123 T2131 T2121 T2132 T2113 H21

21 2111

X32 T3211 T3222 T3233 T3212 T3223 T3231 T3221 T3232 T3213 H32

X13 T1311 T1322 T1333 T1312 T1323 T1331 T1321 T1332 T1313 H13

(1.196)

with the components Tijkl given by Eqn. (1.195). It is clear that this method can be extended

to solve a matrix equation of the type i Ai X Bi = H of any dimension n by inverting an

n2 n2 matrix.

Of particular importance is the case when B = C = I and D = A S Sym. In this

case, Tijkl = Sik lj + ik Slj .

Although the above approach is suitable for numerical purposes, a tensorial solution is

more suitable for proving properties of the solution X. We now present a method, based on

[155], for determining (S I + I S)1 , S Sym, which is applicable for any dimension

n (see Rosati [273] and references therein for other methods).

Consider the tensor T I + I T T , T Lin. If (i , ui ) and (i , vi ) are the eigen-

value/eigenvectors of T and T T , respectively, then

= ( Tui ) v j + ui ( T T v j )

= (i + j )(ui v j ),

T I + I T T . However, unfortunately, these are not the only eigenvalue/eigentensor

pairs, especially when the eigenvectors {ui } are not linearly independent. For example,

when T = e1 e2 + e3 e4 + e4 e5 , with n = 5, all the 25 eigenvalues of T I +

I T T are zero, and there are nine linearly independent eigentensors given by e1 e2 ,

e1 e5 , e3 e2 , e3 e5 , e3 e3 e4 e4 + e5 e5 , e3 e4 + e4 e5 , e3 e1 + e4 e2 ,

e1 e4 + e2 e5 and e1 e1 + e2 e2 . Although the first four eigentensors are of the

form ui v j , i, j = 1, 2, the remaining five are not. A similar situation arises even in the

case of the tensor T I + I T, which again has nine linearly independent eigentensors.

Nevertheless, as we now show, i + j , i, j = 1, 2, . . . , n, are the eigenvalues of T I + I

T T (or T I + I T). Thus, the necessary and sufficient condition for T I + I T T to be

invertible is that (det T ) ni=1 (i + j ) 6= 0. For example, if n = 3, this condition is given

j = i +1

by 1 2 3 (1 + 2 )(2 + 3 )(1 + 3 ) 6= 0.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 75

form. Then

T I + I T T = ( H 1 J 1 H 11 ) I + I ( H 2 J 2 H 21 )

= ( H 1 H 2 )( J 1 I + I J 2 )( H 11 H 21 )

= ( H 1 H 2 )( J 1 I + I J 2 )( H 1 H 2 )1 .

Thus, the eigenvalues of T I + I T T and J 1 I + I J 2 are the same. Since the matrix

form of J 1 I + I J 2 is an upper triangular matrix with i + j , i, j = 1, 2, . . . , n, on the

diagonals, we get the desired result. A similar proof can be devised for T I + I T.

Let (i , ni ) be the eigenvalues/eigenvectors of S Sym. Then, the eigenvalue/eigentensor

pairs of S(S I + I S)S when all the i are distinct are i + j , ni n j + n j ni , i, j =

1, 2, . . . , n (the eigenvalue/eigentensors when the eigenvalues of S are repeated can be

found similar to Eqns. (1.258) and (1.259)). Thus, the necessary and sufficient condition

that S(S I + I S)S be invertible is that (det S) ni=1 (i + j ) 6= 0.

j = i +1

Let i , i = 1, 2, . . . , k, be the distinct eigenvalues of S Sym, and let Pi be the associated

projections, which are given by Eqn. (J.4). We have

" ! ! ! !#

k k k k

S( S I + I S )S = S i P i Pj + Pi j Pj S

i =1 j =1 i =1 j =1

" #

k k

=S ( i + j ) P i P j S.

i =1 j =1

Assuming that the necessary and sufficient condition for invertibility is satisfied, the de-

sired inverse is

" #

k k P P

i j

[S( S I + I S )S] = S

1

S, (1.197)

+ j

i =1 j =1 i

" # " #

k k P P k k (P S) P + P (P S)

i j i j i j

S S [S I + I S] S = S S

+ j

i =1 j =1 i i =1 j =1

i + j

" #

k k

=S Pi P j S

i =1 j =1

" ! !#

k k

=S Pi Pj S

i =1 j =1

= S [I I] S

= S,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

76 Continuum Mechanics

where we have used the properties Pm Pi = im Pi (no sum on i), ik=1 Pi = I and SPi =

(km=1 m Pm )Pi = i Pi = Pi S (no sum on i). Thus, the solution of the tensor equation

S(S I + I S)SX = H is

k k Pi ( H + H T )P j

X= 2( i + j ) . (1.198)

i =1 j =1

The above method can obviously be generalized to the case when T is diagonalizable.

Thus, if the tensor equation to be solved is ( T I + I T ) X = T X + XT T = H, where

T = ik=1 i Pi , with Pi given by Eqn. (J.26), then assuming that T is such that the conditions

for invertibility of T I + I T are satisfied, we have

X = ( T I + I T ) 1 H

" #

k k P P

i j

= H

+ j

i =1 j =1 i

k k Pi HP Tj

= i + j

.

i =1 j =1

We now show how the solutions given by Eqns. (1.198) reduce to some of the solutions

presented in the literature for the case H = W Skw. When n = 3, and when the

eigenvalues are distinct, for a given i, we have Pi W Pi = (ei W ei )Pi = 0. Multiplying

both sides of Eqn. (1.198) by the determinant of

(det S) X = SW S, (1.199)

which is the solution presented by Scheidler [282]. Since the solution has no explicit

dependence on the eigenvalues i , it holds even under the case when the eigenvalues

are repeated. Alternatively, by writing W as (P1 + P2 + P3 )W (P1 + P2 + P3 ), S2 W as

(21 P1 + 22 P2 + 23 P3 )W (P1 + P2 + P3 ), etc. and noting that (1 + 2 )(2 + 3 )(1 + 3 ) =

I1 (S) I2 (S) I3 (S), one can also show that when H = W, Eqn. (1.198) reduces to

( I1 I2 I3 ) X = ( I12 I2 )W (S2 W + W S2 ),

Finally, consider the case when the dimension of the underlying vector space n is two,

and H is arbitrary. Using I1 (S) = 1 + 2 , I2 (S) = 1 2 , and writing H = (P1 +

P2 ) H (P1 + P2 ), SH = (1 P1 + 2 P2 ) H (P1 + P2 ), etc., one can show that Eqn. (1.198)

can be written as

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 77

which is the solution presented by Hoger and Carlson [127]. Alternatively, if S is invertible,

writing S1 = P1 /1 + P2 /2 , we can also show that

A second-order tensor is said to be isotropic if its components are the same in all orthonor-

mal bases, i.e., if T = Tij ei e j = Tij ei e j . Since ei = Q T ei , this condition can also be

written as

or, alternatively, as

QT = TQ Q Orth+ . (1.201)

<.

To prove the converse, let v be an arbitrary vector, and let R be the proper

orthogonal tensor for which v is along its axis, i.e., Rv = v (see Eqn. (1.104)).

Since RT = T R, we have

RTv = T Rv = Tv,

i.e., Tv also lies along the axis of R. Since the axis of R is a one-dimensional

subspace, we have

Tv = v = Iv.

tensors.

QT = T,

to the eigenvalue one. Hence, the above equation implies that

T = 0 ( Q) I + 1 ( Q) Q + 2 ( Q) Q,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

78 Continuum Mechanics

Q, we have

T = 0 I + 1 Q0 + 2 Q20 ,

where, now, 0 , 1 and 2 are constants. Substituting this expression into Eqn. (1.201), we

get 1 = 2 = 0, since, in general, rotations do not commute. This methodology can be

extended to other dimensions as well. For example, if the dimension n is four, then we

get T = 0 I + 1 Q0 + 2 Q20 + 3 Q30 , but again, since rotations do not commute, we get

all the constants other than 0 to be zero. Now consider the case when n is two. In this

case, T = 0 I + 1 Q0 , but now since rotations do commute when n = 2, both 0 and 1

are nonzero. Using the representation of Q0 in terms of cos 0 and sin 0 , we can write the

most general form of an isotropic second-order tensor when n = 2 as T = .

A third-order tensor is isotropic if and only if it is of the form E , where E is the alter-

nate tensor with components eijk . To see the if part, note that

e3 e3 e2 , e2 e2 + e3 e1 e1 e3 , etc.) in the relation

A fourth-order tensor is said to be isotropic if it has the same components in all or-

thonormal bases, i.e., if

C = QCQT , (1.202)

where Q := Q Q, with Q Orth+ . Alternatively, the above condition can also be written

as

the form

C = I I + ( + )I + ( )T (1.204)

= I I + 2S + 2W,

or, in other words, a fourth-order isotropic tensor is a linear combination of tensors with

components ij kl , ik jl and il jk .

If C has the first or second minor symmetry, i.e., if C = SC or C = CS, then using

Eqn. (1.160), it follows as a corollary that

C = I I + 2S. (1.205)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 79

Proof. The proof is based on [157]. To see the if part, note that

QIQT = QQT = I,

Q(CQ) = CQ.

eigenvalue one, i.e., by virtue of Theorem 1.7.2,

CQ = 0 ( Q) I + 1 ( Q) Q + 2 ( Q) Q2 Q Orth+ \Sym.

In the limiting cases following Eqn. (1.105b), namely, (i) 0, we have Q = I,

so that 1 and 2 can be taken to be zero, and (ii) , we have Q2 = I, so

that 2 can be taken to be zero in the above expression. Thus, by virtue of the

continuous dependence of Q on , we can write

CQ = 0 ( Q) I + 1 ( Q) Q + 2 ( Q) Q2 Q Orth+ , (1.206)

We now show that the functions i , i = 0, 1, 2, are in fact functions of the

principal invariants of Q (which is just the trace, since, for proper orthogonal

tensors, the determinant is equal to unity, and the second invariant is equal to

the trace). By letting A R0 Orth+ in Eqn. (1.203), we get

Substituting the expressions for CR0 and C( QR0 Q T ) obtained from Eqn. (1.206)

into the above equation, we get

h i h i

0 ( R0 ) 0 ( QR0 Q T ) I + 1 ( R0 ) 1 ( QR0 Q T )

h i

R0 + 2 ( R0 ) 2 ( QR0 Q T ) R20 = 0.

linearly independent sets, depending on whether R0 Orth+ \Sym, R0

Orth+ Sym { I } (2 = 0), or R0 = I (1 = 2 = 0), we have

i.e., the i s are isotropic scalar-valued functions of R0 . To show that they are

functions of I1 ( R0 ), we have to show that

I1 ( Q1 ) = I1 ( Q2 ) = i ( Q1 ) = i ( Q2 ), i = 0, 1, 2,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

80 Continuum Mechanics

by Theorem 1.5.5, there exists Q0 Orth+ such that Q2 = Q0 Q1 Q0T . Thus,

using Eqn. (1.207), we get i ( Q2 ) = i ( Q0 Q1 Q0T ) = i ( Q1 ). It follows from

Theorem 1.6.7 that there exist functions i such that i ( Q) = i ( I1 ( Q)).

Equation (1.206) can now be written as

CQ = 0 ( I1 ) I + 1 ( I1 ) Q + 2 ( I1 ) Q2 . (1.208)

CQ = [0 ( I1 ) 2 ( I1 ) I1 ] I + [1 ( I1 ) + 2 ( I1 ) I1 ] Q + 2 ( I1 ) Q T .

right-hand side of the above expression has to be linear in Q, i.e., the coefficients

of Q and Q T are constants, and the coefficient of I has to be a multiple of tr Q.

Thus, set 0 ( I1 ) ( + ) I1 , 1 ( I1 ) ( + ) ( ) I1 , and 2 ( I1 )

, where , and are constants. This leads to

CQ = (tr Q) I + ( + ) Q + ( ) Q T

The last step is to show that the above equation implies the expression

given in Eqn. (1.204). Thus, by virtue of Eqn. (1.157), we have to show

that Eqn. (1.209) holds for an arbitrary tensor T in place of Q. Let Ciso :=

[( I I ) + ( + )I + ( )T]. Since fourth-order tensors are a linear trans-

formation from Lin to Lin, Eqn. (1.209) implies that

i

Q Orth+ , the proof is complete.

For various other proofs, see, e.g., [24], [102], [103], [105], [110], [142], [148], [175], [210],

[250] and [251]. Note that this result does not hold when the space dimension n = 2, as

the counterexample C = T T, where T = shows (the above proof fails in the last

step for this case, since an arbitrary tensor cannot be expressed as a linear combination of

rotations). However, the result is applicable for any n if we assume C to be symmetric.

In the subsequent chapters, we shall deal with tensor fields, where a tensor T is a function

of a scalar quantity t, and a vector quantity x (e.g., t representing time, and x the position

vector). In this section, we define the concept of the derivative of a tensor quantity. Let X

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 81

and Y denote normed vector spaces, and let f : X Y. We say that f (u) approaches

zero faster than u, and write f (u) = o (u) if

k f (u)k

lim = 0.

u0 kuk

Let g be a function whose values are scalars, vectors or tensors, and whose domain is an

open interval of <. The derivative Dg (t) is defined by

d 1

Dg (t) := g (t) = lim [ g (t + ) g (t)] . (1.210)

dt 0

1

lim [ g (t + ) g (t) Dg (t)] = 0,

0

or, equivalently,

Thus, we see that the derivative of a vector function is a vector, or the derivative of a tensor

function is a tensor.

We now extend the above concept of a derivative to domains that lie in spaces of di-

mensions higher than one. Let L( X, Y ) denote the vector space of all continuous linear

mappings from X to Y, and let g : X Y. We say that g is differentiable at x if

there exists a linear transformation Dg ( x) L( X, Y ) such that

Dg ( x)[u] Y is called the directional derivative or Gateaux derivative. Note that the di-

rectional derivative is computed by the action of the Frechet derivative on elements of X.

If the directional derivative exists, it is unique, and can be written as

1 d

Dg ( x)[u] = lim [ g ( x + u) g ( x)] = g ( x + u) . (1.213)

0 d =0

dg

Dg (t)[] = <.

dt

Often, the easiest way of computing the directional derivative is to appeal directly to

its definition. We now consider some examples. Consider the function (v) = v v. Then

(v + u) = (v + u) (v + u)

= v v + 2v u + u u

= (v) + 2v u + o (u) ,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

82 Continuum Mechanics

so that

D(v)[u] = 2v u.

order tensor. Then we have

G ( T + U ) = ( T + U )2 = T 2 + TU + UT + U 2 = G ( T ) + TU + UT + o (U ) ,

Similarly for G ( T ) = T 3 , we have

G ( T + U ) = G ( T ) + T 2 U + TUT + UT 2 + o (U ) ,

For ( T ) = tr ( T ), we have ( T + U ) = ( T ) + tr U, which yields

If ( T ) = tr T k , then (with T 0 I)

k

( T + U )k = T k + T ki UT i1 + o (U ) ,

i =1

and on using the linearity of the trace operator, and the fact that tr AB = tr BA, we get

k

( T + U ) = ( T ) + tr [ T ki UT i1 ] + o (U )

i =1

k

= ( T ) + tr [ T k1 U ] + o (U )

i =1

= ( T ) + k ( T k 1 ) T : U + o (U ) ,

so that

D( T )[U ] = k( T k1 ) T : U. (1.215)

= det T + cof T : U + o (U ) ,

which yields

We now discuss the invariance properties of the derivative of a tensor function.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 83

Theorem 1.9.1. Let G : Sym Sym be an isotopic tensor function. Then the deriva-

tive is invariant in the following sense:

(1.217)

Proof. We have already remarked that Sym is invariant under Orth+ . Thus, we

simply need to prove Eqn. (1.217). Since G (S) is isotropic, we have

G Q(S + U ) Q T = QG (S + U ) Q T

= QG (S) Q T + QDG (S)[U ] Q T + o (U )

= G ( QSQ T ) + QDG (S)[U ] Q T + o (U ) .

G Q(S + U ) Q T = G ( QSQ T + QUQ T )

= G ( QSQ T ) + DG ( QSQ T )[ QUQ T ] + o (U ) .

From the above equations, and the uniqueness of the derivative, we get

Eqn. (1.217).

Quite often, we will be required to compute the directional derivative of a bilinear map

( f , g ). Examples of bilinear maps are [108]

the product of a scalar and a vector

(, v) = v,

(u, v) = u v,

(u, v) = u v,

( T, v) = Tv,

and so forth. Assuming that f and g are differentiable at a point x in the common domain

of f and g, the product ( f ( x), g ( x)) is differentiable, and its directional derivative is given

by (see [108] for the proof)

D ( f ( x), g ( x))[u] = ( f ( x), Dg ( x)[u]) + ( D f ( x)[u], g ( x)) u. (1.218)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

84 Continuum Mechanics

( f (t), g (t)) = ( f (t), g (t)) + f (t), g (t) .

Another result that is used frequently is the chain rule, which we now state.

Let f : 2 3 and g : 1 2 , and let g ( x) be differentiable at x 1 . Further, let f

be differentiable at y = g ( x). Then the composition f g is differentiable at x, and

d

D ( f g )(t)[] = ( f g ),

dt

dg

Dg (t)[] = ,

dt

and hence,

d dg

f ( g (t)) = D f ( g (t)) . (1.220)

dt dt

We have already shown how to find the directional derivatives of the first and third in-

variants. To illustrate the product and chain rules, we now compute the directional deriva-

tive of the second invariant of a second-order tensor I2 ( T ) = (det T )(tr T 1 ) given by

h i h i

DI2 ( T )[U ] = det T tr ( T 1 U )tr T 1 tr ( T 1 UT 1 ) = (tr T ) I T T : U. (1.221)

DT [U ] T 1 + T DT 1 [U ] = 0,

where G ( T ) = T 1 , we have

D( T )[U ] = D( G )[ DG ( T )[U ]] = D( G )[ T 1 UT 1 ] = tr ( T 1 UT 1 ).

= det Ttr ( T 1 U )tr T 1 det Ttr ( T 1 UT 1 ),

which proves the first relation in Eqn. (1.221). To get the second relation, we note from

Eqn. (1.80) that I2 is also given by

1h i

(tr T )2 tr ( T 2 ) .

2

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 85

Hence, we have

1

DI2 [U ] = [2(tr T )tr U 2tr ( TU )]

h2 i

= (tr T ) I T T : U.

The second relation can also be obtained directly from the first using the CayleyHamilton

theorem.

One other useful result is the following:

d dT

(det T ) = cof T : . (1.222)

dt dt

To see this note that

d dT

(det T ) = D det( T ) (by Eqn. (1.220))

dt dt

dT

= cof T : (by Eqn. (1.216)).

dt

When T is invertible, then using cof T = (det T ) T T , we can also write Eqn. (1.222) as

d dT 1

(det T ) = det T tr T . (1.223)

dt dt

By taking the directional derivative of the CayleyHamilton equation, we get the fol-

lowing Rivlins identity [68]:

1h i

T 2 U + TUT + UT 2 (tr U ) T 2 (tr T )( TU + UT ) + (tr T )2 tr T 2 U

2

+ [(tr T )(tr U ) tr ( TU )] T (cof T : U ) I = 0.

Similar identities can be derived in other space dimensions by taking the directional deriva-

tives of the relevant CayleyHamilton equation. Identities involving three tensors can be

derived by taking the directional derivative twice. Other identities can be derived by tak-

ing the directional derivative of the CayleyHamilton equation with a tensor function [68].

When X is a Hilbert space, the directional derivative of a real-valued function : X

< can be identified with an element of the space X. The derivative D( x) is an element of

the dual space X 0 = L( X, <) and thus, since the space X is a Hilbert space, there exists by

the Riesz representation theorem [63] a unique element ( x) in the space X that satisfies

where (., .) denotes the inner product of the space X. The element ( x) is called the

gradient of . Note that while D( x) belongs to the dual space X 0 , the gradient ( x)

belongs to the space X.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

86 Continuum Mechanics

As an example, consider the case when X = <3 equipped with the Euclidean inner

product (u, v) = u v. Then we have

( x + u) = ( x) + ( x) u + o (u) .

1 ( x)

()i = D( x)[ei ] = lim [( x + ei ) ( x)] = .

0 xi

Thus, we have

D( x)[u] = u = u.

xi i

As another example, consider the case when X = Lin, with the matrix inner product

( A, B) = A : B. Then the gradient of ( T ) is the matrix /Tij . By definition, we have

D( T )[U ] = :U= U .

T Tij ij

tive of V are related as

V

DV ( T )[U ] = U. (1.225)

T

The gradients of the invariants I1 , I2 and I3 are obtained from Eqns. (1.214), (1.216) and

(1.221), and are given by

I1

= I, (1.226a)

T

I2

= (tr T ) I T T , (1.226b)

T

I3 h iT

= cof T = I2 I I1 T + T 2 . (1.226c)

T

The gradient of ( T ) = tr T k is obtained using Eqn. (1.215), and is given by

(tr T k ) = k( T k1 )T . (1.227)

T

Since Tij /Tkl = Tji /Tlk = ik jl and Tji /Tkl = Tij /Tlk = jk il , we have

T T T

= = I,

T T T (1.228)

T T T

= = T.

T T T

In general, if 1 ( T ) and 2 ( T ) are scalar-valued functions of T, then

(1 2 )

= 1 2 + 2 1 .

T T T

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 87

=T , (1.229)

T T T

U U

T

= T. (1.230)

T T

Since (U )ij = Uij , we have

(U ) U

=U + . (1.231)

T T T

If U ( T ) and V ( T ) are tensor-valued functions of T, then writing U : V as Umn Vmn , we get

T T

(U : V )

U V

= V+ U. (1.232)

T T T

(UV )

Vmj Uim

= Uim + V

T ijkl Tkl Tkl mj

Vmn Unm

= Uim nj + V

Tkl Tkl in mj

Vmn Umn

= Uim nj + V ,

Tkl Tkl im nj

which in tensorial form reads

(UV ) V U

= (U I ) + (I V T) . (1.233)

T T T

Using Eqns. (1.158), (1.173), (1.228) and (1.233), we have the following results:

( T 2 )

= T I + I TT,

T

( T T T )

= ( I T T )T + T T I = 2S( T T I ), (1.234)

T

( T T T )

= ( I T ) + ( T I )T = 2S( I T ), (1.235)

T

( T 1 )

= T 1 T T = ( T T T )1 , (1.236)

T

( T T )

= T( T 1 T T ) = ( T T T 1 )T = T( T T T )1 . (1.237)

T

The fourth relation is obtained by taking U and V to be T 1 and T, respectively, and then

differentiating T 1 T = I. We get

T 1

0 = T 1 I + ( I T T )

T

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

88 Continuum Mechanics

T 1

= = ( I T T )1 ( T 1 I )

T

= ( I T T )( T 1 I ) (by Eqn. (1.185))

1 T

= T T . (by Eqn. (1.168))

The fifth relation is obtained by writing T T as TT 1 and then using Eqn. (1.236).

If ( G ( T )) is a scalar-valued function of a tensor function G ( T ), then by the chain rule,

we have

Gkl

= ,

Tij Gkl Tij

T

G

= . (1.238)

T T G

H H G

= . (1.239)

T G T

Special care has to be exercised in computing the gradients with respect to a symmetric

tensor S. As a specific example, let us consider the computation of the first and second-

order derivatives of a scalar-valued function W (C ), where C Sym. Let S and C denote

these derivatives. We have

1 W W

S(C ) = +

2 C C T

1 W W

= +T (by Eqn. (1.229))

2 C C

1 W

= (I + T)

2 C

W

=S . (by Eqn. (1.158))

C

In indicial notation, the above equation reads

!

1 W W

Sij = + .

2 Cij Cji

(SS) (SS)

1

C= +

2 C C T

1 (SS) (SS)

= + T (by Eqn. (1.230))

2 C C

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 89

1 S

= S (I + T)

2 C

S

=S S (1.240)

C

2 W

=S S. (1.241)

CC

In other words, to find 2 W/CC when C Sym, we compute it as if C is not symmetric,

and then pre- and post-multiply the result with the symmetrizer S. In indicial notation, the

above equations read

!

1 Sij S ji Sij S ji

Cijkl = + + +

4 Ckl Ckl Clk Clk

!

1 2 W 2 W 2 W 2 W

= + + + .

4 Cij Ckl Cji Ckl Cij Clk Cji Clk

Now consider the evaluation of the stretch and rotation tensors of F, and their deriva-

tives for arbitrary space dimension n. Let 2i , i = 1, 2, . . . , k be the distinct eigenval-

ues, and Pi be the corresponding projections of C Psym, so that C = ik=1 2i Pi and

U = ik=1 i Pi , with Pi given via Eqn. (J.4) as

C 2j I U j I

k k

j =1 2 2 = j =1 i j , k > 1

Pi = j 6 =i

i j

j 6 =i (1.242)

I, k = 1.

1 = F k P / . Note that det F =

tensor R is now obtained as FU

The rotation i =1 i i | |

det U = det C.

Now we evaluate the derivatives of these tensors. Differentiating the relation C = UU

using Eqn. (1.233), we get

U

S = S [U I + I U ] S S.

C

Since the eigenvalues i , i = 1, 2, . . . , n, of U are positive, the necessary and sufficient

condition that [U I + I U ] be invertible is satisfied1 . Thus, using (1.197), we get the

unique solution U/C = S/(2) for k = 1, while for k > 1, we have

U

= { S [U I + I U ] S } 1

C " #

k k P P

i j

=S S, (1.243)

+ j

i =1 j =1 i

U U C

=S S

F C F

1 If C is positive semi-definite instead of positive definite, then U/C does not exist when C is singular.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

90 Continuum Mechanics

" #

k

Pi P j kh i

= 2S S FT I

+ j

i =1 j =1 i

" #

k k (P F T ) P

i j

= 2S . (1.244)

i =1 j =1

i + j

By differentiating F = RU, and using Eqns. (1.158), (1.168), (1.185), (1.173), we get

R U

= ( I U ) 1 I ( R I )

F F

k

( P k F T) P

i j

= I U 1 2 [ R U 1 ]S

i =1 j =1

i + j

h i

k k ( RPi F T ) (U 1 P j ) + T (U 1 Pi F T ) ( RP j )

= I U 1

i =1 j =1

i + j

k (i + j )( RPi R T ) (U 1 P j ) i ( RPi R T ) (U 1 P j ) [( RPi ) (P j R T )]T

k

= i + j

i =1 j =1

k ( RPi R T ) P j [( RPi ) (P j R T )]T

k

= i + j

. (1.245)

i =1 j =1

Note that we have interchanged i and j in the last term in the second-to-last step. Now

using Eqn. (1.243), and the fact that U = (U/C )C (or, alternatively, Eqn. (1.244) and the

fact that U = (U/F ) F), and in a similar manner using (1.245), we get U = C/(2) and

T

R T R = ( R T F F R)/(2) for k = 1, while for k > 1,

k k k k P ( F T F + F F ) P T

Pi CP j i j

U = = . (1.246)

i =1 j =1 i

+ j i =1 j =1

i + j

T

Pi ( R T F F R)P j

k k

R = R . (1.247)

i =1 j =1

i + j

ent, then see Eqns. (2.70) and (2.77).

The corresponding results for the left stretch tensor V are

" #

k k G G

V i j

=S S,

B i =1 j =1 i

+ j

" #

k k G (G F )

V i j

= 2S ,

F i =1 j =1

i + j

k k k k G ( FF T + F F ) G T

Gi BG j i j

V = i + j =

i + j

, (1.248)

i =1 j =1 i =1 j =1

k k G ( FR T R F T ) G

" #

i j

R = R, (1.249)

i =1 j =1

i + j

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 91

where Gi = RPi R T are the projections of V , which are obtained using Eqn. (1.242) by

replacing C and U by B and V , respectively. If F is the deformation gradient, then see

Eqns. (2.71) and (2.76).

We now verify the following relations derived by Chen and Wheeler [47] and Wheeler

[352] when n = 3:

T

U

U = F,

F

T

R

F = 0,

F

U 1

L = RT L U ( R T L L T R)UU,

F det U

R 1

L= RU ( R T L L T R)U,

F det U

where L Lin, and

Pi P j = ij P j (no sum on j), PiT = Pi , ik=1 Pi = I, UP j = km=1 m Pm P j = j P j = P j U

(no sum on j), we have

T k k FP j UPi + FPi UP j

U

F

U= i + j

i =1 j =1

" #

k k

P j UPi + Pi UP j

=F

i =1 j =1

i + j

" #

k k (P P + P P )

i j i i j

=F

i =1 j =1

i + j

k

i P i

=F

i =1

i

= F,

RPi R T FP j T(Pi R T FP j R T )

T k k

R

F

F= i + j

i =1 j =1

k k RPi UP j RP j UPi

= i + j

i =1 j =1

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

92 Continuum Mechanics

k k

j R(Pi P j P j Pi )

= i + j

i =1 j =1

k

i R ( P i P i )

= 2i

i =1

= 0,

k k P L T FP + P F T LP

U i j i j

L=

F i =1 j =1

i + j

k Pi L T RUP j + Pi U R T LP j

k

= i + j

i =1 j =1

k j Pi L T RP j + i Pi R T LP j

k

= i + j

i =1 j =1

k k k k j Pi ( R T L L T R)P j

= Pi RT LP j i + j

i =1 j =1 i =1 j =1

! !

k k k k P ( R T L L T R)P

j i j

= Pi RT L Pj i + j

i =1 j =1 i =1 j =1

k k j Pi ( R T L L T R)P j

= RT L

i =1 j =1

i + j

1

= RT L U ( R T L L T R)UU, (by Eqn. (1.250))

det U

R 1 1 T 1 T T

L = ( I U )L (R U ) R L U ( R L L R)UU

F det U

1

= LU 1 RR T LU 1 + RU ( R T L L T R)U

det U

1

= RU ( R T L L T R)U.

det U

diagonalizable tensor T for the case n = 3, and then specialize the results to the case of a

symmetric tensora more comprehensive discussion of the derivatives of the eigenvalues

of a symmetric tensor for the case n = 3 may be found in [344]. By Eqn. (J.25), T can be

expressed as ik=1 i Pi , where k is the number of distinct eigenvalues, Pi are projections

given by Eqn. (J.26), and ik=1 Pi = I. Consider the case when all eigenvalues are distinct,

i.e., 1 6= 2 6= 3 6= 1 . It is known (see, e.g., the remark following (2.2.4) in [37])

that the eigenvalues and associated eigenprojections are differentiable in this case. Using

Eqn. (1.60), we can show, for example, that cof ( T 1 I ) = ( T 2 I ) T ( T 3 I ) T = P1T .

Since det( T I (i ) I ) = det( T i I ) = 0, i = 1, 2, 3 and <, we see that

i , i = 1, 2, 3 are the eigenvalues of T I, <. Thus, (1 1 ), (2 1 ) and

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 93

the eigenvalues of cof ( T 1 I ), which in turn leads to tr cof ( T 1 I ) = (1 2 )(1

3 ). By differentiating the relation det( T i I ) = 0, i = 1, 2, 3, and using the chain rule,

we get

1

(I I )cof ( T 1 I ) = 0, i = 1, 2, 3, (1.251)

T

or, in other words,

1 cof ( T 1 I ) ( T 2 I ) T ( T 3 I ) T

= = = P1T . (1.252)

T tr cof ( T 1 I ) (1 2 )(1 3 )

The corresponding results for 2 /T and 3 /T are obtained by permuting the indices

1, 2 and 3. Equation (1.252) can also be directly obtained by differentiating the relation

31 I1 21 + I2 1 I3 = 0 with respect to T, and using Eqns. (1.226a)(1.226c); this shows

that the second expression for 1 /T in Eqn. (1.252) is valid even for an arbitrary tensor,

provided the denominator is nonzero.

By differentiating the expression P1 = ( T 2 I )( T 3 I )/[(1 2 )(1 3 )] using

Eqns. (1.170), (1.231) and (1.233), we get

P1 1 h

= T I + I T T ( 2 + 3 )I

T (1 2 )(1 3 )

h ii

+ (2 + 3 21 )P1 P1T (2 3 ) P2 P2T P3 P3T .

By using Eqn. (J.29) and the fact that I = 3i=1 Pi = 3i=1 PiT , the above relation simplifies

to

P1 P P2T + P2 P1T P P3T + P3 P1T

= 1 + 1 . (1.253)

T 1 2 1 3

The gradients of P2 and P3 are obtained by cyclically permuting the indices 1, 2 and 3. The

above expressions are valid for an arbitrary tensor with distinct eigenvalues (since such a

tensor is diagonalizable).

Now consider the case when two of the eigenvalues are repeated, i.e., 1 6= 2 = 3 .

Note that now, Eqn. (J.4) yields P1 = ( T 2 I )/(1 2 ). Since P1 P1 = P1 , and since

2 = 3 , (1.252) also holds. Using the fact that 2 + 3 = I1 1 , we get

I TT

(2 + 3 ) = I P1T = 1 . (1.254)

T ( 1 2 )

The gradient of P1 is obtained by setting 2 = 3 and using P2 + P3 = I P1 in Eqn. (1.253)

as

P1 P ( I P1 ) T + ( I P1 ) P1T

= 1 .

T ( 1 2 )

Since P2 + P3 = I P1 , we get

P

(P2 + P3 ) = 1 .

T T

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

94 Continuum Mechanics

When all three eigenvalues are repeated, the gradient of the sum of the eigenvalues is

I, but the individual gradients are undefined. Since symmetric second-order tensors are

diagonalizable, all the above results can be applied to them by setting Pi = PiT . For a

generalization to any underlying space dimension n, see Eqn. (J.27).

The foregoing development can be used to find the gradients of scalar or tensor-valued

functions of a diagonalizable tensor T = ik=1 i Pi , where (i , Pi ), i = 1, 2, . . . , k, are the

distinct eigenvalues and corresponding eigenprojections of T (see, e.g., [37], [47], [104],

[127], [244], [272], [352] and [355] for various other methods). For example, if G ( T ) =

ik=1 f (i )Pi is a tensor-valued function of T, then using the above results, we get G/T =

f 0 ()I for the case k = 1, while for k > 1,

k k k f ( ) f ( )

G f i j

= Pi PiT + Pi P Tj . (1.255)

T i =1

i i =1 j =1

i j

j 6 =i

The above formula is derived by first considering the case of distinct eigenvalues (i.e., k =

n), and then appropriate limits are taken (e.g., 3 2 in case 2 = 3 ) to obtain the

results for repeated eigenvalues. By virtue of Eqn. (J.27), the above result holds for any

arbitrary underlying space dimension n. In case G (S) = ik=1 f (i )Pi is a symmetric-

valued function of S Sym, then the above result reduces to G/S = f 0 ()S for the case

k = 1, while for k > 1,

" #

k k k f ( ) f ( )

G f i j

=S P Pi + Pi P j S. (1.256)

S i =1

i i i =1 j =1

i j

j 6 =i

An application of Eqn. (1.256) can be found towards the end of Section 2.4. We now find

the eigenvalue/eigentensors of G/S in the above equation. First consider the case when

all the eigenvalues i of S are distinct. Then, if ni represent the eigenvectors of S, each Pi

can be represented as ni ni (no sum on i). Using Eqns. (1.172) and (1.342), the eigenval-

ues/eigentensors (the eigentensors have not been normalized) are found to be

f f f

, n n1 , , n2 n2 , , n3 n3 ,

1 1 2 3

f ( 1 ) f ( 2 ) f ( 2 ) f ( 3 )

, n1 n2 + n2 n1 , , n2 n3 + n3 n2 , (1.257)

1 2 2 3

f ( 1 ) f ( 3 )

, n1 n3 + n3 n1 , 1 6 = 2 6 = 3 6 = 1 .

1 3

Next consider the case when two eigenvalues of S are repeated, say, 1 6= 2 = 3 . We

now have

G f f

=S P P1 + ( I P1 ) ( I P1 )

S 1 1 2

f ( 1 ) f ( 2 )

+ [P1 ( I P1 ) + ( I P1 ) P1 ] S.

1 2

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 95

f f

, n1 n1 , , I n1 n1 ,

1 2

f ( 1 ) f ( 2 )

, n1 [ n ( n n1 ) n1 ] + [ n ( n n1 ) n1 ] n1 , 1 6 = 2 = 3 ,

1 2

(1.258)

where n is such that n n1 6= 0, but is otherwise arbitrary. Finally, for the case when all

eigenvalues are repeated, i.e., 1 = 2 = 3 , the distinct eigenvalue/eigentensor pair

is

f

, A , 1 = 2 = 3 , (1.259)

where A Sym.

The derivatives of the eigenvectors {ei } of S Sym are given by [244]

e1 e e1 e2 + e2 e2 e1 e e1 e3 + e3 e3 e1

= 2 + 3 ,

S 2( 1 2 ) 2( 1 3 )

e2 e e1 e2 + e1 e2 e1 e e2 e3 + e3 e3 e2

= 1 + 3 ,

S 2( 2 1 ) 2( 2 3 )

e3 e e1 e3 + e1 e3 e1 e e2 e3 + e2 e3 e2

= 1 + 2 .

S 2( 3 1 ) 2( 3 2 )

Similar to the gradient of a scalar field, we define the gradient of a vector field v as

The terminology gradient used for the above quantity is misleading since the gradient

was defined in terms of the inner product, whereas the above quantity is not [58]. The

quantity v is in reality simply the matrix representation of the Frechet derivative of v,

and hence a second-order tensor. Taking u = e j , and taking the dot product of the above

equation with ei , we get the components of v as

vi

(v)ij = .

x j

Thus, we have

vi

v = e ej.

x j i

vi v v v

v := tr v = tr (ei e j ) = i ei e j = i ij = i , (1.261)

x j x j x j xi

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

96 Continuum Mechanics

to the gradient of a vector field as

Tij

T = e e j ek .

xk i

The divergence of a second-order tensor T, denoted as T, is defined as

( T ) u := ( T T u) constant u V, (1.263)

Tij

T = e.

x j i

( v) u := [v (v)T ]u u V. (1.264)

Thus, v is the axial vector corresponding to the skew tensor [v (v) T ]. In com-

ponent form, we have

vk v

v = eijk (v)kj ei = eijk e = k e j ek .

x j i x j

( T )u := ( T T u) constant u V. (1.265)

In component form,

Tjs Tjs

T = eirs e ej = ( er e s ) e j . (1.266)

xr i xr

The Laplacian of a scalar function ( x) is defined by

2 := (). (1.267)

2

2 = .

xi xi

The Laplacian of a tensor function T ( x), denoted by 2 T, is defined by

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 97

In component form,

2 Tij

(2 T )ij = .

xk xk

In the remaining chapters, we shall assume that a function is sufficiently smooth (with-

out explicitly stating what these smoothness properties are), so that all the equations pre-

sented make sense. For example, in the equation

+ b = 0,

if we are given that and b are continuous functions of the position vector x, then we

implicitly assume to be continuously differentiable.

1.9.5 Examples

Although it is possible to derive tensor identities involving differentiation using the above

definitions of the operators, the proofs can be quite cumbersome, and hence we prefer to

use indicial notation instead. In what follows, u and v are vector fields, and x ei (this

i

is to be interpreted as the del operator acting on a scalar, vector or tensor-valued field,

e.g., = x ei ):

i

1. Show that

= 0. (1.269)

2. Show that

1

(u u) = (u)T u. (1.270)

2

4. Show that

2

u := (u) = ( u) ( u). (1.271b)

is harmonic.

5. Show that (u v) = v ( u) u ( v).

6. Let W Skw, and let w be its axial vector. Then show that

W = w,

W = ( w) I w. (1.272)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

98 Continuum Mechanics

Solution:

1. Consider the ith component of the left-hand side:

2

( )i = eijk

x j xk

2

= eikj (interchanging j and k)

xk x j

2

= eijk ,

x j xk

which implies that = 0.

1 1 ( ui ui ) u

2. (u u) = e j = ui i e j = (u) T u.

2 2 x j x j

3. We have

[(u)v] = j ((u)v) j

u j

= v

x j xi i

vi u j 2 u j

= + vi

x j xi xi x j

= (u)T : v + v ( u).

T u j

[ (u) ]i =

x j xi

!

u j

=

xi x j

= [( u)]i .

To prove the second identity, consider the last term

( u) = eijk j ( u)k ei

= eijk j (ekmn m un )ei

2 u n

= eijk emnk e

x j xm i

2 u n

= (im jn in jm ) e

x j xm i

" #

2 u j 2 u i

= e

xi x j x j x j i

= ( u) (u).

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Introduction to Tensors 99

5. We have

( u v )i

(u v) =

xi

(u j vk )

= eijk

xi

u j v

= eijk vk + eijk u j k

xi xi

u j v

= ekij vk e jik u j k

xi xi

= v ( u) u ( v).

wk

( W ) = eijk e

x j i

= w.

Wjn

( W )ij = eimn

xm

wr

= eimn e jnr

xm

ur

= (ij mr ir mj )

xm

ur u

= i,

xr ij x j

which is the indicial version of Eqn. (1.272).

The exponential of a tensor T (t) can be defined either in terms of its series representation

as

1

e T (t) : = I + T ( t ) + [ T (t)]2 + , (1.273)

2!

or in terms of a solution of the initial value problem

X ( ) = T (t) X ( ) = X ( ) T (t), > 0, (1.274)

X (0) = I, (1.275)

for the tensor function X ( ), where the parameter is independent of t. Note that the su-

perposed dot in the above equation denotes differentiation with respect to . The existence

theorem for linear differential equations tells us that this problem has exactly one solution

X : [0, ) Lin, which we write in the form

X ( ) = e T (t) .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

T

e[T (t)] = (e T (t) ) T , (1.276)

1 BA )

and that if A Lin is invertible, then e( A = A1 e B A for all B Lin.

it follows that (ei (t) , n) is an eigenvalue/eigenvector pair of e T (t) . Hence, the

determinant of e T (t) , which is just the product of the eigenvalues, is given by

n

det(e T (t) ) = in=1 ei (t) = ei=1 i (t) = etr T (t) .

then A(t) B(t) = B(t) A(t). Conversely, if A(t) B(t) = B(t) A(t), then

Proof. We shall write A(t) and B(t) simply as A and B for notational conve-

nience. Let X A ( ) := e A , X B ( ) := e B , and X A+ B := e( A+ B) . Then, we have

X A+ B = ( A + B) X A+ B , (1.280)

X AX B = X A X B + X A X B

= AX A X B + X A BX B , (1.281)

Let Z := X A+ B X A X B . From Eqns. (1.280) and (1.281), we get

Z = ( A + B) X A+ B AX A X B X A BX B . (1.282)

Z = ( A + B) [ X A+ B X A X B ] = ( A + B) Z. (1.283)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Z (0) = 0. The general solution of the differential equation is Z = e( A+ B) C,

where the tensor C is a constant tensor. On using the initial condition, we get

C = 0, from which it follows that Z = 0, thus proving Eqn. (1.279). In an

analogous fashion, by taking Z := X A+ B X B X A , one can also show that if

AB = BA, then e( A+ B) = e B e A . Thus, Eqn. (1.279) follows.

Conversely, if Eqn. (1.278) holds, i.e., if X A+ B = X A X B , then Eqn. (1.282)

reduces to

BX A X B = X A BX B .

B ,

we get BX A = X A B for all . Differentiating this relation with respect to , and

evaluating at = 0, we get BA = AB.

A(t) B(t) = B(t) A(t) = e A(t)+ B(t) = e A(t) e B(t) = e B(t) e A(t) . (1.284)

However, the converse of the above statement may not be true. Indeed, if AB 6= BA, one

can have e A+ B = e A = e B = e A e B , or e A e B = e A+ B 6= e B e A or even e A e B = e B e A 6= e A+ B as

the following examples show:

1. Wood [354] presents the following example:

3

23

0 0 2 0 0

1 ,

A = 2 0 0 B = 2 0 0 12 ,

2

3 1 3 1

2 2 0 2 2 0

Another example presented by Horn and Johnson [136] is

0 0 0 0 0 0 1 0

0 0 0 0 0 0 0 1

A= , B = ,

0 0 0 2 0 0 0 2

0 0 2 0 0 0 2 0

are (0, 0, 2i, 2i ).

2. Wermuth [350] presents the following example where e A e B = e A+ B 6= e B e A : if z =

a + ib is a solution of ez z = 1, e.g., a = 2.088843 . . . and b = 7.461489 . . ., then for

0 0 0 0 0 0 1 0

0 0 0 0 0 0 0 1

A= , B = ,

0 0 a b 0 0 0 0

0 0 b a 0 0 0 0

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

we have

1 0 0 0 1 0 1 0

0 1 0 0 0 1 0 1

eA = , eB = , (1.285)

0 0 a+1 b 0 0 1 0

0 0 b a+1 0 0 0 1

so that

1 0 1 0 1 0 a+1 b

0 1 0 1 0 1 b a + 1

e A e B = e A+ B = , eB e A = .

0 0 a+1 b 0 0 a+1 b

0 0 b a+1 0 0 b a+1

0 0 0 0 0 1 0

0 0 0 0 0 0 1

A= , B = ,

0 0 0 0 0 0 0

0 0 0 0 0 0 0

which yields

1 0 1 0

0 1 0 1

e A = e A+ B = I, eB = ,

0 0 1 0

0 0 0 1

so that e A e B = e B e A 6= e A+ B .

As an application of Eqn. (1.284), since T (t) and T (t) commute, we have e T (t)T (t) =

I = e T (t) eT (t) . Thus,

( e T ( t ) ) 1 = e T ( t ) . (1.286)

As an application of Eqn. (1.286), consider the solution of the tensorial differential equa-

tion

X + AX + X B T = F (t), (1.287)

subject to the initial condition X (0) = C. The differentiation is with respect to t, and A and

B are assumed to be independent of t. Multiplying Eqn. (1.287) by e At e Bt , we get

T Tt T

(e At e Bt ) X + Ae At Xe B t + e At X B T e B = e At F (t)e B t .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

d h At i T

e e Bt X = e At F (t)e B t .

dt

Integrating the above equation between the limits [0, t], we get

Z t

T

(e At e Bt ) X (t) = e A F ( )e B d + C,

0

Z t

T (t ) Tt

X (t) = e A(t ) F ( ) e B d + e At Ce B

0

Z t

T T

= e A F (t )e B d + e At Ce B t . (1.288)

0

e( A I + I B)t = e At e Bt t.

The result given by Eqn. (1.288) also holds if B = 0, and X, F and C are vectors. This

observation allows us to solve the tensorial differential equation

X + AX B = F (t),

X + ( A B T ) X = F (t),

( X ) + ( A B T )( X ) = ( F ),

Z t

e( AB [ F (t )] d + e( AB

T ) T )t

[ X (t)] = (C ).

0

Theorem 1.10.3. Let W (t) Skw for all t. Then eW (t) is a rotation for each t 0.

Conversely, for R(t) Orth+ , there exists a W (t) Skw such that R(t) = eW (t) .

Proof. By Eqn. (1.286),

T (t)

( e W ( t ) ) 1 = e W ( t ) = e W = ( eW ( t ) ) T ,

where the last step follows from Eqn. (1.276). Thus, eW (t) is a orthogonal tensor.

By Theorem 1.10.1, det(eW (t) ) = etr W (t) = e0 = 1, and hence eW (t) is a rotation.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

The converse is proved simply by taking W (t) = log R(t), which we show to

be a skew-symmetric tensor later in this section.

|w(t)|2 W (t), where w(t) is the axial vector of W (t). Thus, W 4 (t) = |w(t)|2 W 2 (t),

W 5 (t) = |w(t)|4 W (t), and so on. Substituting these terms into the series expansion of the

exponential function, and using the representations of sine and cosine functions, we get2

sin(|w(t)|) [1 cos(|w(t)|)] 2

R ( t ) = eW ( t ) = I + W (t) + W ( t ). (1.289)

|w(t)| |w(t)|2

The directional derivative DR(W )[U ] can be obtained from the above expression using the

fact that |w|2 = w w = W : W /2, so that 2 |w| D |w| (W )[U ] = W : U.

Another proof of Eqn. (1.289) is obtained by assuming R( ) eW (t) to be of the form

h1 ( ) I + h2 ( )W + h3 ( )W 2 , and then determining the unknown functions h1 ( ), h2 ( ),

and h3 ( ) by using the governing equation R = RW (where the dot denotes differentiation

with respect to ) subject to the initial conditions R(0) = I and R(0) = W. By using the

fact that W 3 (t) = |w(t)|2 W (t), and that { I, W, W 2 } is a linearly independent set (see

Problem 4), we get

h1 ( ) = 0,

h2 ( ) = h1 |w(t)|2 h3 ( ),

h3 ( ) = h2 ( ),

which are to be solved subject to the initial conditions

h1 (0) = 1, h1 (0) = 0,

h2 (0) = 0, h2 (0) = 1,

h3 (0) = 0, h3 (0) = 0.

We get the solution as

sin(|w(t) |) [1 cos(|w(t) |)]

h1 ( ) = 1, h2 ( ) = , h3 ( ) = ,

|w(t)| |w(t)|2

2 Similarly,

in the two-dimensional case, if

" #

0 (t)

W (t) = ,

(t) 0

where is a parameter which is a function of t, then

" #

sin (t) cos (t) sin (t)

R(t) = eW (t) = cos (t) I + W (t) = .

(t) sin (t) cos (t)

Rt

Let (t) = 0 ( ) d. Then the solution of dz/dt = W (t)z in this two-dimensional case is

" #

cos (t) sin (t)

z(t) = z (0).

sin (t) cos (t)

This result is a special case of the result in Theorem 1.10.4.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Yet another proof is obtained by using Eqn. (1.292) with T W (t), = 0, = i |w(t)|,

= i |w(t)| and = 1 (which implies that the eigenvalues of eW (t) are (1, ei|w(t)| , ei|w(t)| )).

Not surprisingly, Eqn. (1.289) has the same form as Eqn. (1.104) with = |w(t)|. Equa-

tion (1.289) is known as Rodrigues formula.p If W (t) is a complex-valued skew-symmetric

tensor, replace |w(t)| in Eqn. (1.289) by w(t) w(t) if it is nonzero; if it is zero, then W

is nilpotent, and using the series expansion, we get eW (t) = I + W (t) + W 2 (t)/2. A gen-

eralization of Rodrigues formula to the case when n 4 was developed by Gallier and

Xu [86]; we shall present this formula later in this section (see Eqn. (1.295)), although we

use a different method for deriving it than the one that they have used.

Let a = sin(|w(t)|)/ |w(t)|, b = [1 cos(|w(t)|)]/ |w(t)|2 and c = |w(t)|2 , and let

T

R(t) be given by Eqn. (1.289). Using the facts RR T + R R = 0, 2b b2 c a2 = 0,

W WW = cW /2, and W WW 2 W 2 WW = 0 (where now the superposed dot denotes

differentiation with respect to t), we get

1

RR T = (w w)(1 a)W + aW + b(W W WW ), (1.290a)

c

1

R T R = (w w)(1 a)W + aW b(W W WW ). (1.290b)

c

The axial vectors of RR T and R T R are, thus, (w w)(1 a)w/c + aw + b(w w) and

(w w)(1 a)w/c + aw b(w w), respectively. Of course, for the special case when

W (t) = W 0 t, where W 0 is independent of t, the above formulae reduce to RR T = R T R =

W 0 . In this connection, one has the following theorem [354]:

d ( e T (t) )

= Te T (t) = e T (t) T, (1.291)

dt

Rt

if and only if T T = T T. Thus, the solution of Z = T (t) Z is Z (t) = e 0 T ( ) d Z (0)

if and only if T T = T T.

tives of higher powers of T. Thus, by simply differentiating Eqn. (1.273), we get

Eqn. (1.291).

To prove the converse, note from Eqn. (1.273) that e T T = Te T . Differentiating

both sides of this relation, and using the fact that Eqn. (1.291) holds, we get

Te T T = T Te T .

get the desired result.

Rt

By replacing T (t) by 0 T ( ) d in the first relation in Eqn. (1.291), and post-

multiplying by Z (0), we get the stated solution of the differential equation

Z = T (t) Z.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

Explicit formulae for e T (where T can be complex-valued) when n = 3 and n = 4 are

given by Cheng and Yau [52] in terms of the eigenvalues and powers of T.

1. 1 = 2 = 3

(a) If q( T ) = ( T I ), then

e T = e I.

(b) If q( T ) = ( T I )2 , then

e T = e [ ( T I ) + I ] .

(c) If q( T ) = ( T I )3 , then

2

T 2

e =e

( T I ) + ( T I ) + I .

2

2. 1 6= 2 = 3

(a) If q( T ) = ( T I )( T I ), then

e e e e

e T = T+ I.

(b) If q( T ) = ( T I )( T I )2 , then

e e

T e

e = ( T I )2 + e [ ( T I ) + I ] .

( )2

3. 1 6= 2 6= 3 6= 1 . We have q( T ) = ( T I )( T I )( T

I ), and

( T I )( T I ) ( T I )( T I ) ( T I )( T I )

e T = e + e + e .

( )( ) ( )( ) ( )( )

(1.292)

We now present two methods for the explicit determination of the exponential function

for arbitrary n (for a detailed survey, see [225]). The first method is based on the Jordan

canonical form given by Eqn. (J.37), while the second method is based on the definition

given by Eqn. (1.274). The first method yields the formula

m i 1

" !#

k

j j

e T

= e i

Pi + N

j! i

, (1.293)

i =1 j =1

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

where mi is as in Eqn. (J.24). To prove this result, we first note that Pi and N i commute since

they are polynomials in T. Using Eqns. (J.37) and (1.279), and the properties Pi P j = ij Pi

j

(no sum on i), N i N j = Pi N j = 0 for i 6= j, Pi N i = N i (no sum on i), and N i = 0, j mi ,

we have

k k

e T = ei=1 i Pi e i =1 N i

k m i 1 j

" #" #

k

j

= e Pi I +

i

N

i =1 i =1 j =1

j! i

m i 1 j

" !#

k

j

= e i

Pi + N .

i =1 j =1

j! i

e e e +( 1)e

e ( )2

e T = e P1 + e (P2 + N 2 ) =

.

0 e e

0 0 e

The exponential tensor e T for a diagonalizable (in particular, symmetric) tensor T is ob-

tained either as a special case of Eqn. (1.293), or by substituting Eqn. (J.25) into Eqn. (1.273),

and is given by

k

e T = e i P i , (1.294)

i =1

with Pi given by Eqn. (J.26). The method proposed by Putzer [259] essentially yields the

same result as that given by Eqn. (1.293), although more indirectly.

Writing m as (r )m + i (s )m , we can write Eqn. (1.293) as

k h i

eT = e(r )m sin((s )m ) Bm cos((s )m ) B2m + [cos((s )m ) + i sin((s )m )] G m ,

m =1

j

where Bm = iPm and G m = nj=m11 j!1 N m . Note that B3m = iPm = Bm . The above

form allows us to write a Rodrigues-type formula for any dimension n when T is a skew-

symmetric tensor [86]. As discussed earlier, the eigenvalues of a real-valued W Skw

are either zero or of the form im , m = 1, 2, . . . , k, where k is the number of distinct

(1) (2)

nonzero eigenvalues. Also, since W is normal, it is diagonalizable. Let Bm := i [Pm Pm ],

(1) (2)

where Pm and Pm are the projections associated with the eigenvalues im and im ,

(1) (2)

respectively. Since B2m = [Pm + Pm ], it follows that I = km=1 B2m . From Eqn. (1.294),

we get

k h i

eW = I + sin(i ) Bi + [1 cos(i )] B2i , (1.295)

i =1

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

where W = ik=1 i Bi , with each Bi being a real-valued skew tensor. The i are determined

by taking the square roots of the eigenvalues of W 2 , while the matrices Bi are determined

by solving the following system of k equations

k

W= i Bi ,

i =1

k

W 3 = 3i Bi ,

i =1

,

k

(1)k+1 W 2k1 = 2k

i

1

Bi ,

i =1

1 2 ... k

3

32 ... 3k

1

. . . . . . . . . . . . . . . . . . . . . . . . . . . ,

. . . . . . . . . . . . . . . . . . . . . . . . . . .

1 1 1

2k

1 2k2 . . . 2k k

to obtain

W 2 +2j I

(1)k+1 W k

, k>1

i j =1 2i 2j

B i (W ) = j 6 =i

W ,

k = 1.

(which is unitary) since, again, in this case, the real part of the eigenvalues and the Gi s are

zero. We have already noted that some complex-valued skew-symmetric tensors can be

nilpotent, and hence, no Rodrigues-type formula can be derived for its exponential (which

is an orthogonal tensor), in general.

The second method that we discuss is the one devised by Fulmer [85]. Note that since

dk /d k (e T ) = T k e T , for k = 1, 2, . . . , n, we have

" #

T

dk T

e = I and (e ) = T k. (1.296)

=0 d k

=0

( T n I1 T n1 + + (1)n In I )e T = 0,

or, alternatively,

dn d n 1

n T

I1 + + ( 1 ) In e = 0. (1.297)

d n d n1

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

As a result of Eqns. (1.296) and (1.297), we see that e T is the unique solution of the nth

order initial value problem

dk G ( )

DG ( ) = 0, G (0) = I, = T k , k = 1, 2, . . . , n 1, (1.298)

d k

=0

If T has the characteristic polynomial given by Eqn. (J.23), then the general solution of

DG ( ) = 0 is

G ( ) = e1 (C 11 + C 12 + + n1 1 C 1n1 ) + + ek (C k1 + C k2 + + nk 1 C knk ),

(1.299)

I = C 11 + C 21 + + C k1 ,

T = (1 C 11 + C 12 ) + + (k C k1 + C k2 ),

T 2 = (21 C 11 + 21 C 12 ) + + (2k C k1 + 2k C k2 ),

( n 1 ) ! n n1

T n1 = 1n1 C 11 + (n 1)1n2 C 12 + + 1 C 1n1 +

( n n1 ) !

( n 1) ! n n k

+ nk 1 C k1 + (n 1)nk 2 C k2 + + k C knk .

(n nk )!

above equations; substituting these expressions into Eqn. (1.299), we get the desired ex-

pression for e T . Note that the above method is based on the characteristic polynomial3 ,

whereas the first method is based on the minimal polynomial.

As an illustration, let us again find the expression for e T , with T given by Eqn. (J.39).

Since the characteristic polynomial is given by ( T I )( T I )2 , the general solution is

G ( ) = e C 11 + e (C 21 + C 22 ), where the C ij are obtained from the initial conditions

I = C 11 + C 21 ,

T = C 11 + C 21 + C 22 ,

T 2 = 2 C 11 + 2 C 21 + 2C 22 .

sion as before.

Now we discuss the differentiation of e T with respect to T, where T Lin, for any

underlying space dimension n; the treatment is based on [155], alternative treatments may

be found in [143] and [144]. We shall first give an alternate derivation of a formula that

3 Actually, the method can be modified quite easily to work with the minimal polynomial as wellin this case

the number of C ij matrices to be determined is m = ik=1 mi instead of n. This is particularly advantageous, for

example, if T is diagonalizable, since then mi = 1 for i = 1, 2, . . . , k, so that C ij = 0 for i = 1, 2, . . . , k and j 2.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

appears in Mathias [211] and Ortiz et al. [241]. Let H X/T, where X = e T . Differen-

tiating the expression X = T X with respect to T using Eqn. (1.233), we get

H = ( T I ) H + I X T .

d h T i

(e I ) H = eT (e T )T .

d

Integrating this equation subject to the condition H | =0 = 0, and using Eqn. (1.286), we

get

Z h i

H ( ) = (e T I ) eT (e T ) T d

0

Z h i

= e T (t ) (e T ) T d.

0

Z xi h i

T (t )

H ( ) = e T e T d.

0

Since e T /T = H | =1 , we obtain

(e T )

Z 1h i

= e T (1 ) (e T ) T d

T 0

Z 1 T

= e T e T (1 ) d. (1.300)

0

Explicit expressions for e T such as those given by Eqn. (1.293) (or those presented in The-

orem 1.10.5 for n = 3) should be substituted into the above expressions to get explicit

formulae for e T /T. Note that these formulae are valid for any T Lin and any n.

Simpler expressions for a diagonalizable tensor T can be obtained using Eqns. (1.294)

and (1.300), or directly by using Eqn. (1.255) with f (i ) = ei . The final result that we

obtain for a diagonalizable tensor (for any n) is e I for k = 1, and

" #

k k k

(e T ) e i e j

= e Pi Pi +

i T T

P Pj , k > 1

T i =1 i =1 j =1

i j i

j 6 =i

where k is the number of distinct eigenvalues of T, and Pi is given by Eqn. (J.26). When

T Sym, we obviously have PiT = Pi , and the above result should be pre- and post-

multiplied by S.

In what follows, we shall continue to denote T (t) as T; the treatment is based on [158].

Recall that difficulties arise in defining a unique scalar-valued logarithmic function since

e = e+2in where n is an integer. Similar difficulties arise in defining the logarithm of a

tensor. If one defines the solution X of e X = T as log T, then a non-singular T may have

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

an infinite number of real and complex logarithms. In order to avoid this nonuniqueness

and ensure that the logarithm of a real tensor is real, we define the principal logarithmic

function, denoted by log T, of a matrix T with no eigenvalues that are negative or zero (i.e.

/ (, 0], i = 1, 2, . . . , n), as the unique function such that elog T = T. The eigenvalues

i

of log T have imaginary parts that lie strictly between and , since as we shall show

below, if rei , < < , are the eigenvalues of T, then log(rei ) = log r + i, < <

, are the eigenvalues of log T. Thus, if i denote the eigenvalues of T, then

/ (, 0] i.

1. elog T = T if and only if i

2. log e T = T if and only if < Im(i ) < i.

In series form the logarithmic function is given by

(1)i1

log( I + T ) = i

( T )i , (1.301)

i =1

or, alternatively as

(1)i1

log( T ) = i

( T I )i . (1.302)

i =1

The series given by Eqn. (1.301) is absolutely convergent only if k T k < 1. It is immediately

evident from the above representation that log( I ) = 0 and log( T T ) = (log T ) T . Since,

1

for all invertible A Lin, e A BA = A1 e B A for all B Lin, we get log( A1 BA) =

1

A (log B) A.

If H ( ) : [0, 1] Lin is a tensor-valued function of with no eigenvalues on the closed

negative real axis for any [0, 1], an alternative definition can be given as the solution of

the differential equation

dX dH

H ( ) = , X (0) = 0, (1.303)

d d

eigenvalues of T on the closed negative real axis, then, from the above definition, we get

Z

log[ I + ( T I ) ] = [ I + ( T I ) ]1 ( T I ) d. (1.304)

0

If one violates the restriction on the eigenvalues of T, then, from the above definition, one

can get a complex-valued logarithm even for a real tensor as is seen, for example, by taking

T = diag[1, 1, 1]. Let (, n) denote an eigenvalues/eigenvector pair of T. Then, by using

the fact that

In = [ I + ( T I ) ]1 [ I + ( T I ) ]n = (1 + ( 1) )[ I + ( T I ) ]1 n,

1

Z

log[ I + ( T I ) ]n = d n = log[1 + ( 1) ]n,

0 1 + ( 1)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

I ) ]. In particular, for = 1, (log , n) are the eigenvalue/eigenvector pair of log T. Simi-

larly, one can show that ( log , n) are the eigenvalue/eigenvector pair of log( T 1 ). Anal-

ogous to Eqn. (1.277), we get

n

tr log T = log i = log(det T ). (1.305)

i =1

Theorem 1.10.6. Let A, B be such that no eigenvalues of A, B and AB are on the closed

negative real axis. If

[0, 1] (1.306)

then AB = BA.

Conversely, if AB = BA (so that A, B and AB have a common set of eigenvec-

tors), and if < Im(log i + log i ) < for each i,a where i and i denote the

eigenvalues of A and B corresponding to the same eigenvector, then

log {[ I + ( A I ) ][ I + ( B I ) ]} = log {[ I + ( B I ) ][ I + ( A I ) ]} =

log[ I + ( A I ) ] + log[ I + ( B I ) ] [0, 1] (1.307)

Eqn. (1.303), we get

[ I + ( B I ) ]1 [ I + ( A I ) ]1 [ A + B 2I + 2 ( A I )( B I )] =

[ I + ( A I ) ] 1 ( A I ) + [ I + ( B I ) ] 1 ( B I ).

Differentiating the above relation once again with respect to (using the fact that

d( T 1 )/d = T 1 [dT/d ] T 1 ), and evaluating the resulting expression at =

0, we get AB = BA.

Conversely, given that AB = BA, using Eqn. (1.303), we have

log {[ I + ( A I ) ][ I + ( B I ) ]}

Z

= [ I + ( B I ) ]1 [ I + ( A I ) ]1 [ A + B 2I + 2 ( A I )( B I )] d

0

Z

= [ I + ( B I ) ]1 [ I + ( A I ) ]1 {[ I + ( A I ) ]( B I ) + [ I + ( B I ) ]

0

( A I )} d

Z Z

=l [ I + ( B I ) ]1 ( B I ) d + [ I + ( A I ) ]1 ( A I ) d

0 0

= log[ I + ( A I ) ] + log[ I + ( B I ) ].

log {[ I + ( B I ) ][ I + ( A I ) ]}, leading to Eqn. (1.307).

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

a This condition is imposed since, as stated, the imaginary part of the eigenvalues of the log of any

The above implication need not be true if AB = BA, but the constraints on the eigenvalues

of A and B are violated. Cheng et al. [53] have presented the following (complex-valued)

scalar example: if a = b = e( e)i , where e is small and positive, then

Wood [354]. If

3

0 0 2 0.499963 0.866004 0.00866011

A = ( 0.01) 0 0 21 , H = e A 0.866004 0.500012 0.00499992 ,

23 21 0 0.00866011 0.00499992 0.99995

then

0 0 2.71204 0 0 0.0173205

2

log H 0 0 1.5658 , log H 0 0 0.01 ,

so that log H 2 6= 2 log H. The eigenvalues of log H and log H 2 are (approximately) (0, (

0.01)i, ( 0.01)i ) and (0, 0.02i, 0.02i ), respectively.

The converse assertion of Eqn. (1.308) is true for dimension n = 2, as seen in [226, 227],

and also for symmetric tensors (for any n), as is evident by taking the transpose. For n = 4,

it may appear that by taking X = e A , Y = e B , where e A and e B are given by Eqn. (1.285), we

have log( XY ) = log X + log Y, but XY 6= Y X. However, since both A and A + B do not

satisfy the constraint < Im(i ) < , we have log(e A ) 6= A and log(e A+ B ) 6= A + B.

Thus, the question of whether the converse assertion of Eqn. (1.308) is true for n > 2

remains unresolved.

Let T be a tensor with no eigenvalues i on the closed negative real axis. Since T and

T 1 commute, and since log i + log(i )1 = 0, by taking A T and B T 1 , we get

If R Orth+ is such that it does not have any eigenvalues on the closed negative real axis,

i.e., if R Orth+ /Sym + { I }, then log R Skw. This is proved by noting that

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

method may be found in [34]. Analogous to Eqn. (J.38), one can prove that

m i 1

" j #

k

1 N

[ I + ( T I ) ] 1 = Pi + i

.

i =1

[1 + ( i 1) ] j =1

[1 + ( i 1) ]

m i 1 j

( )

Z k

i 1 (1) j1 j1 N i

log[ I + ( T I ) ] =

0 i =1 [ 1 + ( i 1 ) ]

Pi + j +1

d

j =1 [ 1 + ( i 1 ) ]

m i 1 j

( )

k (1) j1 j N i

= log[1 + (i 1) ]Pi + j

. (1.310)

i =1 j =1 j [ 1 + ( i 1 ) ]

m i 1 j

( )

k (1) j1 N i

log T = (log i )Pi + j

. (1.311)

i =1 j =1 ji

An alternative derivation of Eqn. (1.310) can be given as follows. Since

m 1

{P1 , P2 , . . . , Pk , N 1 , . . . , N 1m1 1 , . . . , N k , . . . , N k k }

constitutes a basis for any tensor, we write

m i 1

" #

k

log[ I + ( T I ) ] = f i ( )Pi + f ij ( )( N i ) j

,

i =1 j =1

where the f i ( ) and f ij ( ) are functions to be determined. Substituting into the definition

given by Eqn. (1.303), and using the linear independence of the basis, we get for each i the

equations

[1 + (i 1) ] fi = (i 1),

[1 + (i 1) ] fi1 + fi = 1,

[1 + (i 1) ] fi2 + fi1 = 0,

..

.

[1 + (i 1) ] fi(mi 1) + fi(mi 2) = 0,

which are to be solved subject to the condition that f i (0) = f ij (0) = 0. We first solve for f i

using the first equation, then for f i1 using the second one, and so on. The solution is given

by f i = log[1 + (i 1) ] and f ij = (1) j1 j /[ j(1 + (i 1) ) j ], which yields Eqn. (1.310).

For a diagonalizable tensor T, Eqn. (1.311) simplifies to

k

log T = (log i )Pi , (1.312)

i =1

with Pi given by Eqn. (J.26). By specializing Eqn. (1.311) to the case n = 3, we get the

following analogue of Theorem 1.10.5:

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

1. 1 = 2 = 3

(a) If q( T ) = ( T I ), then

log T = (log ) I.

(b) If q( T ) = ( T I )2 , then

T I

log T = (log ) I + .

(c) If q( T ) = ( T I )3 , then

T I ( T I )2

log T = (log ) I + .

22

2. 1 6= 2 = 3

(a) If q( T ) = ( T I )( T I ), then

T I T I

log T = log + log .

(b) If q( T ) = ( T I )( T I )2 , then

( T I )2 ( T I )( T (2 ) I ) ( T I )( T I )

log T = log log + .

( )2 ( )2 ( )

3. 1 6= 2 6= 3 6= 1 . We have q( T ) = ( T I )( T I )( T

I ), and

( T I )( T I ) ( T I )( T I ) ( T I )( T I )

log T = log + log + log .

( )( ) ( )( ) ( )( )

1

e 1 0 1 e 0

T = 0 e 0 , log T = 0 1 0 , (Case 1(b)),

0 0 e 0 0 1

1 2e1

e 1 1 1 e 2e2

T = 0 e 1 , log T = 0 1 , (Case 1(c)),

1 e

0 0 e 0 0 1

1 e3 e2 1

e 1 1 1 e ( e 1) e3 ( e 1)2

T = 0 e2 1 , log T = 0 1 , (Case 2(b)).

2 e2

0 0 e2 0 0 2

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

(and hence tr R = 1 + 2 cos ) is obtained using Case 3 in the above theorem as

log R = i(P2 P3 )

i( R I ) R ei I R ei I

= i +

e ei ei 1 ei 1

= ( R RT )

2 sin

R RT

= 2
.

R RT

By substituting W log R (|w| = ) into the right-hand side of Eqn. (1.289), we can easily

verify that elog R = R.

By differentiating Eqn. (1.304) with respect to T and using Eqns. (1.233) and (1.236), we

get (for any underlying space dimension n)

Z h

i

log( I + ( T I ) ) = ( I + ( T I ) )1 ( I + ( T I ) )T d.

T 0

log( I + ( T I ) ) =

T

k k k log[1 + ( 1) ] log[1 + ( 1) ]

i j

1 + (i 1) Pi PiT + i j

Pi P Tj ,

i =1 i =1 j =1

j 6 =i

a result that we could also have obtained directly by differentiating Eqn. (1.312) using

Eqn. (1.255). By setting = 1 in the above equation, we get

k k k log log

1 i j

log T = Pi PiT + Pi P Tj .

T

i =1 i i =1 j =1

i j

j 6 =i

For A Sym,

" #

k k k log log

1 i j

A

log A = S Pi Pi + i j Pi P j S, k > 1,

i =1 i i =1 j =1

j 6 =i

1

= S 1 = 2 = 3 ,

(1.313)

An application is presented in Section 2.4.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

We state the divergence, Stokes, potential and localization theorems that are used quite

frequently in the following development. The divergence theorem relates a volume inte-

gral to a surface integral, while the Stokes theorem relates a contour integral to a surface

integral. Let S represent the surface of a volume V, n represent the unit outward normal

to the surface, a scalar field, u a vector field, and T a second-order tensor field. Then we

have

Divergence theorem (also known as the Gauss theorem)

Z Z

dV = n dS. (1.314)

V S

Z Z

u dV = u n dS, (1.315)

Z V ZS

u dV = n u dS,

V

Z ZS

u dV = u n dS.

V S

Z Z

T dV = Tn dS. (1.316)

V S

Note that the divergence theorem is applicable even for multiply connected domains pro-

vided the surfaces are closed.

Stokes theorem

Let C be a contour, and S be the area of any arbitrary surface enclosed by the contour C.

Then

I Z

u dx = ( u) n dS, (1.317)

IC S

Z h i

u dx = ( u)n (u)T n dS. (1.318)

C S

In what follows, we assume n to be C0 continuous over the entire surface S (so that n

makes sense). By taking u as w n in Eqn. (1.317), using (w n) = w (n ), where

= dx/ds is the unit tangent to C, and then applying Eqn. (1.354), we get

I Z

w (n ) ds = [( n)(w n) w + n (w)n] dS. (1.319)

C S

Z Z

( n)(w n) dS = [ I n n] : w dS. (1.320)

S S

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

R

By choosing w to be a constant vector c in the above equation, we get c S ( n ) n dS = 0.

Since the choice of c is arbitrary, we obtain

Z

n dS = 0, (1.321)

S

where = n is the curvature. For a more detailed discussion of this result, see [22]. By

choosing w = x, w = c x, w = ( x x)c and w = ( x x)c in Eqn. (1.320), we get

Z

( x n) dS = 2S, (1.322a)

ZS

( x n) dS = 0, (1.322b)

ZS Z

( x x)n dS = 2 [ I n n] x dS, (1.322c)

ZS Z S

( x n) x dS = [3I n n] x dS. (1.322d)

S S

R R

Note with relation to Eqns. (1.321) and (1.322b) that S n dS and S ( x n ) dS are also zero,

as can be seen by using the divergence theorem.

Potential Theorems

1. Let V be a simply connected region, and let u = be a vector field on V. Then

u = 0. Conversely, if u = 0, then there exists a scalar potential such that

u = .

We have already proved the forwardH assertion (see Eqn. (1.269)). To proveR x the con-

verse, we note from Eqn. (1.317) that C u dx = 0, and thus, the value of x u(y) dy

0

evaluated along any curve in V joining a fixed point x0 and x, just depends on x. Let

Z x

( x) = u(y) dy.

x0

Then u = .

2. If T is a tensor field such that T = 0, then there exists a vector field u such that

T = u.

To see this, using the definition of the curl of a tensor, we have ( T T w) = (

T )w = 0, where w is a constant unit vector. By the above result this implies that there

exists a scalar field such that T T w = , or, alternatively, T = w = (w).

Letting u = w, we get the desired result.

3. If T is a tensor field such that T = 0, and in addition, if tr T = 0, then there

exists W Skw such that T = W.

To prove this, we take the trace of the relation T = u to get u = 0. Let W T =

W be the skew-symmetric tensor whose axial vector is u. Then by Eqn. (1.272), we

get T = u = W.

4. If v and T are vector and tensor fields such that

Z

v n dS = 0,

ZS

T T n dS = 0,

S

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

for every closed surface S in V, then there exist vector and tensor fields u and U such

that v = u and T = U, respectively.

To

R prove the second statement from the first, take the dot product of the equation

T

S T n R = 0 with a constant unit vector w. Then, by the definition of the transpose,

dS

we get S Tw n dS = 0, which by virtue of the first result implies the existence of a

vector field u such that Tw = u, or, alternatively, T = ( u) w = U,

where U = w u.

If the region V does not contain any holes in the interior (i.e., the boundary of V is

comprised of only

R the outer surface), then by means of the divergence theorem, we

conclude that S v n dS = 0 for every closed surface S is equivalent to v = 0.

Thus, for a vector field v that satisfies v = 0 on such a body, there exists a vector

field u such that v = u.

Now we state the localization theorem, which is used in Chapter 3 to obtain the differ-

ential equations from the integral form of the governing equations.

field on V. Then for any given x0 V,

1

Z

( x0 ) = lim dV,

r 0 V ( Br ) Br

R ball of radius r > 0 centered at x0 , and V ( Br ) denotes its volume.

It follows that if B dV = 0 for every closed ball B, then = 0.

Proof. We have

Z

1 1

Z

( x0 ) lim dV = B [( x0 ) ( x)] dV

r 0 V ( Br ) Br V ( Br ) r

1

Z

|( x0 ) ( x)| dV

V ( Br ) Br

sup |( x0 ) ( x)| ,

x Br

R

Note that there is no localization theorem for surfaces, i.e., S dS = 0 for every closed

surface S does not imply that = 0. To see this, take, for example, = a n, where a is a

constant vector.

1.12 Groups

The treatment in this section is based on [35]. A group is a set, say G , equipped with a

function from G G , called combination and denoted by

( a, b) a b,

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

( a b ) c = a ( b c ).

na = an = a a G .

r

3. Existence of reverse elements: For each a G , there exists a G such that

r r

a a = a a = n.

ab = ba a, b G

Since the combination function in the definition of a group may be thought of as taking

two elements from G and producing an element also in G , it is often referred to as a closed

binary operation. The closure property, i.e., the fact that the function value a b is also in G , is

often called the fundamental closure property.

As an example, the set of real numbers is a group with respect to addition. We write

a + b instead of a b, i.e., the combination function is defined on < < to < by

( a, b) a + b.

( a + b) + c = a + (b + c) a, b, c <,

( a) + a = a + ( a) = 0 a <.

Thus, 0 is the neutral element, and reverse elements are the negatives. In fact, < is a com-

mutative group since

Theorem 1.12.1. Let G be a group. Then for any a, b G , there exists a unique x G ,

r

such that a x = b. In fact, x = a b. Similarly, there exists a unique y G such that

r

y a = b. In fact, y = b a.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

r

combining on the left with a, we have

r r

a b = a (a x)

r

= ( a a) x (associativity)

= nx (reverse)

= x. (neutral)

r

Thus, if x exists, it must be precisely a b. This establishes the uniqueness. To

prove existence, one must show that this x has the desired property. First of all,

r

we note that by closure a b G . Next, consider

r r

a ( a b) = ( a a) b (associativity)

= nb (reverse)

= b. (neutral)

r

Hence, x = a b does the job. The result for y can be proved similarly.

b a = c a = b = c.

a b = a c = b = c.

Proof. We have

r r

(b a) a = (c a) a.

b = c.

Proof. Suppose n and n0 are both neutral elements for a group G , i.e., n, n0 G

and for all a G

an = na = a and n0 a = a n0 = a.

Consider a n0 = a. We have

r

n0 = a a = n.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

r r0

Proof. Let a G . Suppose a and a are both reverses of a, i.e.,

r r r0 r0

aa = ar = n and a a = a a = n.

r0

Consider a a = n. We have

r0 r r

a = a n = a.

r

n = n.

r

Proof. For every element a G , the reverse axiom requires a a = n. Taking

r

a = n, we get n n = n. Hence,

r r

n = n n = n.

r

r

a = a.

r

Proof. We have a a = n. Hence,

r r

r r

a = a n = a.

r r r

a b = b a.

Proof. We have

r r r r

( a b) (b a) = a (b (b a)) (associativity)

r r

= a ((b b) a) (associativity)

r

= a (n a) (reverse)

r

= aa (neutral)

= n. (reverse)

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

r r

(b a) ( a b) = n.

r r

Thus, b a is a reverse element of a b. By the uniqueness of the inverse, it is the

reverse.

In general, subsets of groups will not be groups, e.g., the subset G {n}. We introduce

the following terminology:

Definition: Let H be a nonempty [proper] subset of a group G . Then, if H is a group with

respect to the combination function of G , it is called a [proper] subgroup of G .4

Since H H G G , the combination function for H is the restriction of the combina-

tion function for G to H H. However, the function values, which necessarily belong to

G , are not automatically in H, i.e., for an arbitrary subset, closure could fail. An example

of a subgroup of a group G with neutral element n is the singleton set {n}. The following

theorem provides a sufficient condition for a subset to be a subgroup.

r

Theorem 1.12.8. Let H be a nonempty subset of a group G . If for all a, b H, a b

H, then H is a subgroup of G .

Proof. As noted, the combination function for G makes sense for H. Hence, the

associativity requirement is automatically satisfied. We have to show that the

neutral and reverse elements of all a H lie in H, and that H is closed under

the combination operation.

Assume that for a, b H

r

a b H. (1.323)

r r

Choosing b = a, we get a a H. But a a = n. Thus, n H.

r r r

Next, choose a = n. Then, we have n b H. But n b = b. Thus, b H

r

implies b H. So H contains the reverses of all its elements. The reverses

necessarily exist because of the properties of the parent group G .

r r

Finally, let a, b H. We have shown that b H. Letting b play the role of b in

Eqn. (1.323), we have

r

r

a b H.

r

r

But b = b. Thus, a b H, and this proves that H is closed under the combina-

tion operation.

4 The bracket device above is used to make two statements simultaneously. To get the first statement, include

the bracketed material. To get the second leave out the bracketed material.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

We now come to a result that plays a key role in the classification of materials. First,

we define the set of unimodular tensors Unim, and the set of proper unimodular tensors

Unim+ as

Unim : = { H : |det H | = 1},

Theorem 1.12.9. Unim and Unim+ are groups with respect to tensor multiplication,

and Unim+ is a proper subgroup of Unim. Hence, Unim is called the unimodular group

of tensors, while Unim+ is called the proper unimodular group.

Proof. Closure holds since H 1 , H 2 Unim implies that |det( H 1 H 2 )| =

|det H 1 | |det H 2 | = 1. Associativity follows since tensor multiplication is as-

sociative. The reverse element is simply the inverse, while the neutral element

is the identity tensor. Similar arguments can be made for Unim+ . Unim+ is a

proper subgroup of Unim, since there are elements in Unim which are not in

Unim+ , e.g., diag[1, 1, 1].

Theorem 1.12.10. The orthogonal group Orth is a proper subgroup of the unimodular

group Unim, and the proper orthogonal group Orth+ is a proper subgroup of the proper

unimodular group Unim+ .

Proof. The closure property of Orth follows since Q1 , Q2 Orth implies that

Q1 Q2 Orth. The other properties are proved as in the previous proof. Orth is

a proper subgroup since there are unimodular tensors that are not orthogonal.

e.g.,

1 0 0

1 1 0 .

0 0 1

Lemma: Let S0 Psym Unim+ , and let it have at least two distinct eigenvalues, say u

h 2 2 i

and l , with u > l . Then, for every l

u , u , there exists R Orth+ such

l

that the symmetric and proper unimodular tensor

Proof. Since S0 Psym Unim+ , all the eigenvalues of S0 are positive, and we can write

the spectral resolution of S0 as

1

S0 = u e1 e1 + l e2 e2 + e e3 .

l u 3

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

cos sin 0

R( ) = sin cos 0 .

0 0 1

R( )e2 = sin e1 + cos e2 ,

R( )e3 = e3 .

Hence, its eigenvalues are all real. Moreover, +1 is an eigenvalue of T ( ) for all since

= l u S01 R( )S20 R( )e3

= l u S01 R( )S20 e3

1 1

= S R( )e3

l u 0

1 1

= S e

l u 0 3

= e3 .

For = 0, we have R = I, which when substituted into Eqn. (1.324) yields T = I, i.e., the

spectrum of T (0) is {1, 1, 1}. At = /2, we have R(/2)e1 = e2 , R(/2)e2 = e1 , and

R(/2)e3 = e3 . Hence,

T e1 = S01 R S20 R S01 e1

2 2 2

1 1 2

= S R S0 R e

u 0 2 2 1

1

= S01 R S20 e2

u 2

2

= l S01 R e

u 2 2

2l 1

= S e

u 0 1

2l

= e .

2u 1

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

n o

2u 2l 2u

Similarly, T (/2)e2 = e .

2l 2

Thus, the ordered spectrum of T (/2) is 2u

, 1, 2l

.

Since +1 is always an eigenvalue, and since T ( ) depends continuously on , we see that

h 2 2 i

l

given any 2

, u2 , there exists [0, /2] for which the lemma holds.

u l

1

T = e10 e10 + e20 e20 + e30 e30 , (1.325)

h 2 2 i

l

where 2

, u2 . Now we are in a position to prove the main result of this

u l

section. The proof that we present is due to Noll (see [237], page 200).

Group). If G is a group with respect to tensor multiplication such that

Orth G Unim,

then either G = Orth or G = Unim. In other words, there is no group between the

orthogonal group and the unimodular group. The corresponding result for Orth+ and

Unim+ is that Orth+ is maximal in Unim+ .

Proof. There are two possibilities:

1. G contains a tensor S0 Psym Unim with at least two distinct eigenval-

ues, say u > l .

2. G does not contain such a tensor, i.e., the only tensor from Psym Unim

that it contains is the identity tensor.

Consider the first possibility. Using the polar decomposition, any tensor H

Unim can be decomposed as H = QU, where Q is an orthogonal tensor and

U Psym Unim. Hence, if ei are the eigenvectors of U, we can write the

spectral resolution of U as

1

U = 1 e1 e1 + 2 e2 e2 + e3 e3 ,

1 2

U = U 1U 2,

where

1

U 1 = 1 e1 e1 + e2 e2 + e3 e3 ,

1

1

U 2 = e1 e1 + 1 2 e2 e2 + e3 e3 .

1 2

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

1/m1

Now choose integers m1 , m2 large enough so that 1 := 1 and 2 :=

(1 2 )1/m2 satisfy

! !

2l 2u

i .

2u 2l

(Choosing m1 |ln 1 /[2 ln(u /l )]| and m2 |ln(1 2 )/[2 ln(u /l )]| does

the job.) By the Lemma, there exist tensors T 1 and T 2 whose spectral resolution

is given by

1

T 1 = 1 e1 e1 + e2 e2 + e3 e3 ,

1

1

T 2 = e10 e10 + 2 e20 e20 + e30 e30 .

2

Thus,

m 1

T 1 1 = 1 e1 e1 + e2 e2 + e3 e3 ,

1

1 0

T 2m2 = e10 e10 + 1 2 e20 e20 + e e30 .

1 2 3

Let R1 and R2 be the two proper orthogonal tensors that rotate ei and ei0 into ei .

Then, we have

m 1

R1 T 1 1 R1T = R1 (1 e1 e1 + e2 e2 + e3 e3 ) R1T

1

1

= 1 ( R1 e1 ) ( R1 e1 ) + ( Re2 ) ( R1 e2 ) + ( R1 e3 ) ( R1 e3 )

1

1

= 1 e1 e1 + e2 e2 + e3 e3

1

= U 1.

Similarly,

R2 T 2m2 R2T = U 2 .

Hence,

H = QU

= QU 1 U 2

= Q R1 T 1m1 R1T R2 T 2m2 R2T

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

h i m1 h i m2

= QR1 S01 R(1 )S20 R1 (1 )S01 R1T R2 S01 R(2 )S20 R1 (2 )S01 R2T ,

(1.326)

for some 1 , 2 [0, /2]. Thus, any H Unim can be expressed in terms of

powers of S0 Psym Unim, Q Orth, and R(1 ), R(2 ), R1 , R2 Orth+ . But

Orth G and S0 G , and since G is a group, any element generated by powers

of S0 and elements of Orth has to lie in G . Thus, we have Unim G . But, by

hypothesis, G Unim. Hence, for the case when S0 has at least two distinct

eigenvalues G = Unim.

Now consider the case when the only member from Psym Unim in G is the

identity tensor. For any H G , we have H = QU, or, alternatively, U = Q1 H,

where Q Orth G , and U Psym Unim. Since G is a group, and since

H, Q G , we have U = Q1 H G . However, since the only member from

Psym Unim in G is I, we have U = I. Therefore, H = Q Orth, i.e., G Orth.

But, by hypothesis, Orth G . Hence, in this case, we have G = Orth.

To prove that Orth+ is maximal in Unim+ , note that if H Unim+ , then Q in

Eqn. (1.326) belongs to Orth+ . Thus, any H Unim+ can be expressed in terms

of powers of S0 Psym Unim and tensors in Orth+ , and the result follows.

EXERCISES

1. Show that

[u v, v w, w u] = [u, v, w]2 . (1.328)

[n, n u, n v] = [n, u, v] .

2. Which of the spaces Sym, Psym, Skw, Orth+ are linear subspaces of Lin? Justify.

Evaluate if the matrices

0 1 1 0 0 1 0 1 0

1 0 0 , 0 0 1 , 1 0 1 ,

1 0 0 1 1 0 0 1 0

1 1 0 1 0 1 0 0 0

1 1 0 , 0 0 0 , 0 1 1 ,

0 0 0 1 0 1 0 1 1

constitute a basis for Sym. If not, then give a canonical basis for Skw and Sym. Also

give a basis for deviatoric symmetric tensors.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

qi }, i = 1, 2, . . . , n, is linearly dependent or independent if

(a) { pi } is linearly independent, and {qi } is linearly independent;

(b) { pi } is linearly independent, but {qi } is not;

(c) {qi } is linearly independent, but { pi } is not;

(d) { pi } is linearly dependent, and {qi } is also linearly dependent.

independent. Deduce from this result and Eqn. (1.105a) if { I, Q, Q T }, where Q

Orth+ , is linearly dependent or independent.

5. Show that the decomposition of a tensor into a symmetric and skew-symmetric part

as given by Eqn. (1.33) is unique.

6. Show that tr RS = tr SR = tr R T S T = tr S T R T .

7. If S and W are symmetric and skew-symmetric tensors, respectively, and T is an

arbitrary second-order tensor prove that

1 T

S : T = S : TT = S : (T + T ) , (1.329)

2

1

W : T = W : T T = W : (T T T ) ,

2

S : W = 0. (1.330)

(i) If A : B = 0 for every symmetric tensor B, then A Skw.

(ii) If A : B = 0 for every skew tensor B, then A Sym.

9. Using Eqns. (1.35) and (1.52) prove that

det( RS) = (det R)(det S).

1h i

tr (cof T ) = (tr T )2 tr ( T 2 ) , (1.331)

2

[(cof T )u] v = T (u T T v), (1.332)

cof (cof T ) = (det T ) T.

a b + b c + c a,

by showing, in particular, that the identity tensor I cannot be represented in this way.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

(u v) w + (v w) u + (w u) v = [u, v, w] I.

Using Eqn. (1.53), show that Eqn. (1.328) is recovered for the case [u, v, w] 6= 0

(Eqn. (1.328) holds even when [u, v, w] = 0). Note that the principal invariants of

T are

I1 = tr T = u p + v q + w r,

I2 = tr (cof T ) = (u v) ( p q) + (v w) (q r ) + (w u) (r p),

I3 = det T = [u, v, w] [ p, q, r ] ,

13. Let i , i = 1, 2, , n, denote the eigenvalues of T for space-dimension n. Using

Eqn. (J.11), show that the eigenvalues of cof T are nj=1 j , i = 1, 2, . . . , n. For the

j 6 =i

case n = 3, deduce Eqn. (1.331), I2 (cof T ) = tr T (det T ) and I3 (cof T ) = (det T )2 .

Now consider the case when T is diagonalizable with distinct eigenvalues, so that

T = in=1 i Pi (thus, T I = in=1 (i )Pi ; tr Pi = e e = 1).

h Using Eqn. (J.11),

i

n n

show that cof T = i=1 j=1 j Pi , and cof ( T I ) = i=1 nj=1 ( j ) PiT .

n

T

j 6 =i j 6 =i

Thus, for = 1 say, cof ( T 1 I ) = nj=2 ( j 1 )P1T , and tr cof ( T 1 I ) =

nj=2 ( j 1 ).

14. If T, R and S are second-order tensors, then show that

1 1

det T = tr [(cof T ) T T ] = T : cof T, (1.334)

3 3

det( R + S) = det R + cof R : S + R : cof S + det S, (1.335)

cof ( R + S) = cof R + cof S + [(tr R)(tr S) tr ( RS)] I (tr R)S (tr S) R T T

By putting S = u v in the above equation and using Eqn. (1.84), deduce that

cof ( I + u v) = (1 + u v) I v u.

I W +ww

( I + W ) 1 = .

1 + | w |2

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

15. Apply the CayleyHamilton theorem to W Skw, and find the axial vector of W 3 .

16. If w is the axial vector of W Skw, show that w [u W u] 0 u, with equality if

and only if (u, w) are linearly dependent.

17. Let W Skw, and let w be its axial vector. Using Eqn. (1.332), show that (cof T )w is

the axial vector of TW T T . If T Q Orth+ , then deduce using Eqn. (1.93) that Qw

is the axial vector of QW Q T . Also, find the axial vector of QW 3 Q T .

18. Show that a (not necessarily symmetric) tensor T commutes with every skew tensor

W if and only if T = I.

19. Either using the result of the previous problem, or independently, show that a (not

necessarily symmetric) tensor T commutes with every orthogonal tensor Q if and

only if T = I.

20. Show that a (not necessarily symmetric) tensor T commutes with every symmetric

tensor S if and only if T = I.

21. Show that (u, Tv) = 0 u, v that are mutually orthogonal, if an only if T = I.

22. Let a, b be arbitrary vectors, and u, v be unit vectors such that u v 6= 1.

(a) Show that a b is the axial vector of the skew tensor b a a b. It follows

that if b a = a b, then a b = 0, so that by Theorem 1.2.1, a and b are

linearly dependent.

(b) Using (a) and Eqn. (1.104), show that the unique orthogonal tensor that rotates

u into v about an axis perpendicular to u and v is given by

R = I + rv u s(u u + u v + v v).

23. If w is the axial vector of W Skw, then show using Eqns. (1.54) and (1.327) (or

Eqns. (1.60) and (1.90)) that cof W = w w. Deduce using Eqn. (1.273) that

2

cof W ( e | w | 1)

e = I+ w w.

| w |2

2

The determinant of the above tensor is etr (cof W ) = e|w| .

24. Let W 1 and W 2 be skew-symmetric tensors with axial vectors w1 and w2 . Show

using Eqns. (1.21a) and (1.87) that

W 1 W 2 = w2 w1 (w1 w2 ) I, (1.338)

result of Problem 22a (or, independently, by means of Eqns. (1.21)), we see that w1

w2 is the axial vector of W 1 W 2 W 2 W 1 . If w1 = w2 , where <, then, by

Eqn. (1.90), W 1 = W 2 , and hence W 1 W 2 = W 2 W 1 . Conversely, if W 1 W 2 = W 2 W 1 ,

then w1 w2 , the axial vector of W 1 W 2 W 2 W 1 , is 0, and by Theorem 1.2.1, it

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

follows that w1 and w2 are linearly dependent. Thus, two skew-symmetric tensors

W 1 and W 2 commute if and only if their axial vectors are linearly dependent.

It also follows from Eqn. (1.338) that

W 1 W 2 + W 2 W 1 = w1 w2 + w2 w1 2(w1 w2 ) I.

w1 w

(|w1 | |w2 | w1 w2 ), + 2 ,

| w1 | | w2 |

w w

(|w1 | |w2 | + w1 w2 ), 1 2 ,

| w1 | | w2 |

2( w1 w2 ), w1 w2 .

One can now easily compute the principal invariants of W 1 W 2 + W 2 W 1 using these

eigenvalues.

25. For W Skw, find the polar decomposition of I + W in terms of W and its axial

vector w.

26. Prove that there is a one-to-one correspondence between Skw and members of Orth+

with no eigenvalue equal to 1,5 by showing that

(a) For every W Skw, Q = ( I W )1 ( I + W ) Orth+ . If w is the axial vector

of W, then for n = 3, show using Eqn. (1.66) and Problem 14 that the above

expression simplifies to

1 h i

Q= (1 + w w) I + 2W + 2W 2 , (1.339)

1+ww

with w/ |w| as the axis. In expanded form, if

0

W= 0 ,

then

1 + 2 2 2 2( ) 2( + )

1

Q= 2( + ) 1 2 + 2 2 2( ) .

1 + 2 + 2 + 2

2( + ) 2( + ) 1 2 2 + 2

5 For members of Orth+ with at least one eigenvalue equal to 1, the correspondence when the dimension n is

three, is given by

Q = I + 2W 2 ,

where W is of the form

0 sin cos sin

W= sin 0 cos cos .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

By comparing Eqn. (1.104) with Eqn. (1.339), we see that sin 2 |w| /(1 +

|w|2 ) and cos = (1 |w|2 )/(1 + |w|)2 , which corresponds to letting [0, )

in Eqn. (1.104). If w1 / |w1 | and w2 / |w2 | are the axes of R1 and R2 , respectively,

then, from Eqn. (1.107) and the above expressions for cos and sin , the axis of

R2 R1 is e = w/ |w|, where

w = w1 + w2 w1 w2 .

W = ( Q I )( Q + I )1 Skw.

Thus, the number of parameters required to define an orthogonal tensor is the num-

ber of independent components in a skew-symmetric matrix, namely n(n 1)/2; e.g.,

for n = 3, the number of parameters is 3.

27. If u is an arbitrary vector, show that there exists W Skw and a vector v (both

dependent on u) such that W v = u. Use this result to show that if W T 1 = W T 2 W

Skw, then T 1 = T 2 .

28. Show that a scalar-valued function : Skw < is isotropic if and only if there exists

a function : tr (W 2 ) < such that

(W ) = (tr W 2 ) W Skw.

The above result follows directly as a corollary of Theorem 1.6.13, or can be proved

independently by using the representation given by Eqn. (1.89), and mimicking the

proof of Theorem 1.6.12.

29. For S Sym, show using the spectral resolution of S or otherwise that

(u, Su) = 0 u V,

30. A set of Cartesian axes is rotated about the origin to coincide with the unit vectors

3 1 3

e1 = ( , , ),

4 4 2

3 3 1

e2 = ( , , ),

4 4 2

1 3

e3 = ( , , 0).

2 2

Write down the rotation matrix corresponding to this rotation, and transform the

components of the tensor T with respect to {ei } given by

8 4 0

[ T ] = 4 3 1 .

0 1 2

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

31. Using direct notation show that (also verify using indicial notation)

( a b)T = b a, (1.340)

( a b)(c d) = (b c) a d, (1.341)

( a b) : (u v) = ( a u)(b v), (1.342)

T ( a b) = ( T a) b, (1.343)

( a b ) T = a ( T T b ), (1.344)

T : ( a b) = a Tb. (1.345)

Use Eqn. (1.341) to show that for a symmetric S, the spectral decompositions of Sn

and S1 (when it exists) are

Sn = 1n e1 e1 + 2n e2 e2 + 3n e3 e3 ,

S1 = 11 e1 e1 + 21 e2 e2 + 31 e3 e3 .

32. The Fibonacci numbers satisfy the recurrence relation Fn+1 = Fn + Fn1 with F1 =

F2 = 1. In matrix form this recurrence relation can be written as

" # " #

Fn+1 Fn

=S ,

Fn Fn1

where S = 11 10 . Use the above matrix form to first express FnF+1 in terms of FF2 ,

n 1

and then use this to find an explicit expression for Fn in terms of the eigenvalues 1

and 2 of S.

33. Let B be an invertible tensor. Show that if n is an eigenvector of A, then Bn is an

eigenvector of BAB1 corresponding to the same eigenvalue. It follows that if Q

Orth+ and if n is an eigenvector of a tensor T, then Qn is an eigenvector of QTQ T

corresponding to the same eigenvalue.

34. If the eigenvalues of a symmetric tensor S are ordered such that 1 2 3 , show

using the spectral decomposition of S that

u Su u Su

1 = min , 3 = max .

u uu u uu

tensor). Use this result and the CauchySchwartz inequality to show that if R, S

Lin, then

| RS| | R| |S| .

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

ever, among the

two quantities |( R, S)| and | RS|, either can be greater. For

example,

let R = 11 20 ; if S = 10 11 , then |( R, S)| = 3 and | RS| = 2, while if S = 01 11 , then

37. Show that Sn , S1 and (tr S) I + 2S are isotropic functions.

38. Show that for any orthogonal Q, the tensor QQ T is skew at each t.

39. Show using Eqn. (1.336) that if G ( T ) = cof T, then

(1.346)

Gij

= eikm e jln Tmn . (1.348)

Tkl

Equation (1.348) in direct tensorial notation, obtained using Eqns. (1.55), (1.226a),

(1.226b) and (1.233), is given by

(cof T ) = tr T [ I I T] ( I T T + T T I ) + ( I T )T + ( T T I )T.

T

(1.349)

If T C Sym, then, by virtue of Eqn. (1.241), we just pre- and post-multiply the

above result with S, and use Eqn. (1.175) to get

(cof C ) = tr C [ I I S] ( I C + C I ) + S [ I C + C I ] S. (1.350)

C

The indicial notation form of the above equation, obtained using Eqns. (1.161) and

(1.348), is (for the engineering form, see Eqn. (I.6))

1h i

(cof C ) = eikr e jls + eilr e jks + e jkr eils + e jlr eiks Crs . (1.351)

C ijkl 4

Finally, consider the case when C Sym is invertible, i.e., cof C = (det C )C 1 . Show

using Eqns. (1.236) and (1.240) that

h i

(cof C ) = det C C 1 C 1 S(C 1 C 1 )S . (1.352)

C

(cof T ) : (cof T ) = ( T : T )2 ( T T T ) : ( T T T ) /2. Using either Eqns. (1.228)(1.237)

or by means of the directional derivative, show that

1

= 2( T T )3 ,

T

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

i i

i i

2

= ( T 1 : T 1 )cof T 2(det T ) T T T 1 T T ,

T

3

= 2( T : T ) T 2T T T T.

T

(v) = v + v ,

(v) = ( v) + v (), (1.353)

(u v) = ( v)u (v)u ( u)v + (u)v, (1.354)

(u v) = (u)T v + (v)T u,

[(u v)w] = (u v)w + w [(v)T u] + w [(u)T v],

(u v) = u v + (u)v,

( T T v) = T : v + v ( T ), (1.355)

(T ) = T + T ,

(T ) = T + T ,

(Tv) = ( Tv) + ( Tv) ,

2 (u v) = u 2 v + v 2 u + 2u : v,

[ a ( b)] = ( a) ( b) a [ ( b)]. (1.356)

By integrating Eqn. (1.356) over a closed volume V with surface S show that

Z Z

( b) ( a n) dS = {( a) ( b) a [ ( b)]} dV.

S V

u : (u)T = [(u)u ( u)u] + ( u)2 .

Z h i Z

u ( T T ) + (u) T dV = u ( T T n) dS.

V S

44. Show that < {0} is a commutative group with respect to multiplication.

available at https:/www.cambridge.org/core/terms. https://doi.org/10.1017/CBO9781316134054.002

i i

i i

- Statics- Force Vector 3Enviado porDhanis Paramaguru
- Physics I Problems (6)Enviado porBOSS BOSS
- 17 FY13CE Maths Detail SolutionsEnviado porlaukkeas
- LinEnviado porEpic Win
- IIT QUIZEnviado pordivyaevats
- Mathematical Physics1Enviado porSergio Aguilar
- ch1Enviado porSalah Rahimi
- Lesson 1Enviado porMary Joan Miranda
- Moments in 3D.pdfEnviado porPatricia Tungpalan
- 13885_Lecture Notes 2.9Enviado porMonachikós Drákos
- BEZIEREnviado porHardy Widjaja
- Fall-2008-3[1]Enviado porAli Ahmad
- katznelson-linalgebraEnviado porsaradha_ramachandran
- p4 davis alexander my vector lifeEnviado porapi-255204476
- linfun2Enviado porTushar Gupta
- AlgorithmsEnviado porajitkarthik
- At JCM H2 Topic12 Notes 37pagesEnviado porwill bell
- Hello WorldEnviado porMatteo Esposito
- w1-CEnviado porrahul
- Vector ProblemsEnviado porsourav singh
- KernelTrick-LectureNotesPublic.pdfEnviado porEdmond Z
- 4-2Enviado porGautham Shankar
- ch2Enviado porM
- i Semester-upstream 2016Enviado porAkashShukla
- INFORMATION PROCESSING WITH POPULATION CODES.pdfEnviado porrorozainos
- weekly test physics.docxEnviado porAshok Pradhan
- IndependenceEnviado porAndrew Fong
- matrices_rotations.docxEnviado porAndreia Nicolau
- Chapter 2Enviado porArmenion Mark Allen
- MM09ps1Enviado porRoy Vesey

- Multi ObjectiveEnviado porPramodVerma
- Clash of ValuesEnviado porPramodVerma
- AdvHydMotionDesignPractices-Paso.pdfEnviado porPramodVerma
- DRDO Mechanical Engineering Question paper.pdfEnviado porAjay Kumar Verma
- ProcessEnviado porPramodVerma
- High Rate NanomanufacturingEnviado porPramodVerma
- Technology vs EnvironmentEnviado porPramodVerma
- 461_1Enviado porPramodVerma
- Lect5 Detachable Fastener Bolt Nut Locking FoundationEnviado porPramodVerma
- CME2Enviado porPramodVerma
- CME1Enviado porPramodVerma
- Order of Magnitude AnalysisEnviado porPramodVerma
- art%3A10.3938%2Fjkps.62.794.pdfEnviado porPramodVerma
- question_bank.pdfEnviado porPramodVerma
- Slides Synthesis 1Enviado porPramodVerma
- computational mechanics.pdfEnviado porPramodVerma
- Experimental and Numerical Investigations of Turbulent Flow Induced Pipe Vibration in Fully Developed Flow_2004Enviado porPramodVerma
- ECD Flow Through NozzleEnviado porPramodVerma
- Flow and Headlosses in PipeEnviado porPramodVerma
- Flow and headlosses in Pipe.PDFEnviado porPramodVerma

- Linear Algebra BookEnviado porXavii Xordan
- Ch42Enviado porAji Permana Putra
- Linear Network CodingEnviado porSharanappa C Ijeri
- Advanced Optimistion (1)Enviado porMallikarjun Valipi
- Answer Key_CK-12 PreCalculus ConceptsEnviado porFrederichNietszche
- Additions Vectors.pdfEnviado porRene Romero Velasquez
- ch2Enviado porM
- 1608761754 GroupEnviado porDiana Uriza
- Mech 200 Lin Alg Packet SolnEnviado portennisstarbyfar
- Multilinear Algebra PDFEnviado porCarlos Sanchez
- Multilinear Algebra - MITEnviado porasd
- Coordinates, Vector Spaces, And Linear TransformationsEnviado porzcqsmcm
- Projective Geometry Lecture Notes 3Enviado porjjwfish
- ABSTRACT,LINEAREnviado porMonty Sigmund
- Unit 3. Analytic GeometryEnviado porEstherGarcíaJordán
- Chiang - Chapter 5Enviado porMark Anthony Castillano
- 6. Vector Algebra QEnviado porRamchandra Murthy
- CL 701 Lecture Notes 07Enviado porMusa Mohammed
- essential physicsEnviado porputih_138242459
- Continuum Dynamics NotesEnviado porAaronWienkers
- Data Mining Text BookEnviado porsatpalwadhwa
- TutorialMath3171(2)Enviado porbadar
- M. Spivak - Calculus on manifolds.pdfEnviado porEstudante de Ciências
- Synchronous Data Flow AlgorithmsEnviado porcpayne10409
- 129097905-Engineering-Mathematics.pdfEnviado porGowsigan Adhimoolam
- QRG_CS.pdfEnviado porckvirtualize
- m308Enviado porKuber Bakthan
- Basis and DimensionEnviado porLalit Narayan
- [Non-QU] Linear Algebra by Dr.Gabriel Nagy.pdfEnviado poramr
- Solution Manual for Fourier Transforms – Eric HansenEnviado porAfshar Arabi

## Muito mais do que documentos

Descubra tudo o que o Scribd tem a oferecer, incluindo livros e audiolivros de grandes editoras.

Cancele quando quiser.